bcd7890124
* move test factories to libp2p/tools * remove unused inits * move pubsub test utils to tools * cleanup test_interop * fix typing libp2p/tools/utils * add typing to pubsub utils * fix factories typing * fix typing for floodsub_integration_test_settings * fix rest of the typing * fix isort
262 lines
10 KiB
Python
262 lines
10 KiB
Python
# type: ignore
|
|
# To add typing to this module, it's better to do it after refactoring test cases into classes
|
|
|
|
import asyncio
|
|
|
|
import pytest
|
|
|
|
from libp2p.tools.constants import FLOODSUB_PROTOCOL_ID, LISTEN_MADDR
|
|
from libp2p.tools.factories import PubsubFactory
|
|
from libp2p.tools.utils import connect
|
|
|
|
SUPPORTED_PROTOCOLS = [FLOODSUB_PROTOCOL_ID]
|
|
|
|
FLOODSUB_PROTOCOL_TEST_CASES = [
|
|
{
|
|
"name": "simple_two_nodes",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"A": ["B"]},
|
|
"topic_map": {"topic1": ["B"]},
|
|
"messages": [{"topics": ["topic1"], "data": b"foo", "node_id": "A"}],
|
|
},
|
|
{
|
|
"name": "three_nodes_two_topics",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"A": ["B"], "B": ["C"]},
|
|
"topic_map": {"topic1": ["B", "C"], "topic2": ["B", "C"]},
|
|
"messages": [
|
|
{"topics": ["topic1"], "data": b"foo", "node_id": "A"},
|
|
{"topics": ["topic2"], "data": b"Alex is tall", "node_id": "A"},
|
|
],
|
|
},
|
|
{
|
|
"name": "two_nodes_one_topic_single_subscriber_is_sender",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"A": ["B"]},
|
|
"topic_map": {"topic1": ["B"]},
|
|
"messages": [{"topics": ["topic1"], "data": b"Alex is tall", "node_id": "B"}],
|
|
},
|
|
{
|
|
"name": "two_nodes_one_topic_two_msgs",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"A": ["B"]},
|
|
"topic_map": {"topic1": ["B"]},
|
|
"messages": [
|
|
{"topics": ["topic1"], "data": b"Alex is tall", "node_id": "B"},
|
|
{"topics": ["topic1"], "data": b"foo", "node_id": "A"},
|
|
],
|
|
},
|
|
{
|
|
"name": "seven_nodes_tree_one_topics",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"1": ["2", "3"], "2": ["4", "5"], "3": ["6", "7"]},
|
|
"topic_map": {"astrophysics": ["2", "3", "4", "5", "6", "7"]},
|
|
"messages": [{"topics": ["astrophysics"], "data": b"e=mc^2", "node_id": "1"}],
|
|
},
|
|
{
|
|
"name": "seven_nodes_tree_three_topics",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"1": ["2", "3"], "2": ["4", "5"], "3": ["6", "7"]},
|
|
"topic_map": {
|
|
"astrophysics": ["2", "3", "4", "5", "6", "7"],
|
|
"space": ["2", "3", "4", "5", "6", "7"],
|
|
"onions": ["2", "3", "4", "5", "6", "7"],
|
|
},
|
|
"messages": [
|
|
{"topics": ["astrophysics"], "data": b"e=mc^2", "node_id": "1"},
|
|
{"topics": ["space"], "data": b"foobar", "node_id": "1"},
|
|
{"topics": ["onions"], "data": b"I am allergic", "node_id": "1"},
|
|
],
|
|
},
|
|
{
|
|
"name": "seven_nodes_tree_three_topics_diff_origin",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"1": ["2", "3"], "2": ["4", "5"], "3": ["6", "7"]},
|
|
"topic_map": {
|
|
"astrophysics": ["1", "2", "3", "4", "5", "6", "7"],
|
|
"space": ["1", "2", "3", "4", "5", "6", "7"],
|
|
"onions": ["1", "2", "3", "4", "5", "6", "7"],
|
|
},
|
|
"messages": [
|
|
{"topics": ["astrophysics"], "data": b"e=mc^2", "node_id": "1"},
|
|
{"topics": ["space"], "data": b"foobar", "node_id": "4"},
|
|
{"topics": ["onions"], "data": b"I am allergic", "node_id": "7"},
|
|
],
|
|
},
|
|
{
|
|
"name": "three_nodes_clique_two_topic_diff_origin",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"1": ["2", "3"], "2": ["3"]},
|
|
"topic_map": {"astrophysics": ["1", "2", "3"], "school": ["1", "2", "3"]},
|
|
"messages": [
|
|
{"topics": ["astrophysics"], "data": b"e=mc^2", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic", "node_id": "1"},
|
|
],
|
|
},
|
|
{
|
|
"name": "four_nodes_clique_two_topic_diff_origin_many_msgs",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {
|
|
"1": ["2", "3", "4"],
|
|
"2": ["1", "3", "4"],
|
|
"3": ["1", "2", "4"],
|
|
"4": ["1", "2", "3"],
|
|
},
|
|
"topic_map": {
|
|
"astrophysics": ["1", "2", "3", "4"],
|
|
"school": ["1", "2", "3", "4"],
|
|
},
|
|
"messages": [
|
|
{"topics": ["astrophysics"], "data": b"e=mc^2", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar2", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic2", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar3", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic3", "node_id": "1"},
|
|
],
|
|
},
|
|
{
|
|
"name": "five_nodes_ring_two_topic_diff_origin_many_msgs",
|
|
"supported_protocols": SUPPORTED_PROTOCOLS,
|
|
"adj_list": {"1": ["2"], "2": ["3"], "3": ["4"], "4": ["5"], "5": ["1"]},
|
|
"topic_map": {
|
|
"astrophysics": ["1", "2", "3", "4", "5"],
|
|
"school": ["1", "2", "3", "4", "5"],
|
|
},
|
|
"messages": [
|
|
{"topics": ["astrophysics"], "data": b"e=mc^2", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar2", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic2", "node_id": "1"},
|
|
{"topics": ["school"], "data": b"foobar3", "node_id": "2"},
|
|
{"topics": ["astrophysics"], "data": b"I am allergic3", "node_id": "1"},
|
|
],
|
|
},
|
|
]
|
|
|
|
floodsub_protocol_pytest_params = [
|
|
pytest.param(test_case, id=test_case["name"])
|
|
for test_case in FLOODSUB_PROTOCOL_TEST_CASES
|
|
]
|
|
|
|
|
|
async def perform_test_from_obj(obj, router_factory) -> None:
|
|
"""
|
|
Perform pubsub tests from a test obj.
|
|
test obj are composed as follows:
|
|
|
|
{
|
|
"supported_protocols": ["supported/protocol/1.0.0",...],
|
|
"adj_list": {
|
|
"node1": ["neighbor1_of_node1", "neighbor2_of_node1", ...],
|
|
"node2": ["neighbor1_of_node2", "neighbor2_of_node2", ...],
|
|
...
|
|
},
|
|
"topic_map": {
|
|
"topic1": ["node1_subscribed_to_topic1", "node2_subscribed_to_topic1", ...]
|
|
},
|
|
"messages": [
|
|
{
|
|
"topics": ["topic1_for_message", "topic2_for_message", ...],
|
|
"data": b"some contents of the message (newlines are not supported)",
|
|
"node_id": "message sender node id"
|
|
},
|
|
...
|
|
]
|
|
}
|
|
NOTE: In adj_list, for any neighbors A and B, only list B as a neighbor of A
|
|
or B as a neighbor of A once. Do NOT list both A: ["B"] and B:["A"] as the behavior
|
|
is undefined (even if it may work)
|
|
"""
|
|
|
|
# Step 1) Create graph
|
|
adj_list = obj["adj_list"]
|
|
node_map = {}
|
|
pubsub_map = {}
|
|
|
|
async def add_node(node_id_str: str) -> None:
|
|
pubsub_router = router_factory(protocols=obj["supported_protocols"])
|
|
pubsub = PubsubFactory(router=pubsub_router)
|
|
await pubsub.host.get_network().listen(LISTEN_MADDR)
|
|
node_map[node_id_str] = pubsub.host
|
|
pubsub_map[node_id_str] = pubsub
|
|
|
|
tasks_connect = []
|
|
for start_node_id in adj_list:
|
|
# Create node if node does not yet exist
|
|
if start_node_id not in node_map:
|
|
await add_node(start_node_id)
|
|
|
|
# For each neighbor of start_node, create if does not yet exist,
|
|
# then connect start_node to neighbor
|
|
for neighbor_id in adj_list[start_node_id]:
|
|
# Create neighbor if neighbor does not yet exist
|
|
if neighbor_id not in node_map:
|
|
await add_node(neighbor_id)
|
|
tasks_connect.append(
|
|
connect(node_map[start_node_id], node_map[neighbor_id])
|
|
)
|
|
# Connect nodes and wait at least for 2 seconds
|
|
await asyncio.gather(*tasks_connect, asyncio.sleep(2))
|
|
|
|
# Step 2) Subscribe to topics
|
|
queues_map = {}
|
|
topic_map = obj["topic_map"]
|
|
|
|
tasks_topic = []
|
|
tasks_topic_data = []
|
|
for topic, node_ids in topic_map.items():
|
|
for node_id in node_ids:
|
|
tasks_topic.append(pubsub_map[node_id].subscribe(topic))
|
|
tasks_topic_data.append((node_id, topic))
|
|
tasks_topic.append(asyncio.sleep(2))
|
|
|
|
# Gather is like Promise.all
|
|
responses = await asyncio.gather(*tasks_topic)
|
|
for i in range(len(responses) - 1):
|
|
node_id, topic = tasks_topic_data[i]
|
|
if node_id not in queues_map:
|
|
queues_map[node_id] = {}
|
|
# Store queue in topic-queue map for node
|
|
queues_map[node_id][topic] = responses[i]
|
|
|
|
# Allow time for subscribing before continuing
|
|
await asyncio.sleep(0.01)
|
|
|
|
# Step 3) Publish messages
|
|
topics_in_msgs_ordered = []
|
|
messages = obj["messages"]
|
|
tasks_publish = []
|
|
|
|
for msg in messages:
|
|
topics = msg["topics"]
|
|
data = msg["data"]
|
|
node_id = msg["node_id"]
|
|
|
|
# Publish message
|
|
# TODO: Should be single RPC package with several topics
|
|
for topic in topics:
|
|
tasks_publish.append(pubsub_map[node_id].publish(topic, data))
|
|
|
|
# For each topic in topics, add (topic, node_id, data) tuple to ordered test list
|
|
for topic in topics:
|
|
topics_in_msgs_ordered.append((topic, node_id, data))
|
|
|
|
# Allow time for publishing before continuing
|
|
await asyncio.gather(*tasks_publish, asyncio.sleep(2))
|
|
|
|
# Step 4) Check that all messages were received correctly.
|
|
for topic, origin_node_id, data in topics_in_msgs_ordered:
|
|
# Look at each node in each topic
|
|
for node_id in topic_map[topic]:
|
|
# Get message from subscription queue
|
|
msg = await queues_map[node_id][topic].get()
|
|
assert data == msg.data
|
|
# Check the message origin
|
|
assert node_map[origin_node_id].get_id().to_bytes() == msg.from_id
|
|
|
|
# Success, terminate pending tasks.
|