import random import pytest import trio from libp2p.tools.factories import PubsubFactory from libp2p.tools.pubsub.utils import dense_connect, one_to_all_connect from libp2p.tools.utils import connect @pytest.mark.trio async def test_join(): async with PubsubFactory.create_batch_with_gossipsub( 4, degree=4, degree_low=3, degree_high=5 ) as pubsubs_gsub: gossipsubs = [pubsub.router for pubsub in pubsubs_gsub] hosts = [pubsub.host for pubsub in pubsubs_gsub] hosts_indices = list(range(len(pubsubs_gsub))) topic = "test_join" central_node_index = 0 # Remove index of central host from the indices hosts_indices.remove(central_node_index) num_subscribed_peer = 2 subscribed_peer_indices = random.sample(hosts_indices, num_subscribed_peer) # All pubsub except the one of central node subscribe to topic for i in subscribed_peer_indices: await pubsubs_gsub[i].subscribe(topic) # Connect central host to all other hosts await one_to_all_connect(hosts, central_node_index) # Wait 2 seconds for heartbeat to allow mesh to connect await trio.sleep(2) # Central node publish to the topic so that this topic # is added to central node's fanout # publish from the randomly chosen host await pubsubs_gsub[central_node_index].publish(topic, b"data") # Check that the gossipsub of central node has fanout for the topic assert topic in gossipsubs[central_node_index].fanout # Check that the gossipsub of central node does not have a mesh for the topic assert topic not in gossipsubs[central_node_index].mesh # Central node subscribes the topic await pubsubs_gsub[central_node_index].subscribe(topic) await trio.sleep(2) # Check that the gossipsub of central node no longer has fanout for the topic assert topic not in gossipsubs[central_node_index].fanout for i in hosts_indices: if i in subscribed_peer_indices: assert hosts[i].get_id() in gossipsubs[central_node_index].mesh[topic] assert hosts[central_node_index].get_id() in gossipsubs[i].mesh[topic] else: assert ( hosts[i].get_id() not in gossipsubs[central_node_index].mesh[topic] ) assert topic not in gossipsubs[i].mesh @pytest.mark.trio async def test_leave(): async with PubsubFactory.create_batch_with_gossipsub(1) as pubsubs_gsub: gossipsub = pubsubs_gsub[0].router topic = "test_leave" assert topic not in gossipsub.mesh await gossipsub.join(topic) assert topic in gossipsub.mesh await gossipsub.leave(topic) assert topic not in gossipsub.mesh # Test re-leave await gossipsub.leave(topic) @pytest.mark.trio async def test_handle_graft(monkeypatch): async with PubsubFactory.create_batch_with_gossipsub(2) as pubsubs_gsub: gossipsubs = tuple(pubsub.router for pubsub in pubsubs_gsub) index_alice = 0 id_alice = pubsubs_gsub[index_alice].my_id index_bob = 1 id_bob = pubsubs_gsub[index_bob].my_id await connect(pubsubs_gsub[index_alice].host, pubsubs_gsub[index_bob].host) # Wait 2 seconds for heartbeat to allow mesh to connect await trio.sleep(2) topic = "test_handle_graft" # Only lice subscribe to the topic await gossipsubs[index_alice].join(topic) # Monkey patch bob's `emit_prune` function so we can # check if it is called in `handle_graft` event_emit_prune = trio.Event() async def emit_prune(topic, sender_peer_id): event_emit_prune.set() monkeypatch.setattr(gossipsubs[index_bob], "emit_prune", emit_prune) # Check that alice is bob's peer but not his mesh peer assert id_alice in gossipsubs[index_bob].peers_gossipsub assert topic not in gossipsubs[index_bob].mesh await gossipsubs[index_alice].emit_graft(topic, id_bob) # Check that `emit_prune` is called await event_emit_prune.wait() # Check that bob is alice's peer but not her mesh peer assert topic in gossipsubs[index_alice].mesh assert id_bob not in gossipsubs[index_alice].mesh[topic] assert id_bob in gossipsubs[index_alice].peers_gossipsub await gossipsubs[index_bob].emit_graft(topic, id_alice) await trio.sleep(1) # Check that bob is now alice's mesh peer assert id_bob in gossipsubs[index_alice].mesh[topic] @pytest.mark.trio async def test_handle_prune(): async with PubsubFactory.create_batch_with_gossipsub( 2, heartbeat_interval=3 ) as pubsubs_gsub: gossipsubs = tuple(pubsub.router for pubsub in pubsubs_gsub) index_alice = 0 id_alice = pubsubs_gsub[index_alice].my_id index_bob = 1 id_bob = pubsubs_gsub[index_bob].my_id topic = "test_handle_prune" for pubsub in pubsubs_gsub: await pubsub.subscribe(topic) await connect(pubsubs_gsub[index_alice].host, pubsubs_gsub[index_bob].host) # Wait 3 seconds for heartbeat to allow mesh to connect await trio.sleep(3) # Check that they are each other's mesh peer assert id_alice in gossipsubs[index_bob].mesh[topic] assert id_bob in gossipsubs[index_alice].mesh[topic] # alice emit prune message to bob, alice should be removed # from bob's mesh peer await gossipsubs[index_alice].emit_prune(topic, id_bob) # FIXME: This test currently works because the heartbeat interval # is increased to 3 seconds, so alice won't get add back into # bob's mesh peer during heartbeat. await trio.sleep(1) # Check that alice is no longer bob's mesh peer assert id_alice not in gossipsubs[index_bob].mesh[topic] assert id_bob in gossipsubs[index_alice].mesh[topic] @pytest.mark.trio async def test_dense(): async with PubsubFactory.create_batch_with_gossipsub(10) as pubsubs_gsub: hosts = [pubsub.host for pubsub in pubsubs_gsub] num_msgs = 5 # All pubsub subscribe to foobar queues = [await pubsub.subscribe("foobar") for pubsub in pubsubs_gsub] # Densely connect libp2p hosts in a random way await dense_connect(hosts) # Wait 2 seconds for heartbeat to allow mesh to connect await trio.sleep(2) for i in range(num_msgs): msg_content = b"foo " + i.to_bytes(1, "big") # randomly pick a message origin origin_idx = random.randint(0, len(hosts) - 1) # publish from the randomly chosen host await pubsubs_gsub[origin_idx].publish("foobar", msg_content) await trio.sleep(0.5) # Assert that all blocking queues receive the message for queue in queues: msg = await queue.receive() assert msg.data == msg_content @pytest.mark.trio async def test_fanout(): async with PubsubFactory.create_batch_with_gossipsub(10) as pubsubs_gsub: hosts = [pubsub.host for pubsub in pubsubs_gsub] num_msgs = 5 # All pubsub subscribe to foobar except for `pubsubs_gsub[0]` subs = [await pubsub.subscribe("foobar") for pubsub in pubsubs_gsub[1:]] # Sparsely connect libp2p hosts in random way await dense_connect(hosts) # Wait 2 seconds for heartbeat to allow mesh to connect await trio.sleep(2) topic = "foobar" # Send messages with origin not subscribed for i in range(num_msgs): msg_content = b"foo " + i.to_bytes(1, "big") # Pick the message origin to the node that is not subscribed to 'foobar' origin_idx = 0 # publish from the randomly chosen host await pubsubs_gsub[origin_idx].publish(topic, msg_content) await trio.sleep(0.5) # Assert that all blocking queues receive the message for sub in subs: msg = await sub.receive() assert msg.data == msg_content # Subscribe message origin subs.insert(0, await pubsubs_gsub[0].subscribe(topic)) # Send messages again for i in range(num_msgs): msg_content = b"bar " + i.to_bytes(1, "big") # Pick the message origin to the node that is not subscribed to 'foobar' origin_idx = 0 # publish from the randomly chosen host await pubsubs_gsub[origin_idx].publish(topic, msg_content) await trio.sleep(0.5) # Assert that all blocking queues receive the message for sub in subs: msg = await sub.receive() assert msg.data == msg_content @pytest.mark.trio @pytest.mark.slow async def test_fanout_maintenance(): async with PubsubFactory.create_batch_with_gossipsub(10) as pubsubs_gsub: hosts = [pubsub.host for pubsub in pubsubs_gsub] num_msgs = 5 # All pubsub subscribe to foobar queues = [] topic = "foobar" for i in range(1, len(pubsubs_gsub)): q = await pubsubs_gsub[i].subscribe(topic) # Add each blocking queue to an array of blocking queues queues.append(q) # Sparsely connect libp2p hosts in random way await dense_connect(hosts) # Wait 2 seconds for heartbeat to allow mesh to connect await trio.sleep(2) # Send messages with origin not subscribed for i in range(num_msgs): msg_content = b"foo " + i.to_bytes(1, "big") # Pick the message origin to the node that is not subscribed to 'foobar' origin_idx = 0 # publish from the randomly chosen host await pubsubs_gsub[origin_idx].publish(topic, msg_content) await trio.sleep(0.5) # Assert that all blocking queues receive the message for queue in queues: msg = await queue.receive() assert msg.data == msg_content for sub in pubsubs_gsub: await sub.unsubscribe(topic) queues = [] await trio.sleep(2) # Resub and repeat for i in range(1, len(pubsubs_gsub)): q = await pubsubs_gsub[i].subscribe(topic) # Add each blocking queue to an array of blocking queues queues.append(q) await trio.sleep(2) # Check messages can still be sent for i in range(num_msgs): msg_content = b"bar " + i.to_bytes(1, "big") # Pick the message origin to the node that is not subscribed to 'foobar' origin_idx = 0 # publish from the randomly chosen host await pubsubs_gsub[origin_idx].publish(topic, msg_content) await trio.sleep(0.5) # Assert that all blocking queues receive the message for queue in queues: msg = await queue.receive() assert msg.data == msg_content @pytest.mark.trio async def test_gossip_propagation(): async with PubsubFactory.create_batch_with_gossipsub( 2, degree=1, degree_low=0, degree_high=2, gossip_window=50, gossip_history=100 ) as pubsubs_gsub: topic = "foo" await pubsubs_gsub[0].subscribe(topic) # node 0 publish to topic msg_content = b"foo_msg" # publish from the randomly chosen host await pubsubs_gsub[0].publish(topic, msg_content) # now node 1 subscribes queue_1 = await pubsubs_gsub[1].subscribe(topic) await connect(pubsubs_gsub[0].host, pubsubs_gsub[1].host) # wait for gossip heartbeat await trio.sleep(2) # should be able to read message msg = await queue_1.receive() assert msg.data == msg_content