py-libp2p/libp2p/stream_muxer/mplex/mplex.py

260 lines
9.8 KiB
Python
Raw Normal View History

import asyncio
from typing import Any # noqa: F401
from typing import Dict, List, Optional, Tuple
2019-08-02 18:28:04 +08:00
from libp2p.peer.id import ID
2019-08-05 11:22:44 +08:00
from libp2p.security.secure_conn_interface import ISecureConn
from libp2p.stream_muxer.abc import IMuxedConn, IMuxedStream
from libp2p.typing import TProtocol
from libp2p.utils import (
decode_uvarint_from_stream,
encode_uvarint,
encode_varint_prefixed,
read_varint_prefixed_bytes,
)
2019-01-10 02:38:56 +08:00
2019-08-02 17:14:43 +08:00
from .constants import HeaderTags
2019-08-28 21:43:34 +08:00
from .datastructures import StreamID
2019-08-03 13:36:19 +08:00
from .mplex_stream import MplexStream
2018-11-01 05:31:00 +08:00
MPLEX_PROTOCOL_ID = TProtocol("/mplex/6.7.0")
2019-08-02 17:53:51 +08:00
2018-11-21 09:28:41 +08:00
class Mplex(IMuxedConn):
2018-11-01 05:31:00 +08:00
"""
reference: https://github.com/libp2p/go-mplex/blob/master/multiplex.go
"""
secured_conn: ISecureConn
peer_id: ID
# TODO: `dataIn` in go implementation. Should be size of 8.
# TODO: Also, `dataIn` is closed indicating EOF in Go. We don't have similar strategies
# to let the `MplexStream`s know that EOF arrived (#235).
2019-08-28 21:43:34 +08:00
next_channel_id: int
2019-09-05 22:29:33 +08:00
streams: Dict[StreamID, MplexStream]
streams_lock: asyncio.Lock
new_stream_queue: "asyncio.Queue[IMuxedStream]"
shutdown: asyncio.Event
2019-08-02 17:53:51 +08:00
_tasks: List["asyncio.Future[Any]"]
# TODO: `generic_protocol_handler` should be refactored out of mplex conn.
def __init__(self, secured_conn: ISecureConn, peer_id: ID) -> None:
2018-11-01 05:31:00 +08:00
"""
create a new muxed connection
:param secured_conn: an instance of ``ISecureConn``
:param generic_protocol_handler: generic protocol handler
for new muxed streams
[WIP] PubSub and FloodSub development (#133) * Add notifee interface * Add notify function to network interface * Implement notify feature * Add tests for notify * Make notifee functions all async * Fix linting issue * Fix linting issue * Scaffold pubsub router interface * Scaffold pubsub directory * Store peer_id in muxed connection * Implement pubsub notifee * Remove outdated files * Implement pubsub first attempt * Prepare pubsub for floodsub * Add mplex conn to net stream and add conn in notify tests * Implement floodsub * Use NetStream in generic protocol handler * Debugging async issues * Modify test to perform proper assert. Test passes * Remove callbacks. Reduce sleep time * Add simple three node test * Clean up code. Add message classes * Add test for two topics * Add conn to net stream and conn tests * Refactor test setup to remove duplicate code * Fix linting issues * Fix linting issue * Fix linting issue * Fix outstanding unrelated lint issue in multiselect_client * Add connect function * Remove debug prints * Remove debug prints from floodsub * Use MessageTalk in place of direct message breakdown * Remove extra prints * Remove outdated function * Add message to queues for all topics in message * Debugging * Add message self delivery * Increase read timeout to 5 to get pubsub tests passing * Refactor testing helper func. Add tests * Add tests and increase timeout to get tests passing * Add dummy account demo scaffolding * Attempt to use threads. Test fails * Implement basic dummy node tests using threads * Add generic testing function * Add simple seven node tree test * Add more complex seven node tree tests * Add five node ring tests * Remove unnecessary get_message_type func * Add documentation to classes * Add message id to messages * Add documentation to test helper func * Add docs to dummy account node helper func * Add more docs to dummy account node test helper func * fixed linting errors in floodsub * small notify bugfix * move pubsub into libp2p * fixed pubsub linting * fixing pubsub test failures * linting
2019-03-24 01:52:02 +08:00
:param peer_id: peer_id of peer the connection is to
2018-11-01 05:31:00 +08:00
"""
self.secured_conn = secured_conn
2018-11-26 00:05:56 +08:00
2019-08-28 21:43:34 +08:00
self.next_channel_id = 0
[WIP] PubSub and FloodSub development (#133) * Add notifee interface * Add notify function to network interface * Implement notify feature * Add tests for notify * Make notifee functions all async * Fix linting issue * Fix linting issue * Scaffold pubsub router interface * Scaffold pubsub directory * Store peer_id in muxed connection * Implement pubsub notifee * Remove outdated files * Implement pubsub first attempt * Prepare pubsub for floodsub * Add mplex conn to net stream and add conn in notify tests * Implement floodsub * Use NetStream in generic protocol handler * Debugging async issues * Modify test to perform proper assert. Test passes * Remove callbacks. Reduce sleep time * Add simple three node test * Clean up code. Add message classes * Add test for two topics * Add conn to net stream and conn tests * Refactor test setup to remove duplicate code * Fix linting issues * Fix linting issue * Fix linting issue * Fix outstanding unrelated lint issue in multiselect_client * Add connect function * Remove debug prints * Remove debug prints from floodsub * Use MessageTalk in place of direct message breakdown * Remove extra prints * Remove outdated function * Add message to queues for all topics in message * Debugging * Add message self delivery * Increase read timeout to 5 to get pubsub tests passing * Refactor testing helper func. Add tests * Add tests and increase timeout to get tests passing * Add dummy account demo scaffolding * Attempt to use threads. Test fails * Implement basic dummy node tests using threads * Add generic testing function * Add simple seven node tree test * Add more complex seven node tree tests * Add five node ring tests * Remove unnecessary get_message_type func * Add documentation to classes * Add message id to messages * Add documentation to test helper func * Add docs to dummy account node helper func * Add more docs to dummy account node test helper func * fixed linting errors in floodsub * small notify bugfix * move pubsub into libp2p * fixed pubsub linting * fixing pubsub test failures * linting
2019-03-24 01:52:02 +08:00
# Set peer_id
self.peer_id = peer_id
2018-11-26 00:05:56 +08:00
# Mapping from stream ID -> buffer of messages for that stream
2019-09-05 22:29:33 +08:00
self.streams = {}
self.streams_lock = asyncio.Lock()
self.new_stream_queue = asyncio.Queue()
self.shutdown = asyncio.Event()
2018-11-26 00:05:56 +08:00
self._tasks = []
# Kick off reading
self._tasks.append(asyncio.ensure_future(self.handle_incoming()))
2018-11-01 05:31:00 +08:00
@property
def initiator(self) -> bool:
return self.secured_conn.initiator
async def close(self) -> None:
2018-11-01 05:31:00 +08:00
"""
close the stream muxer and underlying secured connection
2018-11-01 05:31:00 +08:00
"""
for task in self._tasks:
task.cancel()
await self.secured_conn.close()
2018-11-01 05:31:00 +08:00
2019-08-02 17:53:51 +08:00
def is_closed(self) -> bool:
2018-11-01 05:31:00 +08:00
"""
check connection is fully closed
:return: true if successful
"""
2019-08-02 17:53:51 +08:00
raise NotImplementedError()
2018-11-01 05:31:00 +08:00
2019-08-28 21:43:34 +08:00
def _get_next_channel_id(self) -> int:
"""
Get next available stream id
:return: next available stream id for the connection
"""
2019-08-28 21:43:34 +08:00
next_id = self.next_channel_id
self.next_channel_id += 1
return next_id
2019-09-05 22:29:33 +08:00
async def _initialize_stream(self, stream_id: StreamID, name: str) -> MplexStream:
stream = MplexStream(name, stream_id, self)
2019-09-05 22:29:33 +08:00
async with self.streams_lock:
self.streams[stream_id] = stream
2019-09-05 22:29:33 +08:00
return stream
async def open_stream(self) -> IMuxedStream:
2018-11-01 05:31:00 +08:00
"""
creates a new muxed_stream
:return: a new ``MplexStream``
2018-11-01 05:31:00 +08:00
"""
2019-08-28 21:43:34 +08:00
channel_id = self._get_next_channel_id()
stream_id = StreamID(channel_id=channel_id, is_initiator=True)
# Default stream name is the `channel_id`
2019-09-05 22:29:33 +08:00
name = str(channel_id)
stream = await self._initialize_stream(stream_id, name)
await self.send_message(HeaderTags.NewStream, name.encode(), stream_id)
return stream
async def accept_stream(self) -> IMuxedStream:
2018-11-01 05:31:00 +08:00
"""
accepts a muxed stream opened by the other end
"""
return await self.new_stream_queue.get()
2019-08-28 21:43:34 +08:00
async def send_message(
self, flag: HeaderTags, data: Optional[bytes], stream_id: StreamID
2019-08-28 21:43:34 +08:00
) -> int:
"""
sends a message over the connection
:param header: header to use
:param data: data to send in the message
2018-11-12 06:38:11 +08:00
:param stream_id: stream the message is in
"""
2018-11-12 06:38:11 +08:00
# << by 3, then or with flag
header = encode_uvarint((stream_id.channel_id << 3) | flag.value)
2018-11-19 00:22:17 +08:00
if data is None:
data = b""
_bytes = header + encode_varint_prefixed(data)
2018-11-13 00:00:43 +08:00
return await self.write_to_stream(_bytes)
2018-11-12 06:38:11 +08:00
async def write_to_stream(self, _bytes: bytes) -> int:
"""
writes a byte array to a secured connection
:param _bytes: byte array to write
:return: length written
"""
await self.secured_conn.write(_bytes)
2018-11-12 06:48:31 +08:00
return len(_bytes)
2019-08-02 17:53:51 +08:00
async def handle_incoming(self) -> None:
"""
Read a message off of the secured connection and add it to the corresponding message buffer
"""
# TODO Deal with other types of messages using flag (currently _)
2018-11-13 00:00:43 +08:00
while True:
2019-08-28 21:43:34 +08:00
channel_id, flag, message = await self.read_message()
if channel_id is not None and flag is not None and message is not None:
stream_id = StreamID(channel_id=channel_id, is_initiator=bool(flag & 1))
2019-09-05 22:29:33 +08:00
is_stream_id_seen: bool
stream: MplexStream
2019-09-05 22:29:33 +08:00
async with self.streams_lock:
is_stream_id_seen = stream_id in self.streams
if is_stream_id_seen:
stream = self.streams[stream_id]
2019-09-05 22:29:33 +08:00
# Other consequent stream message should wait until the stream get accepted
# TODO: Handle more tags, and refactor `HeaderTags`
2019-08-02 17:14:43 +08:00
if flag == HeaderTags.NewStream.value:
2019-09-05 22:29:33 +08:00
if is_stream_id_seen:
# `NewStream` for the same id is received twice...
2019-09-05 23:44:22 +08:00
# TODO: Shutdown
2019-09-05 22:29:33 +08:00
pass
mplex_stream = await self._initialize_stream(
stream_id, message.decode()
)
# TODO: Check if `self` is shutdown.
await self.new_stream_queue.put(mplex_stream)
elif flag in (
HeaderTags.MessageInitiator.value,
HeaderTags.MessageReceiver.value,
):
2019-09-05 22:29:33 +08:00
if not is_stream_id_seen:
# We receive a message of the stream `stream_id` which is not accepted
# before. It is abnormal. Possibly disconnect?
# TODO: Warn and emit logs about this.
continue
2019-09-09 15:45:35 +08:00
async with stream.close_lock:
if stream.event_remote_closed.is_set():
# TODO: Warn "Received data from remote after stream was closed by them. (len = %d)" # noqa: E501
continue
2019-09-05 22:29:33 +08:00
await stream.incoming_data.put(message)
2019-09-05 23:44:22 +08:00
elif flag in (
HeaderTags.CloseInitiator.value,
HeaderTags.CloseReceiver.value,
):
if not is_stream_id_seen:
continue
# NOTE: If remote is already closed, then return: Technically a bug
# on the other side. We should consider killing the connection.
async with stream.close_lock:
if stream.event_remote_closed.is_set():
continue
2019-09-05 23:44:22 +08:00
is_local_closed: bool
async with stream.close_lock:
stream.event_remote_closed.set()
is_local_closed = stream.event_local_closed.is_set()
# If local is also closed, both sides are closed. Then, we should clean up
# the entry of this stream, to avoid others from accessing it.
2019-09-05 23:44:22 +08:00
if is_local_closed:
async with self.streams_lock:
del self.streams[stream_id]
elif flag in (
HeaderTags.ResetInitiator.value,
HeaderTags.ResetReceiver.value,
):
if not is_stream_id_seen:
# This is *ok*. We forget the stream on reset.
continue
async with stream.close_lock:
if not stream.event_remote_closed.is_set():
# TODO: Why? Only if remote is not closed before then reset.
2019-09-05 23:44:22 +08:00
stream.event_reset.set()
2019-09-05 23:44:22 +08:00
stream.event_remote_closed.set()
# If local is not closed, we should close it.
2019-09-05 23:44:22 +08:00
if not stream.event_local_closed.is_set():
stream.event_local_closed.set()
2019-09-05 23:44:22 +08:00
async with self.streams_lock:
del self.streams[stream_id]
else:
# TODO: logging
if is_stream_id_seen:
await stream.reset()
# Force context switch
await asyncio.sleep(0)
2019-08-02 17:53:51 +08:00
async def read_message(self) -> Tuple[int, int, bytes]:
"""
Read a single message off of the secured connection
:return: stream_id, flag, message contents
"""
# FIXME: No timeout is used in Go implementation.
# Timeout is set to a relatively small value to alleviate wait time to exit
# loop in handle_incoming
header = await decode_uvarint_from_stream(self.secured_conn)
# TODO: Handle the case of EOF and other exceptions?
try:
message = await asyncio.wait_for(
read_varint_prefixed_bytes(self.secured_conn), timeout=5
)
except asyncio.TimeoutError:
# TODO: Investigate what we should do if time is out.
return None, None, None
flag = header & 0x07
2019-08-28 21:43:34 +08:00
channel_id = header >> 3
2019-08-28 21:43:34 +08:00
return channel_id, flag, message