Source code for autopilot.core.networking

"""
Classes for network communication.

There are two general types of network objects -

* :class:`.Station` and its children are independent processes that should only be instantiated once
    per piece of hardware. They are used to distribute messages between :class:`.Net_Node` s,
    forward messages up the networking tree, and responding to messages that don't need any input from
    the :class:`~.pilot.Pilot` or :class:`~.terminal.Terminal`.
* :class:`.Net_Node` is a pop-in networking class that can be given to any other object that
    wants to send or receive messages.
"""


import json
import logging
import threading
import zmq
import sys
import datetime
import time
import os
import multiprocessing
import base64
import socket
import struct
import blosc
from copy import copy
from tornado.ioloop import IOLoop
from zmq.eventloop.zmqstream import ZMQStream
from itertools import count
import numpy as np
import pdb
#from pudb.remote import set_trace
if sys.version_info >= (3,0):
    import queue
else:
    import Queue as queue

from autopilot import prefs
from autopilot.core.loggers import init_logger

# TODO: Periodically ping pis to check that they are still responsive

[docs]class Station(multiprocessing.Process): """ Independent networking class used for messaging between computers. These objects send and handle :class:`.networking.Message` s by using a dictionary of :attr:`~.networking.Station.listens`, or methods that are called to respond to different types of messages. Each sent message is given an ID, and a thread is spawned to periodically resend it (up until some time-to-live, typically 5 times) until confirmation is received. By default, the only listen these objects have is :meth:`.l_confirm`, which responds to message confirmations. Accordingly, `listens` should be added by using :meth:`dict.update` rather than reassigning the attribute. Station objects can be made with or without a :attr:`~.networking.Station.pusher`, a :class:`zmq.DEALER` socket that connects to the :class:`zmq.ROUTER` socket of an upstream Station object. This class should not be instantiated on its own, but should instead be subclassed in order to provide methods used by :meth:`~.Station.handle_listen`. Attributes: ctx (:class:`zmq.Context`): zeromq context loop (:class:`tornado.ioloop.IOLoop`): a tornado ioloop pusher (:class:`zmq.Socket`): pusher socket - a dealer socket that connects to other routers push_ip (str): If we have a dealer, IP to push messages to push_port (str): If we have a dealer, port to push messages to push_id (str): :attr:`~zmq.Socket.identity` of the Router we push to listener (:class:`zmq.Socket`): The main router socket to send/recv messages listen_port (str): Port our router listens on logger (:class:`logging.Logger`): Used to log messages and network events. id (str): What are we known as? What do we set our :attr:`~zmq.Socket.identity` as? ip (str): Device IP listens (dict): Dictionary of functions to call for different types of messages. keys match the :attr:`.Message.key`. senders (dict): Identities of other sockets (keys, ie. directly connected) and their state (values) if they keep one outbox (dict): Messages that have been sent but have not been confirmed timers (dict): dict of :class:`threading.Timer` s that will check in on outbox messages msg_counter (:class:`itertools.count`): counter to index our sent messages file_block (:class:`threading.Event`): Event to signal when a file is being received. """ ctx = None # Context loop = None # IOLoop push_ip = None # IP to push to push_port = None # Publisher Port push_id = "" # Identity of the Router we push to listen_port = None # Listener Port pusher = None # pusher socket - a dealer socket that connects to other routers listener = None # Listener socket - a router socket to send/recv messages logger = None # Logger.... id = None # What are we known as? ip = None # whatismy listens = {} # Dictionary of functions to call for different types of messages senders = {} # who has sent us stuff (ie. directly connected) and their state if they keep one push_outbox = {} # Messages that are out with unconfirmed delivery send_outbox = {} # Messages that are out with unconfirmed delivery timers = {} # dict of timer threads that will check in on outbox messages child = False routes = {} # dict of 'to' addressee and the route that should be taken to reach them repeat_interval = 5.0 # seconds to wait before retrying messages def __init__(self): super(Station, self).__init__() # Prefs should be passed by the terminal, if not, try to load from default locatio try: self.ip = self.get_ip() except Exception as e: Warning("Couldn't get IP!: {}".format(e)) self.ip = "" # Setup logging self.logger = init_logger(self) self.file_block = threading.Event() # to wait for file transfer # number messages as we send them self.msg_counter = count() # we have a few builtin listens self.listens = { 'CONFIRM': self.l_confirm, 'STREAM' : self.l_stream } # even tthat signals when we are closing self.closing = threading.Event() self.closing.clear() # start thread that periodically resends messages self.repeat_thread = threading.Thread(target=self.repeat) self.repeat_thread.setDaemon(True) self.repeat_thread.start() def __del__(self): self.closing.set() # Stopping the loop should kill the process, as it's what's holding us in run() self.loop.stop()
[docs] def run(self): """ A :class:`zmq.Context` and :class:`tornado.IOLoop` are spawned, the listener and optionally the pusher are instantiated and connected to :meth:`~.Station.handle_listen` using :meth:`~zmq.eventloop.zmqstream.ZMQStream.on_recv` . The process is kept open by the :class:`tornado.IOLoop` . """ try: # init zmq objects self.context = zmq.Context() self.loop = IOLoop() # Our networking topology is treelike: # each Station object binds one Router to # send and receive messages from its descendants # each Station object may have one Dealer that # connects it with its antecedents. self.listener = self.context.socket(zmq.ROUTER) #self.listener.identity = self.id.encode('utf-8') #self.listener.identity = self.id self.listener.setsockopt_string(zmq.IDENTITY, self.id) self.listener.bind('tcp://*:{}'.format(self.listen_port)) self.listener = ZMQStream(self.listener, self.loop) self.listener.on_recv(self.handle_listen) if self.pusher is True: self.pusher = self.context.socket(zmq.DEALER) #self.pusher.identity = self.id.encode('utf-8') #self.pusher.identity = self.id self.pusher.setsockopt_string(zmq.IDENTITY, self.id) self.pusher.connect('tcp://{}:{}'.format(self.push_ip, self.push_port)) self.pusher = ZMQStream(self.pusher, self.loop) self.pusher.on_recv(self.handle_listen) # TODO: Make sure handle_listen knows how to handle ID-less messages self.logger.info('Starting IOLoop') self.loop.start() except KeyboardInterrupt: # normal quitting behavior pass finally: self.release()
[docs] def prepare_message(self, to, key, value, repeat=True, flags=None): """ If a message originates with us, a :class:`.Message` class is instantiated, given an ID and the rest of its attributes. Args: flags: repeat: to (str): The identity of the socket this message is to key (str): The type of message - used to select which method the receiver uses to process this message. value: Any information this message should contain. Can be any type, but must be JSON serializable. """ msg = Message() msg.sender = self.id try: msg.to = to.decode('utf-8') except AttributeError: msg.to = to try: msg.key = key.decode('utf-8') except AttributeError: msg.key = key msg.value = value msg_num = next(self.msg_counter) msg.id = "{}_{}".format(self.id, msg_num) if not repeat: msg.flags['NOREPEAT'] = True if flags: for k, v in flags.items(): msg.flags[k] = v return msg
[docs] def send(self, to=None, key=None, value=None, msg=None, repeat=True, flags=None): """ Send a message via our :attr:`~.Station.listener` , ROUTER socket. Either an already created :class:`.Message` should be passed as `msg`, or at least `to` and `key` must be provided for a new message created by :meth:`~.Station.prepare_message` . A :class:`threading.Timer` is created to resend the message using :meth:`~.Station.repeat` unless `repeat` is False. Args: flags: to (str): The identity of the socket this message is to key (str): The type of message - used to select which method the receiver uses to process this message. value: Any information this message should contain. Can be any type, but must be JSON serializable. msg (`.Message`): An already created message. repeat (bool): Should this message be resent if confirmation is not received? """ if not msg and not all([to, key]): self.logger.exception('Need either a message or \'to\' and \'key\' fields.\ Got\nto: {}\nkey: {}\nvalue: {}\nmsg: {}'.format(to, key, value, msg)) return manual_to = False if not msg: # we're sending this ourselves, new message. msg = self.prepare_message(to, key, value, repeat, flags) elif to: # if given both message and to, send it to our 'to' # don't want to force a reserialization of the message manual_to = True if 'NOREPEAT' in msg.flags.keys(): repeat = False # if we didn't send it, we shouldn't double-confirm it. if msg.sender not in [self.name, '_'+self.name]: repeat = False # Make sure our message has everything if not msg.validate(): self.logger.error('Message Invalid:\n{}'.format(str(msg))) # encode message msg_enc = msg.serialize() # TODO: try/except here if not msg_enc: #set_trace(term_size=(80,40)) self.logger.error('Message could not be encoded:\n{}'.format(str(msg))) return if manual_to: self.listener.send_multipart([to.encode('utf-8'), msg_enc]) else: if isinstance(msg.to, list): self.listener.send_multipart([msg.to[0].encode('utf-8'), msg_enc]) else: self.listener.send_multipart([msg.to.encode('utf-8'), msg_enc]) # messages can have a flag that says not to log # log_this = True # if 'NOLOG' in msg.flags.keys(): # log_this = False #if msg.key != "CONFIRM": self.logger.debug('MESSAGE SENT - {}'.format(str(msg))) if repeat and not msg.key == "CONFIRM": # add to outbox and spawn timer to resend self.send_outbox[msg.id] = (time.time(), msg)
# self.timers[msg.id] = threading.Timer(5.0, self.repeat, args=(msg.id,'send')) # self.timers[msg.id].start() #self.outbox.put((msg.id))
[docs] def push(self, to=None, key = None, value = None, msg=None, repeat=True, flags=None): """ Send a message via our :attr:`~.Station.pusher` , DEALER socket. Unlike :meth:`~.Station.send` , `to` is not required. Every message is always sent to :attr:`~.Station.push_id` . `to` can be included to send a message further up the network tree to a networking object we're not directly connected to. Either an already created :class:`.Message` should be passed as `msg`, or at least `key` must be provided for a new message created by :meth:`~.Station.prepare_message` . A :class:`threading.Timer` is created to resend the message using :meth:`~.Station.repeat` unless `repeat` is False. Args: flags: to (str): The identity of the socket this message is to. If not included, sent to :meth:`~.Station.push_id` . key (str): The type of message - used to select which method the receiver uses to process this message. value: Any information this message should contain. Can be any type, but must be JSON serializable. msg (`.Message`): An already created message. repeat (bool): Should this message be resent if confirmation is not received? """ # send message via the dealer # even though we only have one connection over our dealer, # we still include 'to' in case we are sending further upstream # but can push without 'to', just fill in with upstream id if not msg and not key: self.logger.exception('Need either a message or a \'key\' field.\ Got\nto: {}\nkey: {}\nvalue: {}\nmsg: {}'.format(to, key, value, msg)) if not msg: if to is None: to = self.push_id msg = self.prepare_message(to, key, value, repeat, flags) if 'NOREPEAT' in msg.flags.keys(): repeat = False log_this = True if 'NOLOG' in msg.flags.keys(): log_this = False # Make sure our message has everything if not msg.validate(): self.logger.error('Message Invalid:\n{}'.format(str(msg))) # encode message msg_enc = msg.serialize() if not msg_enc: self.logger.error('Message could not be encoded:\n{}'.format(str(msg))) return # Even if the message is not to our upstream node, we still send it # upstream because presumably our target is upstream. self.pusher.send_multipart([self.push_id, bytes(msg.to, encoding="utf-8"), msg_enc]) if not (msg.key == "CONFIRM") and log_this: self.logger.debug('MESSAGE PUSHED - {}'.format(str(msg))) if repeat and not msg.key == 'CONFIRM': # add to outbox and spawn timer to resend self.push_outbox[msg.id] = (time.time(), msg)
#self.timers[msg.id] = threading.Timer(5.0, self.repeat, args=(msg.id, 'push')) #self.timers[msg.id].start()
[docs] def repeat(self): """ Periodically (according to :attr:`~.repeat_interval`) resend messages that haven't been confirmed TTL is decremented, and messages are resent until their TTL is 0. """ while not self.closing.is_set(): # make local copies push_outbox = copy(self.push_outbox) send_outbox = copy(self.send_outbox) # try to send any outstanding messages and delete if too old if len(push_outbox)>0: for id in push_outbox.keys(): if push_outbox[id][1].ttl <= 0: self.logger.warning('PUBLISH FAILED {} - {}'.format(id, str(push_outbox[id][1]))) try: del self.push_outbox[id] except KeyError: # fine, already deleted pass else: # if we didn't just put this message in our outbox if (time.time() - push_outbox[id][0]) > self.repeat_interval*2: self.logger.debug('REPUBLISH {} - {}'.format(id, str(push_outbox[id][1]))) self.pusher.send_multipart([self.push_id, push_outbox[id][1].serialize()]) self.push_outbox[id][1].ttl -= 1 if len(send_outbox)>0: for id in send_outbox.keys(): if send_outbox[id][1].ttl <= 0: self.logger.warning('PUBLISH FAILED {} - {}'.format(id, str(send_outbox[id][1]))) try: del self.send_outbox[id] except KeyError: # fine, already deleted pass else: # if we didn't just put this message in our outbox if (time.time() - send_outbox[id][0]) > self.repeat_interval*2: self.logger.debug('REPUBLISH {} - {}'.format(id, str(send_outbox[id][1]))) self.listener.send_multipart([bytes(send_outbox[id][1].to, encoding="utf-8"), send_outbox[id][1].serialize()]) self.send_outbox[id][1].ttl -= 1 # wait to do it again time.sleep(self.repeat_interval)
[docs] def l_confirm(self, msg): """ Confirm that a message was received. Args: msg (:class:`.Message`): A confirmation message - note that this message has its own unique ID, so the value of this message contains the ID of the message that is being confirmed """ # confirmation that a published message was received # value should be the message id # delete message from outbox if we still have it try: if msg.value in self.send_outbox.keys(): del self.send_outbox[msg.value] elif msg.value in self.push_outbox.keys(): del self.push_outbox[msg.value] except KeyError: # fine, already deleted pass # if this is a message to our internal net_node, make sure it gets the memo that shit was confirmed too if msg.to == "_{}".format(self.id): self.send("_{}".format(self.id), 'CONFIRM', msg.value)
#self.logger.info('CONFIRMED MESSAGE {}'.format(msg.value))
[docs] def l_stream(self, msg): """ Reconstitute the original stream of messages and call their handling methods The ``msg`` should contain an ``inner_key`` that indicates the key, and thus the handling method. Args: msg (dict): Compressed stream sent by :meth:`Net_Node._stream` """ listen_fn = self.listens[msg.value['inner_key']] old_value = copy(msg.value) delattr(msg, 'value') msg.key = old_value['inner_key'] for v in old_value['payload']: if isinstance(v, dict) and ('headers' in old_value.keys()): v.update(old_value['headers']) msg.value = v listen_fn(msg)
[docs] def handle_listen(self, msg): """ Upon receiving a message, call the appropriate listen method in a new thread. If the message is :attr:`~.Message.to` us, send confirmation. If the message is not :attr:`~.Message.to` us, attempt to forward it. Args: msg (str): JSON :meth:`.Message.serialize` d message. """ # TODO: This check is v. fragile, pyzmq has a way of sending the stream along with the message #####################33 # Parse the message if len(msg)==1: # from our dealer, these are always to us. send_type = 'dealer' #msg = json.loads(msg[0]) #msg = Message(**msg) msg = Message(msg[0]) elif len(msg)>=2: # from the router send_type = 'router' sender = msg[0] # if this message was a multihop message, store the route if len(msg)>4: self.routes[sender] = msg[0:-3] # # if this is a new sender, add them to the list if sender not in self.senders.keys(): self.senders[sender] = "" self.senders[b'_' + sender] = '' # connection pings are blank frames, # respond to let them know we're alive if msg[-1] == b'': self.listener.send_multipart(msg) return # if this message wasn't to us, forward without deserializing # the second to last should always be the intended recipient unserialized_to = msg[-2] if unserialized_to.decode('utf-8') not in [self.id, "_{}".format(self.id)]: if unserialized_to not in self.senders.keys() and self.pusher: # if we don't know who they are and we have a pusher, try to push it self.pusher.send_multipart([self.push_id, unserialized_to, msg[-1]]) else: #if we know who they are or not, try to send it through router anyway. self.listener.send_multipart([unserialized_to, unserialized_to, msg[-1]]) # self.logger.debug('FORWARDING: to - {}, {}'.format(unserialized_to, msg[-1][:100] if len(msg[-1])>100 else msg[-1])) return #msg = json.loads(msg[-1]) #msg = Message(**msg) #set_trace(term_size=(80, 24)) msg = Message(msg[-1]) # if this is a new sender, add them to the list if msg['sender'] not in self.senders.keys(): self.senders[msg['sender']] = "" self.senders['_' + msg['sender']] = '' else: self.logger.error('Dont know what this message is:{}'.format(msg)) return # Check if our listen was sent properly # if not msg.validate(): # self.logger.error('Message failed to validate:\n{}'.format(str(msg))) # return ################################### # Handle the message # if this message has a multihop 'to' field, forward it along # some messages have a flag not to log them # log_this = True # if 'NOLOG' in msg.flags.keys(): # log_this = False if isinstance(msg.to, list): if len(msg.to) == 1: msg.to = msg.to[0] if isinstance(msg.to, list): # pop ourselves off the list _ = msg.to.pop(0) # if the next recipient in the list is our push-parent, push it if msg.to[0] == self.push_id: self.push(msg=msg) else: self.send(msg=msg) # if this message is to us, just handle it and return elif msg.to in [self.id, "_{}".format(self.id)]: if (msg.key != "CONFIRM"): self.logger.debug('RECEIVED: {}'.format(str(msg))) # Log and spawn thread to respond to listen try: listen_funk = self.listens[msg.key] listen_thread = threading.Thread(target=listen_funk, args=(msg,)) listen_thread.start() except KeyError: self.logger.exception('ERROR: No function could be found for msg id {} with key: {}'.format(msg.id, msg.key)) # send a return message that confirms even if we except # don't confirm confirmations if (msg.key != "CONFIRM") and ('NOREPEAT' not in msg.flags.keys()): if send_type == 'router': self.send(msg.sender, 'CONFIRM', msg.id) elif send_type == 'dealer': self.push(msg.sender, 'CONFIRM', msg.id) return # otherwise, if it's to someone we know about, send it there elif self.child and (msg.to == 'T'): # FIXME UGLY HACK self.push(msg=msg) elif msg.to in self.senders.keys(): self.send(msg=msg) # otherwise, if we have a pusher, send it there # it's either for them or some other upstream node we don't know about elif self.pusher: self.push(msg=msg) else: self.logger.warning('Message to unconfirmed recipient, attempting to send: {}'.format(str(msg))) self.send(msg=msg) # finally, if there's something we're supposed to do, do it # even if the message is not to us, # sometimes we do work en passant to reduce effort doubling # FIXME Seems like a really bad idea. if msg.key in self.listens.keys(): listen_funk = self.listens[msg.key] listen_thread = threading.Thread(target=listen_funk, args=(msg,)) listen_thread.start() # since we return if it's to us before, confirm is repeated down here. # FIXME: Inelegant if (msg.key != "CONFIRM") and ('NOREPEAT' not in msg.flags.keys()): if send_type == 'router': self.send(sender, 'CONFIRM', msg.id) elif send_type == 'dealer': self.push(msg.sender, 'CONFIRM', msg.id)
[docs] def get_ip(self): """ Find our IP address returns (str): our IPv4 address. """ # shamelessly stolen from https://www.w3resource.com/python-exercises/python-basic-exercise-55.php # variables are badly named because this is just a rough unwrapping of what was a monstrous one-liner # (and i don't really understand how it works) # get ips that aren't the loopback unwrap00 = [ip for ip in socket.gethostbyname_ex(socket.gethostname())[2] if not ip.startswith("127.")][:1] # ??? truly dk unwrap01 = [[(s.connect(('8.8.8.8', 53)), s.getsockname()[0], s.close()) for s in [socket.socket(socket.AF_INET, socket.SOCK_DGRAM)]][0][1]] unwrap2 = [l for l in (unwrap00, unwrap01) if l][0][0] return unwrap2
[docs] def release(self): try: self.closing.set() self.terminate() except AttributeError: # already been called, NoneTypes have no attribute terminate pass
# Stopping the loop should kill the process, as it's what's holding us in run() #self.loop.stop()
[docs]class Terminal_Station(Station): """ :class:`~.networking.Station` object used by :class:`~.Terminal` objects. Spawned without a :attr:`~.Station.pusher`. **Listens** +-------------+-------------------------------------------+-----------------------------------------------+ | Key | Method | Description | +=============+===========================================+===============================================+ | 'PING' | :meth:`~.Terminal_Station.l_ping` | We are asked to confirm that we are alive | +-------------+-------------------------------------------+-----------------------------------------------+ | 'INIT' | :meth:`~.Terminal_Station.l_init` | Ask all pilots to confirm that they are alive | +-------------+-------------------------------------------+-----------------------------------------------+ | 'CHANGE' | :meth:`~.Terminal_Station.l_change` | Change a parameter on the Pi | +-------------+-------------------------------------------+-----------------------------------------------+ | 'STOPALL' | :meth:`~.Terminal_Station.l_stopall` | Stop all pilots and plots | +-------------+-------------------------------------------+-----------------------------------------------+ | 'KILL' | :meth:`~.Terminal_Station.l_kill` | Terminal wants us to die :( | +-------------+-------------------------------------------+-----------------------------------------------+ | 'DATA' | :meth:`~.Terminal_Station.l_data` | Stash incoming data from a Pilot | +-------------+-------------------------------------------+-----------------------------------------------+ | 'STATE' | :meth:`~.Terminal_Station.l_state` | A Pilot has changed state | +-------------+-------------------------------------------+-----------------------------------------------+ | 'HANDSHAKE' | :meth:`~.Terminal_Station.l_handshake` | A Pi is telling us it's alive and its IP | +-------------+-------------------------------------------+-----------------------------------------------+ | 'FILE' | :meth:`~.Terminal_Station.l_file` | The pi needs some file from us | +-------------+-------------------------------------------+-----------------------------------------------+ """ plot_timer = None #send_plot = threading.Event() #send_plot.clear() # dict of threading events that determine how frequently we send plot updates sent_plot = {} def __init__(self, pilots): """ Args: pilots (dict): The :attr:`.Terminal.pilots` dictionary. """ super(Terminal_Station, self).__init__() # by default terminal doesn't have a pusher, everything connects to it self.pusher = False # Store some prefs values self.listen_port = prefs.get('MSGPORT') self.id = 'T' # Message dictionary - What method to call for each type of message received by the terminal class self.listens.update({ 'PING': self.l_ping, # We are asked to confirm that we are alive 'INIT': self.l_init, # We should ask all the pilots to confirm that they are alive 'CHANGE': self.l_change, # Change a parameter on the Pi 'STOPALL': self.l_stopall, # Stop all pilots and plots 'KILL': self.l_kill, # Terminal wants us to die :( 'DATA': self.l_data, # Stash incoming data from an autopilot 'CONTINUOUS': self.l_continuous, # handle incoming continuous data 'STATE': self.l_state, # The Pi is confirming/notifying us that it has changed state 'HANDSHAKE': self.l_handshake, # initial connection with some initial info 'FILE': self.l_file, # The pi needs some file from us }) # dictionary that keeps track of our pilots self.pilots = pilots # start a timer at the draw FPS of the terminal -- only send if prefs.get( 'DRAWFPS'): self.data_fps = float(prefs.get('DRAWFPS')) else: self.data_fps = 20 self.data_ifps = 1.0/self.data_fps
[docs] def start_plot_timer(self): """ Start a timer that controls how often streamed video frames are sent to :class:`.gui.Video` plots. """ self.plot_timer = threading.Thread(target=self._fps_clock) self.plot_timer.setDaemon(True) self.plot_timer.start()
def _fps_clock(self): while not self.closing.is_set(): for k, v in self.sent_plot.items(): try: v.set() except: pass # TODO: Too General #self.send_plot.set() time.sleep(self.data_ifps) ########################## # Message Handling Methods
[docs] def l_ping(self, msg): """ We are asked to confirm that we are alive Respond with a blank 'STATE' message. Args: msg (:class:`.Message`): """ # we are being asked if we're alive # respond with blank message since the terminal isn't really stateful self.send(msg.sender, 'STATE', flags={'NOLOG':True})
[docs] def l_init(self, msg): """ Ask all pilots to confirm that they are alive Sends a "PING" to everyone in the pilots dictionary. Args: msg (:class:`.Message`): """ # Ping all pis that we are expecting given our pilot db # Responses will be handled with l_state so not much needed here for p in self.pilots.keys(): self.send(p, 'PING', flags={'NOLOG':True})
[docs] def l_change(self, msg): """ Change a parameter on the Pi Warning: Not Implemented Args: msg (:class:`.Message`): """ # TODO: Should also handle param changes to GUI objects like ntrials, etc. pass
[docs] def l_stopall(self, msg): """ Stop all pilots and plots Args: msg (:class:`.Message`): """ # let all the pilots and plot objects know that they should stop for p in self.pilots.keys(): self.send(p, 'STOP') self.send("P_{}".format(p), 'STOP')
[docs] def l_kill(self, msg): """ Terminal wants us to die :( Stop the :attr:`.Station.loop` Args: msg (:class:`.Message`): """ self.logger.info('Received kill request') self.closing.set() # Stopping the loop should kill the process, as it's what's holding us in run() self.loop.stop()
[docs] def l_data(self, msg): """ Stash incoming data from a Pilot Just forward this along to the internal terminal object ('_T') and a copy to the relevant plot. Args: msg (:class:`.Message`): """ # Send through to terminal #self.send('_T', 'DATA', msg.value, flags=msg.flags) self.send(to='_T', msg=msg) # Send to plot widget, which should be listening to "P_{pilot_name}" #self.send('P_{}'.format(msg.value['pilot']), 'DATA', msg.value, flags=msg.flags) self.send(to='P_{}'.format(msg.value['pilot']), msg=msg)
[docs] def l_continuous(self, msg): """ Handle the storage of continuous data Forwards all data on to the Terminal's internal :class:`Net_Node`, send to :class:`.Plot` according to update rate in ``prefs.get('DRAWFPS')`` Args: msg (dict): A continuous data message """ if not self.plot_timer: self.start_plot_timer() # Send through to terminal #msg.value.update({'continuous':True}) self.send(to='_T', msg=msg) # Send to plot widget, which should be listening to "P_{pilot_name}" if msg.sender not in self.sent_plot.keys(): self.sent_plot[msg.sender] = threading.Event() if self.sent_plot[msg.sender].is_set(): self.send(to='P_{}'.format(msg.value['pilot']), msg=msg) self.sent_plot[msg.sender].clear()
# def l_continuous(self, msg): # # # Send through to terminal # msg.value.update({'continuous':True}) # self.send('_T', 'DATA', msg.value, flags=msg.flags) # # # Send to plot widget, which should be listening to "P_{pilot_name}" # self.send('P_{}'.format(msg.value['pilot']), 'DATA', msg.value, flags=msg.flags)
[docs] def l_state(self, msg): """ A Pilot has changed state. Stash in 'state' field of pilot dict and send along to _T Args: msg (:class:`.Message`): """ if msg.sender in self.pilots.keys(): #if 'state' in self.pilots[msg.sender].keys(): # if msg.value == self.pilots[msg.sender]['state']: # # if we've already gotten this one, don't send to terminal # return self.pilots[msg.sender]['state'] = msg.value # Tell the terminal so it can update the pilot_db file state = {'state':msg.value, 'pilot':msg.sender} self.send('_T', 'STATE', state) # Tell the plot self.send("P_{}".format(msg.sender), 'STATE', msg.value) self.senders[msg.sender] = msg.value
[docs] def l_handshake(self, msg): """ A Pi is telling us it's alive and its IP. Send along to _T Args: msg (:class:`.Message`): """ # only rly useful for our terminal object self.send('_T', 'HANDSHAKE', value=msg.value)
[docs] def l_file(self, msg): """ A Pilot needs some file from us. Send it back after :meth:`base64.b64encode` ing it. TODO: Split large files into multiple messages... Args: msg (:class:`.Message`): The value field of the message should contain some relative path to a file contained within `prefs.get('SOUNDDIR')` . eg. `'/songs/sadone.wav'` would return `'os.path.join(prefs.get('SOUNDDIR')/songs.sadone.wav'` """ # The <target> pi has requested some file <value> from us, let's send it back # This assumes the file is small, if this starts crashing we'll have to split the message... full_path = os.path.join(prefs.get('SOUNDDIR'), msg.value) with open(full_path, 'rb') as open_file: # encode in base64 so json doesn't complain file_contents = base64.b64encode(open_file.read()) file_message = {'path':msg.value, 'file':file_contents} self.send(msg.sender, 'FILE', file_message)
[docs]class Pilot_Station(Station): """ :class:`~.networking.Station` object used by :class:`~.Pilot` objects. Spawned with a :attr:`~.Station.pusher` connected back to the :class:`~.Terminal` . **Listens** +-------------+-------------------------------------+-----------------------------------------------+ | Key | Method | Description | +=============+=====================================+===============================================+ | 'STATE' | :meth:`~.Pilot_Station.l_state` | Pilot has changed state | | 'COHERE' | :meth:`~.Pilot_Station.l_cohere` | Make sure our data and the Terminal's match. | | 'PING' | :meth:`~.Pilot_Station.l_ping` | The Terminal wants to know if we're listening | | 'START' | :meth:`~.Pilot_Station.l_start` | We are being sent a task to start | | 'STOP' | :meth:`~.Pilot_Station.l_stop` | We are being told to stop the current task | | 'PARAM' | :meth:`~.Pilot_Station.l_change` | The Terminal is changing some task parameter | | 'FILE' | :meth:`~.Pilot_Station.l_file` | We are receiving a file | +-------------+-------------------------------------+-----------------------------------------------+ """ def __init__(self): # Pilot has a pusher - connects back to terminal self.pusher = True if prefs.get('LINEAGE') == 'CHILD': self.push_id = prefs.get('PARENTID').encode('utf-8') self.push_port = prefs.get('PARENTPORT') self.push_ip = prefs.get('PARENTIP') self.child = True else: self.push_id = b'T' self.push_port = prefs.get('PUSHPORT') self.push_ip = prefs.get('TERMINALIP') self.child = False # Store some prefs values self.listen_port = prefs.get('MSGPORT') #self.id = prefs.get('NAME').encode('utf-8') self.id = prefs.get('NAME') self.pi_id = "_{}".format(self.id) self.subject = None # Store current subject ID self.state = None # store current pi state self.child = False # Are we acting as a child right now? self.parent = False # Are we acting as a parent right now? super(Pilot_Station, self).__init__() self.listens.update({ 'STATE': self.l_state, # Confirm or notify terminal of state change 'COHERE': self.l_cohere, # Sending our temporary data table at the end of a run to compare w/ terminal's copy 'PING': self.l_ping, # The Terminal wants to know if we're listening 'START': self.l_start, # We are being sent a task to start 'STOP': self.l_stop, # We are being told to stop the current task 'PARAM': self.l_change, # The Terminal is changing some task parameter 'FILE': self.l_file, # We are receiving a file 'CONTINUOUS': self.l_continuous, # we are sending continuous data to the terminal 'CHILD': self.l_child, 'HANDSHAKE': self.l_noop, 'CALIBRATE_PORT': self.l_forward, 'CALIBRATE_RESULT': self.l_forward, 'BANDWIDTH': self.l_forward }) ###########################3 # Message/Listen handling methods
[docs] def l_noop(self, msg): pass
[docs] def l_state(self, msg): """ Pilot has changed state Stash it and alert the Terminal Args: msg (:class:`.Message`): """ # Save locally so we can respond to queries on our own, then push 'er on through # Value will just have the state, we want to add our name self.state = msg.value self.push(to=self.push_id, key='STATE', value=msg.value)
[docs] def l_cohere(self, msg): """ Send our local version of the data table so the terminal can double check Warning: Not Implemented Args: msg (:class:`.Message`): """ pass
[docs] def l_ping(self, msg): """ The Terminal wants to know our status Push back our current state. Args: msg (:class:`.Message`): """ # The terminal wants to know if we are alive, respond with our name and IP # don't bother the pi self.push(key='STATE', value=self.state, flags={'NOLOG':True})
[docs] def l_start(self, msg): """ We are being sent a task to start If we need any files, request them. Then send along to the pilot. Args: msg (:class:`.Message`): value will contain a dictionary containing a task description. """ self.subject = msg.value['subject'] # TODO: Refactor into a general preflight check. # First make sure we have any sound files that we need # TODO: stim managers need to be able to return list of stimuli and this is a prime reason why if 'stim' in msg.value.keys(): if 'sounds' in msg.value['stim'].keys(): # nested list comprehension to get value['sounds']['L/R'][0-n] f_sounds = [sound for sounds in msg.value['stim']['sounds'].values() for sound in sounds if sound['type'] in ['File', 'Speech']] elif 'manager' in msg.value['stim'].keys(): # we have a manager if msg.value['stim']['type'] == 'sounds': f_sounds = [] for group in msg.value['stim']['groups']: f_sounds.extend([sound for sounds in group['sounds'].values() for sound in sounds if sound['type'] in ['File', 'Speech']]) else: f_sounds = [] if len(f_sounds)>0: # check to see if we have these files, if not, request them for sound in f_sounds: full_path = os.path.join(prefs.get('SOUNDDIR'), sound['path']) if not os.path.exists(full_path): # We ask the terminal to send us the file and then wait. self.logger.info('REQUESTING SOUND {}'.format(sound['path'])) self.push(key='FILE', value=sound['path']) # wait here to get the sound, # the receiving thread will set() when we get it. self.file_block.clear() self.file_block.wait() # If we're starting the task as a child, stash relevant params if 'child' in msg.value.keys(): self.child = True self.parent_id = msg.value['child']['parent'] self.subject = msg.value['child']['subject'] else: self.child = False # once we make sure we have everything, tell the Pilot to start. self.send(self.pi_id, 'START', msg.value)
[docs] def l_stop(self, msg): """ Tell the pi to stop the task Args: msg (:class:`.Message`): """ self.send(self.pi_id, 'STOP')
[docs] def l_change(self, msg): """ The terminal is changing a parameter Warning: Not implemented Args: msg (:class:`.Message`): """ # TODO: Changing some task parameter from the Terminal pass
[docs] def l_file(self, msg): """ We are receiving a file. Decode from b64 and save. Set the file_block. Args: msg (:class:`.Message`): value will have 'path' and 'file', where the path determines where in `prefs.get('SOUNDDIR')` the b64 encoded 'file' will be saved. """ # The file should be of the structure {'path':path, 'file':contents} full_path = os.path.join(prefs.get('SOUNDDIR'), msg.value['path']) # TODO: give Message full deserialization capabilities including this one file_data = base64.b64decode(msg.value['file']) try: os.makedirs(os.path.dirname(full_path)) except: # TODO: Make more specific - only if dir already exists pass with open(full_path, 'wb') as open_file: open_file.write(file_data) self.logger.info('SOUND RECEIVED {}'.format(msg.value['path'])) # If we requested a file, some poor start fn is probably waiting on us self.file_block.set()
[docs] def l_continuous(self, msg): """ Forwards continuous data sent by children back to terminal. Continuous data sources from this pilot should be streamed directly to the terminal. Args: msg (:class:`Message`): Continuous data message """ if self.child: msg.value['pilot'] = self.parent_id msg.value['subject'] = self.subject msg.value['continuous'] = True self.push(to='T', key='DATA', value=msg.value, repeat=False) else: self.logger.warning('Received continuous data but no child found, \ continuous data should be streamed directly to terminal \ from pilot')
[docs] def l_child(self, msg): """ Telling our child to run a task. Args: msg (): Returns: """ if 'KEY' in msg.value.keys(): KEY = msg.value['keys'] else: KEY = 'START' self.send(to=prefs.get('CHILDID'), key=KEY, value=msg.value)
[docs] def l_forward(self, msg): """ Just forward the message to the pi. """ self.send(to=self.pi_id, key=msg.key, value=msg.value)
#####################################
[docs]class Net_Node(object): """ Drop in networking object to be given to any sub-object behind some external-facing :class:`.Station` object. These objects are intended to communicate locally, within a piece of hardware, though not necessarily within the same process. To minimize the complexity of the network topology, Net_Nodes must communicate through a :class:`.Station` ROUTER, rather than address each other directly. Args: id (str): What are we known as? What do we set our :attr:`~zmq.Socket.identity` as? upstream (str): The identity of the ROUTER socket used by our upstream :class:`.Station` object. port (int): The port that our upstream ROUTER socket is bound to listens (dict): Dictionary of functions to call for different types of messages. keys match the :attr:`.Message.key`. instance (bool): Should the node try and use the existing zmq context and tornado loop? upstream_ip (str): If this Net_Node is being used on its own (ie. not behind a :class:`.Station`), it can directly connect to another node at this IP. Otherwise use 'localhost' to connect to a station. route_port (int): Typically, Net_Nodes only have a single Dealer socket and receive messages from their encapsulating :class:`.Station`, but if you want to take this node offroad and use it independently, an int here binds a Router to the port. Attributes: context (:class:`zmq.Context`): zeromq context loop (:class:`tornado.ioloop.IOLoop`): a tornado ioloop sock (:class:`zmq.Socket`): Our DEALER socket. id (str): What are we known as? What do we set our :attr:`~zmq.Socket.identity` as? upstream (str): The identity of the ROUTER socket used by our upstream :class:`.Station` object. port (int): The port that our upstream ROUTER socket is bound to listens (dict): Dictionary of functions to call for different types of messages. keys match the :attr:`.Message.key`. outbox (dict): Messages that have been sent but have not been confirmed timers (dict): dict of :class:`threading.Timer` s that will check in on outbox messages logger (:class:`logging.Logger`): Used to log messages and network events. msg_counter (:class:`itertools.count`): counter to index our sent messages loop_thread (:class:`threading.Thread`): Thread that holds our loop. initialized with `daemon=True` """ context = None loop = None id = None upstream = None # ID of router we connect to port = None listens = {} outbox = {} timers = {} #connected = False logger = None sock = None loop_thread = None repeat_interval = 5 # how many seconds to wait before trying to repeat a message def __init__(self, id, upstream, port, listens, instance=True, upstream_ip='localhost', daemon=True, expand_on_receive=True): """ """ if instance: self.context = zmq.Context.instance() self.loop = IOLoop.current() else: self.context = zmq.Context() self.loop = IOLoop() self.closing = threading.Event() self.closing.clear() # we have a few builtin listens self.listens = { 'CONFIRM': self.l_confirm, #'STREAM' : self.l_stream } # then add the rest self.listens.update(listens) #self.id = id.encode('utf-8') self.id = id #self.upstream = upstream.encode('utf-8') self.upstream = upstream self.port = int(port) # self.connected = False self.msg_counter = count() # try to get a logger self.logger = init_logger(self) # If we were given an explicit IP to connect to, stash it self.upstream_ip = upstream_ip # # If we want to be able to have messages sent to us directly, make a router at this port # self.route_port = route_port self.daemon = daemon self.streams = {} self.expand = expand_on_receive if prefs.get( 'SUBJECT'): self.subject = prefs.get('SUBJECT').encode('utf-8') else: self.subject = None self.init_networking() def __del__(self): self.release()
[docs] def init_networking(self): """ Creates socket, connects to specified port on localhost, and starts the :meth:`~Net_Node.threaded_loop` as a daemon thread. """ self.sock = self.context.socket(zmq.DEALER) #self.sock.identity = self.id self.sock.setsockopt_string(zmq.IDENTITY, self.id) #self.sock.probe_router = 1 # if used locally (typical case), connect to localhost self.sock.connect('tcp://{}:{}'.format(self.upstream_ip, self.port)) # wrap in zmqstreams and start loop thread self.sock = ZMQStream(self.sock, self.loop) self.sock.on_recv(self.handle_listen) # if self.route_port: # # if want to directly receive messages, bind a router port # self.router = self.context.socket(zmq.ROUTER) # self.router.identity = self.id # self.router.bind('tcp://*:{}'.format(self.route_port)) # self.router = ZMQStream(self.router, self.loop) # self.router.on_recv(self.handle_listen) self.loop_thread = threading.Thread(target=self.threaded_loop) if self.daemon: self.loop_thread.daemon = True self.loop_thread.start()
# self.repeat_thread = threading.Thread(target=self.repeat) # if self.daemon: # self.repeat_thread.daemon = True # self.repeat_thread.start() #self.connected = True
[docs] def threaded_loop(self): """ Run in a thread, either starts the IOLoop, or if it is already started (ie. running in another thread), breaks. """ while not self.closing.is_set(): try: self.loop.start() except RuntimeError: # loop already started break
[docs] def handle_listen(self, msg): """ Upon receiving a message, call the appropriate listen method in a new thread and send confirmation it was received. Note: Unlike :meth:`.Station.handle_listen` , only the :attr:`.Message.value` is given to listen methods. This was initially intended to simplify these methods, but this might change in the future to unify the messaging system. Args: msg (str): JSON :meth:`.Message.serialize` d message. """ # messages from dealers are single frames because we only have one connected partner # and that's the dealer spec lol #msg = json.loads(msg[0]) #msg = Message(**msg) # Nodes expand arrays by default as they're expected to msg = Message(msg[-1], expand_arrays=self.expand) # Check if our listen was sent properly if not msg.validate(): if self.logger: self.logger.error('Message failed to validate:\n{}'.format(str(msg))) return # if msg.key == 'CONFIRM': # if msg.value in self.outbox.keys(): # del self.outbox[msg.value] # # # stop a timer thread if we have it # if msg.value in self.timers.keys(): # self.timers[msg.value].cancel() # del self.timers[msg.value] # # self.logger.info('CONFIRMED MESSAGE {}'.format(msg.value)) # else: # Log and spawn thread to respond to listen if isinstance(msg.to, list): if len(msg.to) == 1: msg.to = msg.to[0] if isinstance(msg.to, list): # not to us, just keep it going _ = msg.to.pop(0) self.send(msg=msg, repeat=False) try: listen_funk = self.listens[msg.key] listen_thread = threading.Thread(target=listen_funk, args=(msg.value,)) listen_thread.start() except KeyError: if msg.key=="STREAM": try: listen_thread = threading.Thread(target=self.l_stream, args=(msg,)) listen_thread.start() except Exception as e: self.logger.exception(e) self.logger.error('MSG ID {} - No listen function found for key: {}'.format(msg.id, msg.key)) if (msg.key != "CONFIRM") and ('NOREPEAT' not in msg.flags.keys()) : # send confirmation self.send(msg.sender, 'CONFIRM', msg.id) log_this = True if 'NOLOG' in msg.flags.keys(): log_this = False if self.logger and log_this: self.logger.debug('RECEIVED: {}'.format(str(msg)))
[docs] def send(self, to=None, key=None, value=None, msg=None, repeat=True, flags = None, force_to = False): """ Send a message via our :attr:`~.Net_Node.sock` , DEALER socket. `to` is not required. Every message is always sent to :attr:`~.Net_Node.upstream` . `to` can be included to send a message further up the network tree to a networking object we're not directly connected to. Either an already created :class:`.Message` should be passed as `msg`, or at least `key` must be provided for a new message created by :meth:`~.Net_Node.prepare_message` . A :class:`threading.Timer` is created to resend the message using :meth:`~.Net_Node.repeat` unless `repeat` is False. Args: to (str, list): The identity of the socket this message is to. If not included, sent to :meth:`~.Net_Node.upstream` . key (str): The type of message - used to select which method the receiver uses to process this message. value: Any information this message should contain. Can be any type, but must be JSON serializable. msg (`.Message`): An already created message. repeat (bool): Should this message be resent if confirmation is not received? flags (dict): force_to (bool): If we really really want to use the 'to' field to address messages (eg. node being used for direct communication), overrides default behavior of sending to upstream. """ # send message via the dealer # even though we only have one connection over our dealer, # we still include 'to' in case we are sending further upstream # but can push without 'to', just fill in with upstream id if to is None: to = self.upstream if (key is None) and (msg is None): if self.logger: self.logger.error('Push sent without Key') return if not msg: msg = self.prepare_message(to, key, value, repeat, flags) log_this = True if 'NOLOG' in msg.flags.keys(): log_this = False # Make sure our message has everything # if not msg.validate(): # if self.logger: # self.logger.error('Message Invalid:\n{}'.format(str(msg))) # return # encode message msg_enc = msg.serialize() #pdb.set_trace() if not msg_enc: self.logger.error('Message could not be encoded:\n{}'.format(str(msg))) return if force_to: self.sock.send_multipart([bytes(msg.to, encoding="utf-8"), bytes(msg.to, encoding="utf-8"), msg_enc]) else: self.sock.send_multipart([self.upstream.encode('utf-8'), bytes(msg.to, encoding="utf-8"), msg_enc]) if self.logger and log_this: self.logger.debug("MESSAGE SENT - {}".format(str(msg))) if repeat and not msg.key == "CONFIRM": # add to outbox and spawn timer to resend self.outbox[msg.id] = (time.time(), msg)
# self.timers[msg.id] = threading.Timer(5.0, self.repeat, args=(msg.id,)) # self.timers[msg.id].start()
[docs] def repeat(self): """ Periodically (according to :attr:`~.repeat_interval`) resend messages that haven't been confirmed TTL is decremented, and messages are resent until their TTL is 0. """ while not self.closing.is_set(): # try to send any outstanding messages and delete if too old # make a local copy of dict outbox = copy(self.outbox) if len(outbox) > 0: for id in outbox.keys(): if outbox[id][1].ttl <= 0: self.logger.warning('PUBLISH FAILED {} - {}'.format(id, str(outbox[id][1]))) try: del self.outbox[id] except KeyError: # fine, already deleted pass else: # if we didn't just put this message in the outbox... if (time.time() - outbox[id][0]) > (self.repeat_interval*2): self.logger.debug('REPUBLISH {} - {}'.format(id, str(outbox[id][1]))) self.sock.send_multipart([self.upstream.encode('utf-8'), outbox[id][1].serialize()]) self.outbox[id][1].ttl -= 1 # wait to do it again time.sleep(self.repeat_interval)
[docs] def l_confirm(self, value): """ Confirm that a message was received. Args: value (str): The ID of the message we are confirming. """ # delete message from outbox if we still have it # msg.value should contain the if of the message that was confirmed try: if value in self.outbox.keys(): del self.outbox[value] except KeyError: # already deleted pass # # stop a timer thread if we have it # if value in self.timers.keys(): # self.timers[value].cancel() # del self.timers[value] self.logger.debug('CONFIRMED MESSAGE {}'.format(value))
[docs] def l_stream(self, msg): """ Reconstitute the original stream of messages and call their handling methods The ``msg`` should contain an ``inner_key`` that indicates the key, and thus the handling method. Args: msg (dict): Compressed stream sent by :meth:`Net_Node._stream` """ listen_fn = self.listens[msg.value['inner_key']] old_value = copy(msg.value) delattr(msg, 'value') for v in old_value['payload']: # if isinstance(v, dict) and ('headers' in old_value.keys()): # v.update(old_value['headers']) #msg.value = v listen_fn(v)
# # def l_stream(self, value): # listen_fn = self.listens[value['inner_key']] # for v in value['payload']: # listen_fn(v) # #
[docs] def prepare_message(self, to, key, value, repeat, flags=None): """ Instantiate a :class:`.Message` class, give it an ID and the rest of its attributes. Args: flags: repeat: to (str): The identity of the socket this message is to key (str): The type of message - used to select which method the receiver uses to process this message. value: Any information this message should contain. Can be any type, but must be JSON serializable. """ msg = Message() # if our name is _{something} and our upstream is {something}, replace sender with our upstream node # upstream node should handle all incoming information to those types of nodes #if self.id == "_{}".format(self.upstream): # msg.sender = self.upstream #else: msg.sender = self.id try: msg.to = to.decode('utf-8') except AttributeError: msg.to = to try: msg.key = key.decode('utf-8') except AttributeError: msg.key = key msg.value = value msg_num = next(self.msg_counter) msg.id = "{}_{}".format(self.id, msg_num) if not repeat: msg.flags['NOREPEAT'] = True if flags: for k, v in flags.items(): msg.flags[k] = v return msg
[docs] def get_stream(self, id, key, min_size=5, upstream=None, port = None, ip=None, subject=None): """ Make a queue that another object can dump data into that sends on its own socket. Smarter handling of continuous data than just hitting 'send' a shitload of times. Returns: Queue: Place to dump ur data """ if upstream is None: upstream = self.upstream if port is None: port = self.port if ip is None: ip = self.upstream_ip if subject is None: if self.subject: subject = self.subject elif prefs.get( 'SUBJECT'): subject = prefs.get('SUBJECT') # make a queue q = queue.Queue() stream_thread = threading.Thread(target=self._stream, args=(id, key, min_size, upstream, port, ip, subject, q)) stream_thread.setDaemon(True) stream_thread.start() self.streams[id] = stream_thread self.streams[id] = stream_thread self.logger.info(("Stream started with configuration:\n"+ "ID: {}\n".format(self.id+"_"+id)+ "Key: {}\n".format(key)+ "Min Chunk Size: {}\n".format(min_size)+ "Upstream ID: {}\n".format(upstream) + "Port: {}\n".format(port) + "IP: {}\n".format(ip) + "Subject: {}\n".format(subject))) return q
def _stream(self, id, msg_key, min_size, upstream, port, ip, subject, q): # create a new context and socket #context = zmq.Context() #loop = IOLoop() socket = self.context.socket(zmq.DEALER) socket_id = "{}_{}".format(self.id, id) #socket.identity = socket_id socket.setsockopt_string(zmq.IDENTITY, socket_id) socket.connect('tcp://{}:{}'.format(ip, port)) socket = ZMQStream(socket, self.loop) upstream = upstream.encode('utf-8') if subject is None: if prefs.get( 'SUBJECT'): subject = prefs.get('SUBJECT') else: subject = "" if isinstance(subject, bytes): subject = subject.decode('utf-8') if prefs.get('LINEAGE') == "CHILD": # pilot = bytes(prefs.get('PARENTID'), encoding="utf-8") pilot = prefs.get('PARENTID') else: # pilot = bytes(prefs.get('NAME'), encoding="utf-8") pilot = prefs.get('NAME') msg_counter = count() pending_data = [] if min_size > 1: for data in iter(q.get, 'END'): if isinstance(data, tuple): # tuples are immutable, so can't serialize numpy arrays they contain data = list(data) pending_data.append(data) if not socket.sending() and len(pending_data)>=min_size: msg = Message(to=upstream.decode('utf-8'), key="STREAM", value={'inner_key' : msg_key, 'headers' : {'subject': subject, 'pilot' : pilot, 'continuous': True}, 'payload' : pending_data}, id="{}_{}".format(id, next(msg_counter)), flags={'NOREPEAT':True, 'MINPRINT':True}, sender=socket_id).serialize() last_msg = socket.send_multipart((upstream, upstream, msg), track=True, copy=True) self.logger.debug("STREAM {}: Sent {} items".format(self.id+'_'+id, len(pending_data))) pending_data = [] else: # just send like normal messags for data in iter(q.get, 'END'): if isinstance(data, tuple): # tuples are immutable, so can't serialize numpy arrays they contain data = list(data) if not socket.sending(): msg = Message(to=upstream.decode('utf-8'), key=msg_key, subject=subject, pilot=pilot, continuous=True, value=data, flags={'NOREPEAT': True, 'MINPRINT': True}, id="{}_{}".format(id, next(msg_counter)), sender=socket_id).serialize() last_msg = socket.send_multipart((upstream, upstream, msg), track=True, copy=True) self.logger.debug("STREAM {}: Sent 1 item".format(self.id + '_' + id))
[docs] def release(self): self.closing.set() self.loop.stop()
[docs]class Message(object): """ A formatted message. `id`, `to`, `sender`, and `key` are required attributes, but any other key-value pair passed on init is added to the message's attributes and included in the message. Can be indexed and set like a dictionary (message['key'], etc.) Attributes: id (str): ID that uniquely identifies a message. format {sender.id}_{number} to (str): ID of socket this message is addressed to sender (str): ID of socket where this message originates key (str): Type of message, used to select a listen method to process it value: Body of message, can be any type but must be JSON serializable. timestamp (str): Timestamp of message creation ttl (int): Time-To-Live, each message is sent this many times at max, each send decrements ttl. """ # TODO: just make serialization handle all attributes except Files which need to be b64 encoded first. id = None # number of message, format {sender.id}_{number} to = None sender = None key = None # value is the only attribute that can be left None, # ie. with signal-type messages like "STOP" value = None timestamp = None flags = {} ttl = 2 # every message starts with 2 retries. only relevant to the sender so not serialized. changed = False serialized = None def __init__(self, msg=None, expand_arrays = False, **kwargs): # type: (object, object) -> None # Messages don't need to have all attributes on creation, # but do need them to serialize """ Args: *args: **kwargs: """ # optional attrs should be instance attributes so they are caught by _-dict__ self.flags = {} self.timestamp = None self.ttl = 5 #set_trace(term_size=(120,40)) #if len(args)>1: # Exception("Messages can only be constructed with a single positional argument, which is assumed to be a serialized message") #elif len(args)>0: if msg: self.serialized = msg if expand_arrays: deserialized = json.loads(msg, object_pairs_hook=self._deserialize_numpy) else: deserialized = json.loads(msg) kwargs.update(deserialized) for k, v in kwargs.items(): setattr(self, k, v) #self[k] = v # if we're not a previous message being recreated, get a timestamp for our creation if 'timestamp' not in kwargs.keys(): self.get_timestamp() # self.DETECTED_MINPRINT = False def __str__(self): # type: () -> str # if len(str(self.value))>100: # self.DETECTED_MINPRINT = True # TODO: Make verbose/debugging mode, print value in that case. if self.key == 'FILE' or ('MINPRINT' in self.flags.keys()): me_string = "ID: {}; TO: {}; SENDER: {}; KEY: {}, FLAGS: {}".format(self.id, self.to, self.sender, self.key, self.flags) else: me_string = "ID: {}; TO: {}; SENDER: {}; KEY: {}; FLAGS: {}; VALUE: {}".format(self.id, self.to, self.sender, self.key, self.flags, self.value) #me_string = "ID: {}; TO: {}; SENDER: {}; KEY: {}".format(self.id, self.to, self.sender, self.key) return me_string # enable dictionary-like behavior
[docs] def __getitem__(self, key): """ Args: key: """ #value = self._check_dec(self.__dict__[key]) # TODO: Recursively walk looking for 'NUMPY ARRAY' and expand before giving return self.__dict__[key]
[docs] def __setitem__(self, key, value): """ Args: key: value: """ # self.changed=True #value = self._check_enc(value) self.__dict__[key] = value
# def __setattr__(self, key, value): # self.changed=True # #value = self._check_enc(value) # super(Message, self).__setattr__(self, key, value) # self.__dict__[key] = value # def __getattr__(self, key): # #value = self._check_dec(self.__dict__[key]) # return self.__dict__[key] # # def _check_enc(self, value): # if isinstance(value, np.ndarray): # value = json_tricks.dumps(value) # elif isinstance(value, dict): # for k, v in value.items(): # value[k] = self._check_enc(v) # elif isinstance(value, list): # value = [self._check_enc(v) for v in value] # return value # # def _check_dec(self, value): # # # if numpy array, reconstitute # if isinstance(value, basestring): # if value.startswith('{"__ndarray__'): # value = json_tricks.loads(value) # elif isinstance(value, dict): # for k, v in value.items(): # value[k] = self._check_dec(v) # elif isinstance(value, list): # value = [self._check_dec(v) for v in value] # return value
[docs] def _serialize_numpy(self, array): """ Serialize a numpy array for sending over the wire Args: array: Returns: """ compressed = base64.b64encode(blosc.pack_array(array)).decode('ascii') return {'NUMPY_ARRAY': compressed}
def _deserialize_numpy(self, obj_pairs): # print(len(obj_pairs), obj_pairs) if (len(obj_pairs) == 1) and obj_pairs[0][0] == "NUMPY_ARRAY": return blosc.unpack_array(base64.b64decode(obj_pairs[0][1])) else: return dict(obj_pairs)
[docs] def expand(self): """ Don't decompress numpy arrays by default for faster IO, explicitly expand them when needed :return: """ pass
[docs] def __delitem__(self, key): """ Args: key: """ self.changed=True del self.__dict__[key]
[docs] def __contains__(self, key): """ Args: key: """ return key in self.__dict__
def __len__(self): return len(self.__dict__)
[docs] def get_timestamp(self): """ Get a Python timestamp Returns: str: Isoformatted timestamp from ``datetime`` """ self.timestamp = datetime.datetime.now().isoformat()
[docs] def validate(self): """ Checks if `id`, `to`, `sender`, and `key` are all defined. Returns: bool (True): Does message have all required attributes set? """ valid = True for prop in (self.id, self.to, self.sender, self.key): if prop is None: valid = False return valid
[docs] def serialize(self): """ Serializes all attributes in `__dict__` using json. Returns: str: JSON serialized message. """ if not self.changed and self.serialized: return self.serialized valid = self.validate() if not valid: Exception("""Message invalid at the time of serialization!\n {}""".format(str(self))) return False # msg = { # 'id': self.id, # 'to': self.to, # 'sender': self.sender, # 'key': self.key, # 'value': self.value # } msg = self.__dict__ # exclude 'serialized' so it's not in there twice try: del msg['serialized'] except KeyError: pass try: msg_enc = json.dumps(msg, default=self._serialize_numpy).encode('utf-8') self.serialized = msg_enc self.changed=False return msg_enc except: return False
[docs]def serialize_array(array): """ Pack an array with :func:`blosc.pack_array` and serialize with :func:`base64.b64encode` Args: array (:class:`numpy.ndarray`): Array to serialize Returns: dict: {'NUMPY_ARRAY': base-64 encoded, blosc-compressed array.} """ compressed = base64.b64encode(blosc.pack_array(array)).decode('ascii') return {'NUMPY_ARRAY': compressed}