X-Git-Url: https://git.novaco.in/?a=blobdiff_plain;f=p2pool%2Fp2p.py;h=bc61982eca3c0548e4f19aa7f18d6fef705230c9;hb=9541a77695bc98cccff001541ccb3880bca6d875;hp=2089d55392d4d9f2d06d88356d497cd86865fc1e;hpb=38a461338e2ba55c7cdf28e360209d7771b03086;p=p2pool.git diff --git a/p2pool/p2p.py b/p2pool/p2p.py index 2089d55..bc61982 100644 --- a/p2pool/p2p.py +++ b/p2pool/p2p.py @@ -1,42 +1,49 @@ from __future__ import division +import math import random import sys import time -from twisted.internet import defer, error, protocol, reactor -from twisted.python import log +from twisted.internet import defer, protocol, reactor +from twisted.python import failure, log import p2pool from p2pool import data as p2pool_data -from p2pool.bitcoin import p2p as bitcoin_p2p from p2pool.bitcoin import data as bitcoin_data -from p2pool.util import deferral, pack +from p2pool.util import deferral, p2protocol, pack, variable -class Protocol(bitcoin_p2p.BaseProtocol): - version = 2 - sub_version = p2pool.__version__ +class PeerMisbehavingError(Exception): + pass + + +def fragment(f, **kwargs): + try: + f(**kwargs) + except p2protocol.TooLong: + fragment(f, **dict((k, v[:len(v)//2]) for k, v in kwargs.iteritems())) + fragment(f, **dict((k, v[len(v)//2:]) for k, v in kwargs.iteritems())) + +class Protocol(p2protocol.Protocol): + max_remembered_txs_size = 2500000 def __init__(self, node, incoming): + p2protocol.Protocol.__init__(self, node.net.PREFIX, 1000000, node.traffic_happened) self.node = node self.incoming = incoming - self._prefix = self.node.net.PREFIX - - max_payload_length = 1000000 - - other_version = None - connected2 = False + self.other_version = None + self.connected2 = False def connectionMade(self): - bitcoin_p2p.BaseProtocol.connectionMade(self) - self.factory.proto_made_connection(self) + self.connection_lost_event = variable.Event() + self.addr = self.transport.getPeer().host, self.transport.getPeer().port self.send_version( - version=self.version, + version=1300, services=0, addr_to=dict( services=0, @@ -49,50 +56,52 @@ class Protocol(bitcoin_p2p.BaseProtocol): port=self.transport.getHost().port, ), nonce=self.node.nonce, - sub_version=self.sub_version, + sub_version=p2pool.__version__, mode=1, best_share_hash=self.node.best_share_hash_func(), ) - reactor.callLater(10, self._connect_timeout) - self.timeout_delayed = reactor.callLater(100, self._timeout) + self.timeout_delayed = reactor.callLater(10, self._connect_timeout) - old_dataReceived = self.dataReceived - def new_dataReceived(data): - if not self.timeout_delayed.called: - self.timeout_delayed.reset(100) - old_dataReceived(data) - self.dataReceived = new_dataReceived + self.get_shares = deferral.GenericDeferrer( + max_id=2**256, + func=lambda id, hashes, parents, stops: self.send_sharereq(id=id, hashes=hashes, parents=parents, stops=stops), + timeout=15, + on_timeout=self.disconnect, + ) + + self.remote_tx_hashes = set() # view of peer's known_txs # not actually initially empty, but sending txs instead of tx hashes won't hurt + self.remote_remembered_txs_size = 0 + + self.remembered_txs = {} # view of peer's mining_txs + self.remembered_txs_size = 0 + self.known_txs_cache = {} def _connect_timeout(self): - if not self.connected2 and self.transport.connected: - print 'Handshake timed out, disconnecting from %s:%i' % self.addr - self.transport.loseConnection() + self.timeout_delayed = None + print 'Handshake timed out, disconnecting from %s:%i' % self.addr + self.disconnect() def packetReceived(self, command, payload2): - if command != 'version' and not self.connected2: - self.transport.loseConnection() - return - - bitcoin_p2p.BaseProtocol.packetReceived(self, command, payload2) + try: + if command != 'version' and not self.connected2: + raise PeerMisbehavingError('first message was not version message') + p2protocol.Protocol.packetReceived(self, command, payload2) + except PeerMisbehavingError, e: + print 'Peer %s:%i misbehaving, will drop and ban. Reason:' % self.addr, e.message + self.badPeerHappened() + + def badPeerHappened(self): + if p2pool.DEBUG: + print "Bad peer banned:", self.addr + self.disconnect() + if self.transport.getPeer().host != '127.0.0.1': # never ban localhost + self.node.bans[self.transport.getPeer().host] = time.time() + 60*60 def _timeout(self): - if self.transport.connected: - print 'Connection timed out, disconnecting from %s:%i' % self.addr - self.transport.loseConnection() - - @defer.inlineCallbacks - def _think(self): - while self.connected2: - self.send_ping() - yield deferral.sleep(random.expovariate(1/100)) - - @defer.inlineCallbacks - def _think2(self): - while self.connected2: - self.send_addrme(port=self.node.port) - #print 'sending addrme' - yield deferral.sleep(random.expovariate(1/(100*len(self.node.peers) + 1))) + self.timeout_delayed = None + print 'Connection timed out, disconnecting from %s:%i' % self.addr + self.disconnect() message_version = pack.ComposedType([ ('version', pack.IntType(32)), @@ -105,32 +114,82 @@ class Protocol(bitcoin_p2p.BaseProtocol): ('best_share_hash', pack.PossiblyNoneType(0, pack.IntType(256))), ]) def handle_version(self, version, services, addr_to, addr_from, nonce, sub_version, mode, best_share_hash): - if self.other_version is not None or version < 2: - self.transport.loseConnection() - return + if self.other_version is not None: + raise PeerMisbehavingError('more than one version message') + if version < 1300: + raise PeerMisbehavingError('peer too old') self.other_version = version self.other_sub_version = sub_version[:512] self.other_services = services if nonce == self.node.nonce: - #print 'Detected connection to self, disconnecting from %s:%i' % self.addr - self.transport.loseConnection() - return + raise PeerMisbehavingError('was connected to self') if nonce in self.node.peers: - #print 'Detected duplicate connection, disconnecting from %s:%i' % self.addr - self.transport.loseConnection() + if p2pool.DEBUG: + print 'Detected duplicate connection, disconnecting from %s:%i' % self.addr + self.disconnect() return self.nonce = nonce self.connected2 = True + + self.timeout_delayed.cancel() + self.timeout_delayed = reactor.callLater(100, self._timeout) + + old_dataReceived = self.dataReceived + def new_dataReceived(data): + if self.timeout_delayed is not None: + self.timeout_delayed.reset(100) + old_dataReceived(data) + self.dataReceived = new_dataReceived + self.factory.proto_connected(self) - self._think() - self._think2() + self._stop_thread = deferral.run_repeatedly(lambda: [ + self.send_ping(), + random.expovariate(1/100)][-1]) + + self._stop_thread2 = deferral.run_repeatedly(lambda: [ + self.send_addrme(port=self.node.serverfactory.listen_port.getHost().port) if self.node.serverfactory.listen_port is not None else None, + random.expovariate(1/(100*len(self.node.peers) + 1))][-1]) if best_share_hash is not None: self.node.handle_share_hashes([best_share_hash], self) + + def update_remote_view_of_my_known_txs(before, after): + added = set(after) - set(before) + removed = set(before) - set(after) + if added: + self.send_have_tx(tx_hashes=list(added)) + if removed: + self.send_losing_tx(tx_hashes=list(removed)) + + # cache forgotten txs here for a little while so latency of "losing_tx" packets doesn't cause problems + key = max(self.known_txs_cache) + 1 if self.known_txs_cache else 0 + self.known_txs_cache[key] = dict((h, before[h]) for h in removed) + reactor.callLater(20, self.known_txs_cache.pop, key) + watch_id = self.node.known_txs_var.transitioned.watch(update_remote_view_of_my_known_txs) + self.connection_lost_event.watch(lambda: self.node.known_txs_var.transitioned.unwatch(watch_id)) + + self.send_have_tx(tx_hashes=self.node.known_txs_var.value.keys()) + + def update_remote_view_of_my_mining_txs(before, after): + added = set(after) - set(before) + removed = set(before) - set(after) + if added: + self.remote_remembered_txs_size += sum(100 + bitcoin_data.tx_type.packed_size(after[x]) for x in added) + assert self.remote_remembered_txs_size <= self.max_remembered_txs_size + fragment(self.send_remember_tx, tx_hashes=[x for x in added if x in self.remote_tx_hashes], txs=[after[x] for x in added if x not in self.remote_tx_hashes]) + if removed: + self.send_forget_tx(tx_hashes=list(removed)) + self.remote_remembered_txs_size -= sum(100 + bitcoin_data.tx_type.packed_size(before[x]) for x in removed) + watch_id2 = self.node.mining_txs_var.transitioned.watch(update_remote_view_of_my_mining_txs) + self.connection_lost_event.watch(lambda: self.node.mining_txs_var.transitioned.unwatch(watch_id2)) + + self.remote_remembered_txs_size += sum(100 + bitcoin_data.tx_type.packed_size(x) for x in self.node.mining_txs_var.value.values()) + assert self.remote_remembered_txs_size <= self.max_remembered_txs_size + fragment(self.send_remember_tx, tx_hashes=[], txs=self.node.mining_txs_var.value.values()) message_ping = pack.ComposedType([]) def handle_ping(self): @@ -179,7 +238,7 @@ class Protocol(bitcoin_p2p.BaseProtocol): count = 100 self.send_addrs(addrs=[ dict( - timestamp=self.node.addr_store[host, port][2], + timestamp=int(self.node.addr_store[host, port][2]), address=dict( services=self.node.addr_store[host, port][0], address=host, @@ -189,40 +248,192 @@ class Protocol(bitcoin_p2p.BaseProtocol): self.node.get_good_peers(count) ]) - message_getshares = pack.ComposedType([ + message_shares = pack.ComposedType([ + ('shares', pack.ListType(p2pool_data.share_type)), + ]) + def handle_shares(self, shares): + result = [] + for wrappedshare in shares: + if wrappedshare['type'] < 9: continue + share = p2pool_data.load_share(wrappedshare, self.node.net, self.addr) + if wrappedshare['type'] >= 13: + txs = [] + for tx_hash in share.share_info['new_transaction_hashes']: + if tx_hash in self.node.known_txs_var.value: + tx = self.node.known_txs_var.value[tx_hash] + else: + for cache in self.known_txs_cache.itervalues(): + if tx_hash in cache: + tx = cache[tx_hash] + print 'Transaction %064x rescued from peer latency cache!' % (tx_hash,) + break + else: + print >>sys.stderr, 'Peer referenced unknown transaction %064x, disconnecting' % (tx_hash,) + self.disconnect() + return + txs.append(tx) + else: + txs = None + + result.append((share, txs)) + + self.node.handle_shares(result, self) + + def sendShares(self, shares, tracker, known_txs, include_txs_with=[]): + tx_hashes = set() + for share in shares: + if share.VERSION >= 13: + # send full transaction for every new_transaction_hash that peer does not know + for tx_hash in share.share_info['new_transaction_hashes']: + assert tx_hash in known_txs, 'tried to broadcast share without knowing all its new transactions' + if tx_hash not in self.remote_tx_hashes: + tx_hashes.add(tx_hash) + continue + if share.hash in include_txs_with: + x = share.get_other_tx_hashes(tracker) + if x is not None: + tx_hashes.update(x) + + hashes_to_send = [x for x in tx_hashes if x not in self.node.mining_txs_var.value and x in known_txs] + + new_remote_remembered_txs_size = self.remote_remembered_txs_size + sum(100 + bitcoin_data.tx_type.packed_size(known_txs[x]) for x in hashes_to_send) + if new_remote_remembered_txs_size > self.max_remembered_txs_size: + raise ValueError('shares have too many txs') + self.remote_remembered_txs_size = new_remote_remembered_txs_size + + fragment(self.send_remember_tx, tx_hashes=[x for x in hashes_to_send if x in self.remote_tx_hashes], txs=[known_txs[x] for x in hashes_to_send if x not in self.remote_tx_hashes]) + + fragment(self.send_shares, shares=[share.as_share() for share in shares]) + + self.send_forget_tx(tx_hashes=hashes_to_send) + + self.remote_remembered_txs_size -= sum(100 + bitcoin_data.tx_type.packed_size(known_txs[x]) for x in hashes_to_send) + + + message_sharereq = pack.ComposedType([ + ('id', pack.IntType(256)), ('hashes', pack.ListType(pack.IntType(256))), ('parents', pack.VarIntType()), ('stops', pack.ListType(pack.IntType(256))), ]) - def handle_getshares(self, hashes, parents, stops): - self.node.handle_get_shares(hashes, parents, stops, self) - - message_shares = pack.ComposedType([ + def handle_sharereq(self, id, hashes, parents, stops): + shares = self.node.handle_get_shares(hashes, parents, stops, self) + try: + self.send_sharereply(id=id, result='good', shares=[share.as_share() for share in shares]) + except p2protocol.TooLong: + self.send_sharereply(id=id, result='too long', shares=[]) + + message_sharereply = pack.ComposedType([ + ('id', pack.IntType(256)), + ('result', pack.EnumType(pack.VarIntType(), {0: 'good', 1: 'too long', 2: 'unk2', 3: 'unk3', 4: 'unk4', 5: 'unk5', 6: 'unk6'})), ('shares', pack.ListType(p2pool_data.share_type)), ]) - def handle_shares(self, shares): - res = [] - for share in shares: - share_obj = p2pool_data.Share.from_share(share, self.node.net) - share_obj.peer = self - res.append(share_obj) - self.node.handle_shares(res, self) - - def sendShares(self, shares, full=False): - def att(f, **kwargs): - try: - f(**kwargs) - except bitcoin_p2p.TooLong: - att(f, **dict((k, v[:len(v)//2]) for k, v in kwargs.iteritems())) - att(f, **dict((k, v[len(v)//2:]) for k, v in kwargs.iteritems())) - if shares: - att(self.send_shares, shares=[share.as_share() for share in shares]) + def handle_sharereply(self, id, result, shares): + if result == 'good': + res = [p2pool_data.load_share(share, self.node.net, self.addr) for share in shares if share['type'] >= 9] + else: + res = failure.Failure("sharereply result: " + result) + self.get_shares.got_response(id, res) + + + message_bestblock = pack.ComposedType([ + ('header', bitcoin_data.block_header_type), + ]) + def handle_bestblock(self, header): + self.node.handle_bestblock(header, self) + + + message_have_tx = pack.ComposedType([ + ('tx_hashes', pack.ListType(pack.IntType(256))), + ]) + def handle_have_tx(self, tx_hashes): + #assert self.remote_tx_hashes.isdisjoint(tx_hashes) + self.remote_tx_hashes.update(tx_hashes) + while len(self.remote_tx_hashes) > 10000: + self.remote_tx_hashes.pop() + message_losing_tx = pack.ComposedType([ + ('tx_hashes', pack.ListType(pack.IntType(256))), + ]) + def handle_losing_tx(self, tx_hashes): + #assert self.remote_tx_hashes.issuperset(tx_hashes) + self.remote_tx_hashes.difference_update(tx_hashes) + + + message_remember_tx = pack.ComposedType([ + ('tx_hashes', pack.ListType(pack.IntType(256))), + ('txs', pack.ListType(bitcoin_data.tx_type)), + ]) + def handle_remember_tx(self, tx_hashes, txs): + for tx_hash in tx_hashes: + if tx_hash in self.remembered_txs: + print >>sys.stderr, 'Peer referenced transaction twice, disconnecting' + self.disconnect() + return + + if tx_hash in self.node.known_txs_var.value: + tx = self.node.known_txs_var.value[tx_hash] + else: + for cache in self.known_txs_cache.itervalues(): + if tx_hash in cache: + tx = cache[tx_hash] + print 'Transaction %064x rescued from peer latency cache!' % (tx_hash,) + break + else: + print >>sys.stderr, 'Peer referenced unknown transaction %064x, disconnecting' % (tx_hash,) + self.disconnect() + return + + self.remembered_txs[tx_hash] = tx + self.remembered_txs_size += 100 + bitcoin_data.tx_type.packed_size(tx) + new_known_txs = dict(self.node.known_txs_var.value) + warned = False + for tx in txs: + tx_hash = bitcoin_data.hash256(bitcoin_data.tx_type.pack(tx)) + if tx_hash in self.remembered_txs: + print >>sys.stderr, 'Peer referenced transaction twice, disconnecting' + self.disconnect() + return + + if tx_hash in self.node.known_txs_var.value and not warned: + print 'Peer sent entire transaction %064x that was already received' % (tx_hash,) + warned = True + + self.remembered_txs[tx_hash] = tx + self.remembered_txs_size += 100 + bitcoin_data.tx_type.packed_size(tx) + new_known_txs[tx_hash] = tx + self.node.known_txs_var.set(new_known_txs) + if self.remembered_txs_size >= self.max_remembered_txs_size: + raise PeerMisbehavingError('too much transaction data stored') + message_forget_tx = pack.ComposedType([ + ('tx_hashes', pack.ListType(pack.IntType(256))), + ]) + def handle_forget_tx(self, tx_hashes): + for tx_hash in tx_hashes: + self.remembered_txs_size -= 100 + bitcoin_data.tx_type.packed_size(self.remembered_txs[tx_hash]) + assert self.remembered_txs_size >= 0 + del self.remembered_txs[tx_hash] + def connectionLost(self, reason): + self.connection_lost_event.happened() + if self.timeout_delayed is not None: + self.timeout_delayed.cancel() if self.connected2: self.factory.proto_disconnected(self, reason) + self._stop_thread() + self._stop_thread2() self.connected2 = False self.factory.proto_lost_connection(self, reason) + if p2pool.DEBUG: + print "Peer connection lost:", self.addr, reason + self.get_shares.respond_all(reason) + + @defer.inlineCallbacks + def do_ping(self): + start = reactor.seconds() + yield self.get_shares(hashes=[0], parents=0, stops=[]) + end = reactor.seconds() + defer.returnValue(end - start) class ServerFactory(protocol.ServerFactory): def __init__(self, node, max_conns): @@ -231,12 +442,17 @@ class ServerFactory(protocol.ServerFactory): self.conns = {} self.running = False + self.listen_port = None def buildProtocol(self, addr): if sum(self.conns.itervalues()) >= self.max_conns or self.conns.get(self._host_to_ident(addr.host), 0) >= 3: return None + if addr.host in self.node.bans and self.node.bans[addr.host] > time.time(): + return None p = Protocol(self.node, True) p.factory = self + if p2pool.DEBUG: + print "Got peer connection from:", addr return p def _host_to_ident(self, host): @@ -262,20 +478,15 @@ class ServerFactory(protocol.ServerFactory): self.running = True def attempt_listen(): - if not self.running: - return - try: + if self.running: self.listen_port = reactor.listenTCP(self.node.port, self) - except error.CannotListenError, e: - print >>sys.stderr, 'Error binding to P2P port: %s. Retrying in 3 seconds.' % (e.socketError,) - reactor.callLater(3, attempt_listen) - attempt_listen() + deferral.retry('Error binding to P2P port:', traceback=False)(attempt_listen)() def stop(self): assert self.running self.running = False - self.listen_port.stopListening() + return self.listen_port.stopListening() class ClientFactory(protocol.ClientFactory): def __init__(self, node, desired_conns, max_attempts): @@ -323,25 +534,28 @@ class ClientFactory(protocol.ClientFactory): def start(self): assert not self.running self.running = True - self._think() + self._stop_thinking = deferral.run_repeatedly(self._think) def stop(self): assert self.running self.running = False + self._stop_thinking() - @defer.inlineCallbacks def _think(self): - while self.running: - try: - if len(self.conns) < self.desired_conns and len(self.attempts) < self.max_attempts and self.node.addr_store: - (host, port), = self.node.get_good_peers(1) - - if self._host_to_ident(host) not in self.attempts: - #print 'Trying to connect to', host, port - reactor.connectTCP(host, port, self, timeout=5) - except: - log.err() - - yield deferral.sleep(random.expovariate(1/1)) + try: + if len(self.conns) < self.desired_conns and len(self.attempts) < self.max_attempts and self.node.addr_store: + (host, port), = self.node.get_good_peers(1) + + if self._host_to_ident(host) in self.attempts: + pass + elif host in self.node.bans and self.node.bans[host] > time.time(): + pass + else: + #print 'Trying to connect to', host, port + reactor.connectTCP(host, port, self, timeout=5) + except: + log.err() + + return random.expovariate(1/1) class SingleClientFactory(protocol.ReconnectingClientFactory): def __init__(self, node): @@ -364,16 +578,20 @@ class SingleClientFactory(protocol.ReconnectingClientFactory): self.node.lost_conn(proto, reason) class Node(object): - def __init__(self, best_share_hash_func, port, net, addr_store={}, connect_addrs=set(), desired_outgoing_conns=10, max_outgoing_attempts=30, max_incoming_conns=50, preferred_storage=1000): + def __init__(self, best_share_hash_func, port, net, addr_store={}, connect_addrs=set(), desired_outgoing_conns=10, max_outgoing_attempts=30, max_incoming_conns=50, preferred_storage=1000, known_txs_var=variable.Variable({}), mining_txs_var=variable.Variable({})): self.best_share_hash_func = best_share_hash_func self.port = port self.net = net self.addr_store = dict(addr_store) self.connect_addrs = connect_addrs self.preferred_storage = preferred_storage + self.known_txs_var = known_txs_var + self.mining_txs_var = mining_txs_var + self.traffic_happened = variable.Event() self.nonce = random.randrange(2**64) self.peers = {} + self.bans = {} # address -> end_time self.clientfactory = ClientFactory(self, desired_outgoing_conns, max_outgoing_attempts) self.serverfactory = ServerFactory(self, max_incoming_conns) self.running = False @@ -388,29 +606,30 @@ class Node(object): self.running = True - self._think2() + self._stop_thinking = deferral.run_repeatedly(self._think) - @defer.inlineCallbacks - def _think2(self): - while self.running: - try: - if len(self.addr_store) < self.preferred_storage and self.peers: - random.choice(self.peers.values()).send_getaddrs(count=8) - except: - log.err() - - yield deferral.sleep(random.expovariate(1/20)) + def _think(self): + try: + if len(self.addr_store) < self.preferred_storage and self.peers: + random.choice(self.peers.values()).send_getaddrs(count=8) + except: + log.err() + + return random.expovariate(1/20) + @defer.inlineCallbacks def stop(self): if not self.running: raise ValueError('already stopped') self.running = False - self.clientfactory.stop() - self.serverfactory.stop() + self._stop_thinking() + yield self.clientfactory.stop() + yield self.serverfactory.stop() for singleclientconnector in self.singleclientconnectors: - singleclientconnector.factory.stopTrying() # XXX will this disconnect a current connection? + yield singleclientconnector.factory.stopTrying() + yield singleclientconnector.disconnect() del self.singleclientconnectors def got_conn(self, conn): @@ -435,7 +654,8 @@ class Node(object): old_services, old_first_seen, old_last_seen = self.addr_store[host, port] self.addr_store[host, port] = services, old_first_seen, max(old_last_seen, timestamp) else: - self.addr_store[host, port] = services, timestamp, timestamp + if len(self.addr_store) < 10000: + self.addr_store[host, port] = services, timestamp, timestamp def handle_shares(self, shares, peer): print 'handle_shares', (shares, peer) @@ -446,17 +666,11 @@ class Node(object): def handle_get_shares(self, hashes, parents, stops, peer): print 'handle_get_shares', (hashes, parents, stops, peer) + def handle_bestblock(self, header, peer): + print 'handle_bestblock', header + def get_good_peers(self, max_count): t = time.time() - return [x[0] for x in sorted(self.addr_store.iteritems(), key=lambda (k, (services, first_seen, last_seen)): -max(3600, last_seen - first_seen)/max(3600, t - last_seen)*random.expovariate(1))][:max_count] - -if __name__ == '__main__': - p = random.randrange(2**15, 2**16) - for i in xrange(5): - p2 = random.randrange(2**15, 2**16) - print p, p2 - n = Node(p2, True, {addrdb_key.pack(dict(address='127.0.0.1', port=p)): addrdb_value.pack(dict(services=0, first_seen=int(time.time())-10, last_seen=int(time.time())))}) - n.start() - p = p2 - - reactor.run() + return [x[0] for x in sorted(self.addr_store.iteritems(), key=lambda (k, (services, first_seen, last_seen)): + -math.log(max(3600, last_seen - first_seen))/math.log(max(3600, t - last_seen))*random.expovariate(1) + )][:max_count]