broadcast shares in serial
[p2pool.git] / p2pool / main.py
index d44db83..4073f7f 100644 (file)
@@ -20,10 +20,10 @@ from twisted.web import server
 from twisted.python import log
 from nattraverso import portmapper, ipdiscover
 
-import bitcoin.p2p as bitcoin_p2p, bitcoin.getwork as bitcoin_getwork, bitcoin.data as bitcoin_data
+import bitcoin.p2p as bitcoin_p2p, bitcoin.data as bitcoin_data
 from bitcoin import worker_interface, height_tracker
-from util import expiring_dict, fixargparse, jsonrpc, variable, deferral, math, logging, pack
-from . import p2p, networks, web
+from util import expiring_dict, fixargparse, jsonrpc, variable, deferral, math, logging
+from . import p2p, networks, web, work
 import p2pool, p2pool.data as p2pool_data
 
 @deferral.retry('Error getting work from bitcoind:', 3)
@@ -39,288 +39,17 @@ def getwork(bitcoind):
     packed_transactions = [x.decode('hex') for x in work['transactions']]
     defer.returnValue(dict(
         version=work['version'],
-        previous_block_hash=int(work['previousblockhash'], 16),
+        previous_block=int(work['previousblockhash'], 16),
         transactions=map(bitcoin_data.tx_type.unpack, packed_transactions),
         merkle_link=bitcoin_data.calculate_merkle_link([None] + map(bitcoin_data.hash256, packed_transactions), 0),
         subsidy=work['coinbasevalue'],
         time=work['time'],
         bits=bitcoin_data.FloatingIntegerType().unpack(work['bits'].decode('hex')[::-1]) if isinstance(work['bits'], (str, unicode)) else bitcoin_data.FloatingInteger(work['bits']),
         coinbaseflags=work['coinbaseflags'].decode('hex') if 'coinbaseflags' in work else ''.join(x.decode('hex') for x in work['coinbaseaux'].itervalues()) if 'coinbaseaux' in work else '',
+        clock_offset=time.time() - work['time'],
+        last_update=time.time(),
     ))
 
-class WorkerBridge(worker_interface.WorkerBridge):
-    def __init__(self, lp_signal, my_pubkey_hash, net, donation_percentage, current_work, merged_work, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, worker_fee, p2p_node, submit_block, set_best_share, shared_share_hashes):
-        worker_interface.WorkerBridge.__init__(self)
-        self.new_work_event = lp_signal
-        self.recent_shares_ts_work = []
-        
-        self.my_pubkey_hash = my_pubkey_hash
-        self.net = net
-        self.donation_percentage = donation_percentage
-        self.current_work = current_work
-        self.merged_work = merged_work
-        self.best_share_var = best_share_var
-        self.tracker = tracker
-        self.my_share_hashes = my_share_hashes
-        self.my_doa_share_hashes = my_doa_share_hashes
-        self.worker_fee = worker_fee
-        self.p2p_node = p2p_node
-        self.submit_block = submit_block
-        self.set_best_share = set_best_share
-        self.shared_share_hashes = shared_share_hashes
-        
-        self.pseudoshare_received = variable.Event()
-        self.share_received = variable.Event()
-        self.local_rate_monitor = math.RateMonitor(10*60)
-        
-        self.removed_unstales_var = variable.Variable((0, 0, 0))
-        self.removed_doa_unstales_var = variable.Variable(0)
-        
-        @tracker.verified.removed.watch
-        def _(share):
-            if share.hash in self.my_share_hashes and tracker.is_child_of(share.hash, self.best_share_var.value):
-                assert share.share_data['stale_info'] in [None, 'orphan', 'doa'] # we made these shares in this instance
-                self.removed_unstales_var.set((
-                    self.removed_unstales_var.value[0] + 1,
-                    self.removed_unstales_var.value[1] + (1 if share.share_data['stale_info'] == 'orphan' else 0),
-                    self.removed_unstales_var.value[2] + (1 if share.share_data['stale_info'] == 'doa' else 0),
-                ))
-            if share.hash in self.my_doa_share_hashes and self.tracker.is_child_of(share.hash, self.best_share_var.value):
-                self.removed_doa_unstales_var.set(self.removed_doa_unstales_var.value + 1)
-    
-    def get_stale_counts(self):
-        '''Returns (orphans, doas), total, (orphans_recorded_in_chain, doas_recorded_in_chain)'''
-        my_shares = len(self.my_share_hashes)
-        my_doa_shares = len(self.my_doa_share_hashes)
-        delta = self.tracker.verified.get_delta_to_last(self.best_share_var.value)
-        my_shares_in_chain = delta.my_count + self.removed_unstales_var.value[0]
-        my_doa_shares_in_chain = delta.my_doa_count + self.removed_doa_unstales_var.value
-        orphans_recorded_in_chain = delta.my_orphan_announce_count + self.removed_unstales_var.value[1]
-        doas_recorded_in_chain = delta.my_dead_announce_count + self.removed_unstales_var.value[2]
-        
-        my_shares_not_in_chain = my_shares - my_shares_in_chain
-        my_doa_shares_not_in_chain = my_doa_shares - my_doa_shares_in_chain
-        
-        return (my_shares_not_in_chain - my_doa_shares_not_in_chain, my_doa_shares_not_in_chain), my_shares, (orphans_recorded_in_chain, doas_recorded_in_chain)
-    
-    def get_user_details(self, request):
-        user = request.getUser() if request.getUser() is not None else ''
-        
-        desired_pseudoshare_target = None
-        if '+' in user:
-            user, desired_pseudoshare_difficulty_str = user.rsplit('+', 1)
-            try:
-                desired_pseudoshare_target = bitcoin_data.difficulty_to_target(float(desired_pseudoshare_difficulty_str))
-            except:
-                pass
-        
-        desired_share_target = 2**256 - 1
-        if '/' in user:
-            user, min_diff_str = user.rsplit('/', 1)
-            try:
-                desired_share_target = bitcoin_data.difficulty_to_target(float(min_diff_str))
-            except:
-                pass
-        
-        if random.uniform(0, 100) < self.worker_fee:
-            pubkey_hash = self.my_pubkey_hash
-        else:
-            try:
-                pubkey_hash = bitcoin_data.address_to_pubkey_hash(user, self.net.PARENT)
-            except: # XXX blah
-                pubkey_hash = self.my_pubkey_hash
-        
-        return user, pubkey_hash, desired_share_target, desired_pseudoshare_target
-    
-    def preprocess_request(self, request):
-        user, pubkey_hash, desired_share_target, desired_pseudoshare_target = self.get_user_details(request)
-        return pubkey_hash, desired_share_target, desired_pseudoshare_target
-    
-    def get_work(self, pubkey_hash, desired_share_target, desired_pseudoshare_target):
-        if len(self.p2p_node.peers) == 0 and self.net.PERSIST:
-            raise jsonrpc.Error(-12345, u'p2pool is not connected to any peers')
-        if self.best_share_var.value is None and self.net.PERSIST:
-            raise jsonrpc.Error(-12345, u'p2pool is downloading shares')
-        if time.time() > self.current_work.value['last_update'] + 60:
-            raise jsonrpc.Error(-12345, u'lost contact with bitcoind')
-        
-        if self.merged_work.value:
-            tree, size = bitcoin_data.make_auxpow_tree(self.merged_work.value)
-            mm_hashes = [self.merged_work.value.get(tree.get(i), dict(hash=0))['hash'] for i in xrange(size)]
-            mm_data = '\xfa\xbemm' + bitcoin_data.aux_pow_coinbase_type.pack(dict(
-                merkle_root=bitcoin_data.merkle_hash(mm_hashes),
-                size=size,
-                nonce=0,
-            ))
-            mm_later = [(aux_work, mm_hashes.index(aux_work['hash']), mm_hashes) for chain_id, aux_work in self.merged_work.value.iteritems()]
-        else:
-            mm_data = ''
-            mm_later = []
-        
-        if True:
-            share_info, generate_tx = p2pool_data.Share.generate_transaction(
-                tracker=self.tracker,
-                share_data=dict(
-                    previous_share_hash=self.best_share_var.value,
-                    coinbase=(mm_data + self.current_work.value['coinbaseflags'])[:100],
-                    nonce=random.randrange(2**32),
-                    pubkey_hash=pubkey_hash,
-                    subsidy=self.current_work.value['subsidy'],
-                    donation=math.perfect_round(65535*self.donation_percentage/100),
-                    stale_info=(lambda (orphans, doas), total, (orphans_recorded_in_chain, doas_recorded_in_chain):
-                        'orphan' if orphans > orphans_recorded_in_chain else
-                        'doa' if doas > doas_recorded_in_chain else
-                        None
-                    )(*self.get_stale_counts()),
-                    desired_version=3,
-                ),
-                block_target=self.current_work.value['bits'].target,
-                desired_timestamp=int(time.time() - self.current_work.value['clock_offset']),
-                desired_target=desired_share_target,
-                ref_merkle_link=dict(branch=[], index=0),
-                net=self.net,
-            )
-        
-        if desired_pseudoshare_target is None:
-            target = 2**256-1
-            if len(self.recent_shares_ts_work) == 50:
-                hash_rate = sum(work for ts, work in self.recent_shares_ts_work[1:])//(self.recent_shares_ts_work[-1][0] - self.recent_shares_ts_work[0][0])
-                if hash_rate:
-                    target = min(target, int(2**256/hash_rate))
-        else:
-            target = desired_pseudoshare_target
-        target = max(target, share_info['bits'].target)
-        for aux_work in self.merged_work.value.itervalues():
-            target = max(target, aux_work['target'])
-        target = math.clip(target, self.net.PARENT.SANE_TARGET_RANGE)
-        
-        transactions = [generate_tx] + list(self.current_work.value['transactions'])
-        packed_generate_tx = bitcoin_data.tx_type.pack(generate_tx)
-        merkle_root = bitcoin_data.check_merkle_link(bitcoin_data.hash256(packed_generate_tx), self.current_work.value['merkle_link'])
-        
-        getwork_time = time.time()
-        merkle_link = self.current_work.value['merkle_link']
-        
-        print 'New work for worker! Difficulty: %.06f Share difficulty: %.06f Total block value: %.6f %s including %i transactions' % (
-            bitcoin_data.target_to_difficulty(target),
-            bitcoin_data.target_to_difficulty(share_info['bits'].target),
-            self.current_work.value['subsidy']*1e-8, self.net.PARENT.SYMBOL,
-            len(self.current_work.value['transactions']),
-        )
-        
-        bits = self.current_work.value['bits']
-        previous_block = self.current_work.value['previous_block']
-        ba = bitcoin_getwork.BlockAttempt(
-            version=self.current_work.value['version'],
-            previous_block=self.current_work.value['previous_block'],
-            merkle_root=merkle_root,
-            timestamp=self.current_work.value['time'],
-            bits=self.current_work.value['bits'],
-            share_target=target,
-        )
-        
-        received_header_hashes = set()
-        
-        def got_response(header, request):
-            header_hash = bitcoin_data.hash256(bitcoin_data.block_header_type.pack(header))
-            pow_hash = self.net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(header))
-            try:
-                if pow_hash <= header['bits'].target or p2pool.DEBUG:
-                    self.submit_block(dict(header=header, txs=transactions), ignore_failure=False)
-                    if pow_hash <= header['bits'].target:
-                        print
-                        print 'GOT BLOCK FROM MINER! Passing to bitcoind! %s%064x' % (self.net.PARENT.BLOCK_EXPLORER_URL_PREFIX, header_hash)
-                        print
-            except:
-                log.err(None, 'Error while processing potential block:')
-            
-            user, _, _, _ = self.get_user_details(request)
-            assert header['merkle_root'] == merkle_root
-            assert header['previous_block'] == previous_block
-            assert header['bits'] == bits
-            
-            on_time = self.best_share_var.value == share_info['share_data']['previous_share_hash']
-            
-            for aux_work, index, hashes in mm_later:
-                try:
-                    if pow_hash <= aux_work['target'] or p2pool.DEBUG:
-                        df = deferral.retry('Error submitting merged block: (will retry)', 10, 10)(aux_work['merged_proxy'].rpc_getauxblock)(
-                            pack.IntType(256, 'big').pack(aux_work['hash']).encode('hex'),
-                            bitcoin_data.aux_pow_type.pack(dict(
-                                merkle_tx=dict(
-                                    tx=transactions[0],
-                                    block_hash=header_hash,
-                                    merkle_link=merkle_link,
-                                ),
-                                merkle_link=bitcoin_data.calculate_merkle_link(hashes, index),
-                                parent_block_header=header,
-                            )).encode('hex'),
-                        )
-                        @df.addCallback
-                        def _(result):
-                            if result != (pow_hash <= aux_work['target']):
-                                print >>sys.stderr, 'Merged block submittal result: %s Expected: %s' % (result, pow_hash <= aux_work['target'])
-                            else:
-                                print 'Merged block submittal result: %s' % (result,)
-                        @df.addErrback
-                        def _(err):
-                            log.err(err, 'Error submitting merged block:')
-                except:
-                    log.err(None, 'Error while processing merged mining POW:')
-            
-            if pow_hash <= share_info['bits'].target and header_hash not in received_header_hashes:
-                min_header = dict(header);del min_header['merkle_root']
-                hash_link = p2pool_data.prefix_to_hash_link(packed_generate_tx[:-32-4], p2pool_data.Share.gentx_before_refhash)
-                share = p2pool_data.Share(self.net, None, dict(
-                    min_header=min_header, share_info=share_info, hash_link=hash_link,
-                    ref_merkle_link=dict(branch=[], index=0),
-                ), merkle_link=merkle_link, other_txs=transactions[1:] if pow_hash <= header['bits'].target else None)
-                
-                print 'GOT SHARE! %s %s prev %s age %.2fs%s' % (
-                    request.getUser(),
-                    p2pool_data.format_hash(share.hash),
-                    p2pool_data.format_hash(share.previous_hash),
-                    time.time() - getwork_time,
-                    ' DEAD ON ARRIVAL' if not on_time else '',
-                )
-                self.my_share_hashes.add(share.hash)
-                if not on_time:
-                    self.my_doa_share_hashes.add(share.hash)
-                
-                self.tracker.add(share)
-                if not p2pool.DEBUG:
-                    self.tracker.verified.add(share)
-                self.set_best_share()
-                
-                try:
-                    if pow_hash <= header['bits'].target or p2pool.DEBUG:
-                        for peer in self.p2p_node.peers.itervalues():
-                            peer.sendShares([share])
-                        self.shared_share_hashes.add(share.hash)
-                except:
-                    log.err(None, 'Error forwarding block solution:')
-                
-                self.share_received.happened(bitcoin_data.target_to_average_attempts(share.target), not on_time)
-            
-            if pow_hash > target:
-                print 'Worker %s submitted share with hash > target:' % (request.getUser(),)
-                print '    Hash:   %56x' % (pow_hash,)
-                print '    Target: %56x' % (target,)
-            elif header_hash in received_header_hashes:
-                print >>sys.stderr, 'Worker %s @ %s submitted share more than once!' % (request.getUser(), request.getClientIP())
-            else:
-                received_header_hashes.add(header_hash)
-                
-                self.pseudoshare_received.happened(bitcoin_data.target_to_average_attempts(target), not on_time, user)
-                self.recent_shares_ts_work.append((time.time(), bitcoin_data.target_to_average_attempts(target)))
-                while len(self.recent_shares_ts_work) > 50:
-                    self.recent_shares_ts_work.pop(0)
-                self.local_rate_monitor.add_datum(dict(work=bitcoin_data.target_to_average_attempts(target), dead=not on_time, user=user))
-            
-            return on_time
-        
-        return ba, got_response
-
 @defer.inlineCallbacks
 def main(args, net, datadir_path, merged_urls, worker_endpoint):
     try:
@@ -334,7 +63,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         @deferral.retry('Error while checking Bitcoin connection:', 1)
         @defer.inlineCallbacks
         def check():
-            if not (yield net.PARENT.RPC_CHECK)(bitcoind):
+            if not (yield net.PARENT.RPC_CHECK(bitcoind)):
                 print >>sys.stderr, "    Check failed! Make sure that you're connected to the right bitcoind with --bitcoind-rpc-port!"
                 raise deferral.RetrySilentlyException()
             temp_work = yield getwork(bitcoind)
@@ -352,7 +81,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         task.LoopingCall(poll_height).start(60*60)
         
         print '    ...success!'
-        print '    Current block hash: %x' % (temp_work['previous_block_hash'],)
+        print '    Current block hash: %x' % (temp_work['previous_block'],)
         print '    Current block height: %i' % (block_height_var.value,)
         print
         
@@ -404,24 +133,24 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         print "Loading shares..."
         for i, (mode, contents) in enumerate(ss.get_shares()):
             if mode == 'share':
-                if contents.hash in tracker.shares:
+                if contents.hash in tracker.items:
                     continue
                 shared_share_hashes.add(contents.hash)
                 contents.time_seen = 0
                 tracker.add(contents)
-                if len(tracker.shares) % 1000 == 0 and tracker.shares:
-                    print "    %i" % (len(tracker.shares),)
+                if len(tracker.items) % 1000 == 0 and tracker.items:
+                    print "    %i" % (len(tracker.items),)
             elif mode == 'verified_hash':
                 known_verified.add(contents)
             else:
                 raise AssertionError()
         print "    ...inserting %i verified shares..." % (len(known_verified),)
         for h in known_verified:
-            if h not in tracker.shares:
+            if h not in tracker.items:
                 ss.forget_verified_share(h)
                 continue
-            tracker.verified.add(tracker.shares[h])
-        print "    ...done loading %i shares!" % (len(tracker.shares),)
+            tracker.verified.add(tracker.items[h])
+        print "    ...done loading %i shares!" % (len(tracker.items),)
         print
         tracker.removed.watch(lambda share: ss.forget_share(share.hash))
         tracker.verified.removed.watch(lambda share: ss.forget_verified_share(share.hash))
@@ -432,31 +161,13 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         
         # BITCOIND WORK
         
-        bitcoind_work = variable.Variable(None)
-        
-        @defer.inlineCallbacks
-        def poll_bitcoind():
-            work = yield getwork(bitcoind)
-            bitcoind_work.set(dict(
-                version=work['version'],
-                previous_block=work['previous_block_hash'],
-                bits=work['bits'],
-                coinbaseflags=work['coinbaseflags'],
-                time=work['time'],
-                transactions=work['transactions'],
-                merkle_link=work['merkle_link'],
-                subsidy=work['subsidy'],
-                clock_offset=time.time() - work['time'],
-                last_update=time.time(),
-            ))
-        yield poll_bitcoind()
-        
+        bitcoind_work = variable.Variable((yield getwork(bitcoind)))
         @defer.inlineCallbacks
         def work_poller():
             while True:
                 flag = factory.new_block.get_deferred()
                 try:
-                    yield poll_bitcoind()
+                    bitcoind_work.set((yield getwork(bitcoind)))
                 except:
                     log.err()
                 yield defer.DeferredList([flag, deferral.sleep(15)], fireOnOneCallback=True)
@@ -486,57 +197,6 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         bitcoind_work.changed.watch(lambda _: poll_header())
         yield poll_header()
         
-        # MERGED WORK
-        
-        merged_work = variable.Variable({})
-        
-        @defer.inlineCallbacks
-        def set_merged_work(merged_url, merged_userpass):
-            merged_proxy = jsonrpc.Proxy(merged_url, dict(Authorization='Basic ' + base64.b64encode(merged_userpass)))
-            while True:
-                auxblock = yield deferral.retry('Error while calling merged getauxblock:', 1)(merged_proxy.rpc_getauxblock)()
-                merged_work.set(dict(merged_work.value, **{auxblock['chainid']: dict(
-                    hash=int(auxblock['hash'], 16),
-                    target=pack.IntType(256).unpack(auxblock['target'].decode('hex')),
-                    merged_proxy=merged_proxy,
-                )}))
-                yield deferral.sleep(1)
-        for merged_url, merged_userpass in merged_urls:
-            set_merged_work(merged_url, merged_userpass)
-        
-        @merged_work.changed.watch
-        def _(new_merged_work):
-            print 'Got new merged mining work!'
-        
-        # COMBINE WORK
-        
-        current_work = variable.Variable(None)
-        def compute_work():
-            t = dict(bitcoind_work.value)
-            
-            if (best_block_header.value is not None and
-            best_block_header.value['previous_block'] == t['previous_block'] and
-            net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(best_block_header.value)) <= t['bits'].target):
-                print 'Skipping from block %x to block %x!' % (best_block_header.value['previous_block'],
-                    bitcoin_data.hash256(bitcoin_data.block_header_type.pack(best_block_header.value)))
-                t = dict(
-                    version=best_block_header.value['version'],
-                    previous_block=bitcoin_data.hash256(bitcoin_data.block_header_type.pack(best_block_header.value)),
-                    bits=best_block_header.value['bits'], # not always true
-                    coinbaseflags='',
-                    time=best_block_header.value['timestamp'] + 600, # better way?
-                    transactions=[],
-                    merkle_link=bitcoin_data.calculate_merkle_link([None], 0),
-                    subsidy=net.PARENT.SUBSIDY_FUNC(block_height_var.value),
-                    clock_offset=current_work.value['clock_offset'],
-                    last_update=current_work.value['last_update'],
-                )
-            
-            current_work.set(t)
-        bitcoind_work.changed.watch(lambda _: compute_work())
-        best_block_header.changed.watch(lambda _: compute_work())
-        compute_work()
-        
         # BEST SHARE
         
         get_height_rel_highest = yield height_tracker.get_height_rel_highest_func(bitcoind, factory, lambda: bitcoind_work.value['previous_block'], net)
@@ -579,18 +239,6 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         bitcoind_work.changed.watch(lambda _: set_best_share())
         set_best_share()
         
-        # LONG POLLING
-        
-        lp_signal = variable.Event()
-        
-        @current_work.transitioned.watch
-        def _(before, after):
-            # trigger LP if version/previous_block/bits changed or transactions changed from nothing
-            if any(before[x] != after[x] for x in ['version', 'previous_block', 'bits']) or (not before['transactions'] and after['transactions']):
-                lp_signal.happened()
-        merged_work.changed.watch(lambda _: lp_signal.happened())
-        best_share_var.changed.watch(lambda _: lp_signal.happened())
-        
         
         print '    ...success!'
         print
@@ -604,7 +252,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                 
                 new_count = 0
                 for share in shares:
-                    if share.hash in tracker.shares:
+                    if share.hash in tracker.items:
                         #print 'Got duplicate share, ignoring. Hash: %s' % (p2pool_data.format_hash(share.hash),)
                         continue
                     
@@ -621,13 +269,13 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                     set_best_share()
                 
                 if len(shares) > 5:
-                    print '... done processing %i shares. New: %i Have: %i/~%i' % (len(shares), new_count, len(tracker.shares), 2*net.CHAIN_LENGTH)
+                    print '... done processing %i shares. New: %i Have: %i/~%i' % (len(shares), new_count, len(tracker.items), 2*net.CHAIN_LENGTH)
             
             def handle_share_hashes(self, hashes, peer):
                 t = time.time()
                 get_hashes = []
                 for share_hash in hashes:
-                    if share_hash in tracker.shares:
+                    if share_hash in tracker.items:
                         continue
                     last_request_time, count = requested.get(share_hash, (None, 0))
                     if last_request_time is not None and last_request_time - 5 < t < last_request_time + 10 * 1.5**count:
@@ -686,7 +334,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                 print
                 def spread():
                     if (get_height_rel_highest(share.header['previous_block']) > -5 or
-                        current_work.value['previous_block'] in [share.header['previous_block'], share.header_hash]):
+                        bitcoind_work.value['previous_block'] in [share.header['previous_block'], share.header_hash]):
                         broadcast_share(share.hash)
                 spread()
                 reactor.callLater(5, spread) # so get_height_rel_highest can update
@@ -708,11 +356,6 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                     addrs.update(dict((tuple(k), v) for k, v in json.loads(f.read())))
             except:
                 print >>sys.stderr, 'error parsing addrs'
-        elif os.path.exists(os.path.join(datadir_path, 'addrs.txt')):
-            try:
-                addrs.update(dict(eval(x) for x in open(os.path.join(datadir_path, 'addrs.txt'))))
-            except:
-                print >>sys.stderr, "error reading addrs.txt"
         for addr_df in map(parse, net.BOOTSTRAP_ADDRS):
             try:
                 addr = yield addr_df
@@ -748,6 +391,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
             for peer in p2p_node.peers.itervalues():
                 peer.send_bestblock(header=header)
         
+        @defer.inlineCallbacks
         def broadcast_share(share_hash):
             shares = []
             for share in tracker.get_chain(share_hash, min(5, tracker.get_height(share_hash))):
@@ -756,8 +400,8 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                 shared_share_hashes.add(share.hash)
                 shares.append(share)
             
-            for peer in p2p_node.peers.itervalues():
-                peer.sendShares([share for share in shares if share.peer is not peer])
+            for peer in list(p2p_node.peers.itervalues()):
+                yield peer.sendShares([share for share in shares if share.peer is not peer])
         
         # send share when the chain changes to their chain
         best_share_var.changed.watch(broadcast_share)
@@ -765,7 +409,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         def save_shares():
             for share in tracker.get_chain(best_share_var.value, min(tracker.get_height(best_share_var.value), 2*net.CHAIN_LENGTH)):
                 ss.add_share(share)
-                if share.hash in tracker.verified.shares:
+                if share.hash in tracker.verified.items:
                     ss.add_verified_hash(share.hash)
         task.LoopingCall(save_shares).start(60)
         
@@ -793,10 +437,10 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         
         print 'Listening for workers on %r port %i...' % (worker_endpoint[0], worker_endpoint[1])
         
-        get_current_txouts = lambda: p2pool_data.get_expected_payouts(tracker, best_share_var.value, current_work.value['bits'].target, current_work.value['subsidy'], net)
+        get_current_txouts = lambda: p2pool_data.get_expected_payouts(tracker, best_share_var.value, bitcoind_work.value['bits'].target, bitcoind_work.value['subsidy'], net)
         
-        wb = WorkerBridge(lp_signal, my_pubkey_hash, net, args.donation_percentage, current_work, merged_work, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, args.worker_fee, p2p_node, submit_block, set_best_share, shared_share_hashes)
-        web_root = web.get_web_root(tracker, current_work, get_current_txouts, datadir_path, net, wb.get_stale_counts, my_pubkey_hash, wb.local_rate_monitor, args.worker_fee, p2p_node, wb.my_share_hashes, wb.pseudoshare_received, wb.share_received, best_share_var)
+        wb = work.WorkerBridge(my_pubkey_hash, net, args.donation_percentage, bitcoind_work, best_block_header, merged_urls, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, args.worker_fee, p2p_node, submit_block, set_best_share, broadcast_share, block_height_var)
+        web_root = web.get_web_root(tracker, bitcoind_work, get_current_txouts, datadir_path, net, wb.get_stale_counts, my_pubkey_hash, wb.local_rate_monitor, args.worker_fee, p2p_node, wb.my_share_hashes, wb.pseudoshare_received, wb.share_received, best_share_var)
         worker_interface.WorkerInterface(wb).attach_to(web_root, get_handler=lambda request: request.redirect('/static/'))
         
         deferral.retry('Error binding to worker port:', traceback=False)(reactor.listenTCP)(worker_endpoint[1], server.Site(web_root), interface=worker_endpoint[0])
@@ -872,14 +516,14 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
             while True:
                 yield deferral.sleep(3)
                 try:
-                    if time.time() > current_work.value['last_update'] + 60:
-                        print >>sys.stderr, '''---> LOST CONTACT WITH BITCOIND for %s! Check that it isn't frozen or dead! <---''' % (math.format_dt(time.time() - current_work.value['last_update']),)
+                    if time.time() > bitcoind_work.value['last_update'] + 60:
+                        print >>sys.stderr, '''---> LOST CONTACT WITH BITCOIND for %s! Check that it isn't frozen or dead! <---''' % (math.format_dt(time.time() - bitcoind_work.value['last_update']),)
                     
                     height = tracker.get_height(best_share_var.value)
                     this_str = 'P2Pool: %i shares in chain (%i verified/%i total) Peers: %i (%i incoming)' % (
                         height,
-                        len(tracker.verified.shares),
-                        len(tracker.shares),
+                        len(tracker.verified.items),
+                        len(tracker.items),
                         len(p2p_node.peers),
                         sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming),
                     ) + (' FDs: %i R/%i W' % (len(reactor.getReaders()), len(reactor.getWriters())) if p2pool.DEBUG else '')
@@ -890,7 +534,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                         math.format(int(my_att_s)),
                         math.format_dt(dt),
                         math.format_binomial_conf(sum(1 for datum in datums if datum['dead']), len(datums), 0.95),
-                        math.format_dt(2**256 / tracker.shares[best_share_var.value].max_target / my_att_s) if my_att_s and best_share_var.value else '???',
+                        math.format_dt(2**256 / tracker.items[best_share_var.value].max_target / my_att_s) if my_att_s and best_share_var.value else '???',
                     )
                     
                     if height > 2:
@@ -907,7 +551,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                         this_str += '\n Pool: %sH/s Stale rate: %.1f%% Expected time to block: %s' % (
                             math.format(int(real_att_s)),
                             100*stale_prop,
-                            math.format_dt(2**256 / current_work.value['bits'].target / real_att_s),
+                            math.format_dt(2**256 / bitcoind_work.value['bits'].target / real_att_s),
                         )
                         
                         for warning in p2pool_data.get_warnings(tracker, best_share_var.value, net):
@@ -927,7 +571,7 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         log.err(None, 'Fatal error:')
 
 def run():
-    realnets=dict((name, net) for name, net in networks.nets.iteritems() if '_testnet' not in name)
+    realnets = dict((name, net) for name, net in networks.nets.iteritems() if '_testnet' not in name)
     
     parser = fixargparse.FixedArgumentParser(description='p2pool (version %s)' % (p2pool.__version__,), fromfile_prefix_chars='@')
     parser.add_argument('--version', action='version', version=p2pool.__version__)
@@ -1006,6 +650,7 @@ def run():
     
     if args.debug:
         p2pool.DEBUG = True
+        defer.setDebugging(True)
     
     net_name = args.net_name + ('_testnet' if args.testnet else '')
     net = networks.nets[net_name]