moved all web related stuff into p2pool.web
authorForrest Voight <forrest.voight@gmail.com>
Wed, 22 Feb 2012 08:39:36 +0000 (03:39 -0500)
committerForrest Voight <forrest.voight@gmail.com>
Wed, 22 Feb 2012 08:39:36 +0000 (03:39 -0500)
p2pool/main.py
p2pool/web.py [new file with mode: 0644]

index 7ef534d..4245844 100644 (file)
@@ -8,20 +8,19 @@ import random
 import struct
 import sys
 import time
-import json
 import signal
 import traceback
 import urlparse
 
 from twisted.internet import defer, error, reactor, protocol, task
-from twisted.web import server, resource
+from twisted.web import server
 from twisted.python import log
 from nattraverso import portmapper, ipdiscover
 
 import bitcoin.p2p as bitcoin_p2p, bitcoin.getwork as bitcoin_getwork, bitcoin.data as bitcoin_data
 from bitcoin import worker_interface
 from util import expiring_dict, jsonrpc, variable, deferral, math, logging, pack
-from . import p2p, networks, graphs
+from . import p2p, networks, web
 import p2pool, p2pool.data as p2pool_data
 
 @deferral.retry('Error getting work from bitcoind:', 3)
@@ -51,12 +50,6 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
     try:
         print 'p2pool (version %s)' % (p2pool.__version__,)
         print
-        try:
-            from . import draw
-        except ImportError:
-            draw = None
-            print "Install Pygame and PIL to enable visualizations! Visualizations disabled."
-            print
         
         # connect to bitcoind over JSON-RPC and do initial getmemorypool
         url = 'http://%s:%i/' % (args.bitcoind_address, args.bitcoind_rpc_port)
@@ -384,8 +377,6 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
         print '    ...success!'
         print
         
-        start_time = time.time()
-        
         @defer.inlineCallbacks
         def upnp_thread():
             while True:
@@ -679,22 +670,6 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
                 
                 return ba, got_response
         
-        web_root = resource.Resource()
-        worker_interface.WorkerInterface(WorkerBridge()).attach_to(web_root)
-        
-        def get_rate():
-            if tracker.get_height(current_work.value['best_share_hash']) < 720:
-                return json.dumps(None)
-            return json.dumps(p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], 720)
-                / (1 - p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], 720)))
-        
-        def get_users():
-            height, last = tracker.get_height_and_last(current_work.value['best_share_hash'])
-            weights, total_weight, donation_weight = tracker.get_cumulative_weights(current_work.value['best_share_hash'], min(height, 720), 65535*2**256)
-            res = {}
-            for script in sorted(weights, key=lambda s: weights[s]):
-                res[bitcoin_data.script2_to_human(script, net.PARENT)] = weights[script]/total_weight
-            return json.dumps(res)
         
         def get_current_txouts():
             weights, total_weight, donation_weight = tracker.get_cumulative_weights(current_work.value['best_share_hash'],
@@ -705,202 +680,8 @@ def main(args, net, datadir_path, merged_urls, worker_endpoint):
             res[p2pool_data.DONATION_SCRIPT] = res.get(p2pool_data.DONATION_SCRIPT, 0) + current_work2.value['subsidy'] - sum(res.itervalues())
             return res
         
-        def get_current_scaled_txouts(scale, trunc=0):
-            txouts = get_current_txouts()
-            total = sum(txouts.itervalues())
-            results = dict((script, value*scale//total) for script, value in txouts.iteritems())
-            if trunc > 0:
-                total_random = 0
-                random_set = set()
-                for s in sorted(results, key=results.__getitem__):
-                    if results[s] >= trunc:
-                        break
-                    total_random += results[s]
-                    random_set.add(s)
-                if total_random:
-                    winner = math.weighted_choice((script, results[script]) for script in random_set)
-                    for script in random_set:
-                        del results[script]
-                    results[winner] = total_random
-            if sum(results.itervalues()) < int(scale):
-                results[math.weighted_choice(results.iteritems())] += int(scale) - sum(results.itervalues())
-            return results
-        
-        def get_current_payouts():
-            return json.dumps(dict((bitcoin_data.script2_to_human(script, net.PARENT), value/1e8) for script, value in get_current_txouts().iteritems()))
-        
-        def get_patron_sendmany(this):
-            try:
-                if '/' in this:
-                    this, trunc = this.split('/', 1)
-                else:
-                    trunc = '0.01'
-                return json.dumps(dict(
-                    (bitcoin_data.script2_to_address(script, net.PARENT), value/1e8)
-                    for script, value in get_current_scaled_txouts(scale=int(float(this)*1e8), trunc=int(float(trunc)*1e8)).iteritems()
-                    if bitcoin_data.script2_to_address(script, net.PARENT) is not None
-                ))
-            except:
-                return json.dumps(None)
-        
-        def get_global_stats():
-            # averaged over last hour
-            lookbehind = 3600//net.SHARE_PERIOD
-            if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
-                return None
-            
-            nonstale_hash_rate = p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], lookbehind)
-            stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
-            return json.dumps(dict(
-                pool_nonstale_hash_rate=nonstale_hash_rate,
-                pool_hash_rate=nonstale_hash_rate/(1 - stale_prop),
-                pool_stale_prop=stale_prop,
-            ))
-        
-        def get_local_stats():
-            lookbehind = 3600//net.SHARE_PERIOD
-            if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
-                return None
-            
-            global_stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
-            
-            my_unstale_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes)
-            my_orphan_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes and share.share_data['stale_info'] == 253)
-            my_doa_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes and share.share_data['stale_info'] == 254)
-            my_share_count = my_unstale_count + my_orphan_count + my_doa_count
-            my_stale_count = my_orphan_count + my_doa_count
-            
-            my_stale_prop = my_stale_count/my_share_count if my_share_count != 0 else None
-            
-            my_work = sum(bitcoin_data.target_to_average_attempts(share.target)
-                for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind - 1)
-                if share.hash in my_share_hashes)
-            actual_time = (tracker.shares[current_work.value['best_share_hash']].timestamp -
-                tracker.shares[tracker.get_nth_parent_hash(current_work.value['best_share_hash'], lookbehind - 1)].timestamp)
-            share_att_s = my_work / actual_time
-            
-            miner_hash_rates = {}
-            miner_dead_hash_rates = {}
-            datums, dt = local_rate_monitor.get_datums_in_last()
-            for datum in datums:
-                miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt
-                if datum['dead']:
-                    miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
-            
-            return json.dumps(dict(
-                my_hash_rates_in_last_hour=dict(
-                    note="DEPRECATED",
-                    nonstale=share_att_s,
-                    rewarded=share_att_s/(1 - global_stale_prop),
-                    actual=share_att_s/(1 - my_stale_prop) if my_stale_prop is not None else 0, # 0 because we don't have any shares anyway
-                ),
-                my_share_counts_in_last_hour=dict(
-                    shares=my_share_count,
-                    unstale_shares=my_unstale_count,
-                    stale_shares=my_stale_count,
-                    orphan_stale_shares=my_orphan_count,
-                    doa_stale_shares=my_doa_count,
-                ),
-                my_stale_proportions_in_last_hour=dict(
-                    stale=my_stale_prop,
-                    orphan_stale=my_orphan_count/my_share_count if my_share_count != 0 else None,
-                    dead_stale=my_doa_count/my_share_count if my_share_count != 0 else None,
-                ),
-                miner_hash_rates=miner_hash_rates,
-                miner_dead_hash_rates=miner_dead_hash_rates,
-            ))
-        
-        def get_peer_addresses():
-            return ' '.join(peer.transport.getPeer().host + (':' + str(peer.transport.getPeer().port) if peer.transport.getPeer().port != net.P2P_PORT else '') for peer in p2p_node.peers.itervalues())
-        
-        def get_uptime():
-            return json.dumps(time.time() - start_time)
-        
-        class WebInterface(resource.Resource):
-            def __init__(self, func, mime_type, *fields):
-                self.func, self.mime_type, self.fields = func, mime_type, fields
-            
-            def render_GET(self, request):
-                request.setHeader('Content-Type', self.mime_type)
-                request.setHeader('Access-Control-Allow-Origin', '*')
-                return self.func(*(request.args[field][0] for field in self.fields))
-        
-        web_root.putChild('rate', WebInterface(get_rate, 'application/json'))
-        web_root.putChild('users', WebInterface(get_users, 'application/json'))
-        web_root.putChild('fee', WebInterface(lambda: json.dumps(args.worker_fee), 'application/json'))
-        web_root.putChild('current_payouts', WebInterface(get_current_payouts, 'application/json'))
-        web_root.putChild('patron_sendmany', WebInterface(get_patron_sendmany, 'text/plain', 'total'))
-        web_root.putChild('global_stats', WebInterface(get_global_stats, 'application/json'))
-        web_root.putChild('local_stats', WebInterface(get_local_stats, 'application/json'))
-        web_root.putChild('peer_addresses', WebInterface(get_peer_addresses, 'text/plain'))
-        web_root.putChild('payout_addr', WebInterface(lambda: json.dumps(bitcoin_data.pubkey_hash_to_address(my_pubkey_hash, net.PARENT)), 'application/json'))
-        web_root.putChild('recent_blocks', WebInterface(lambda: json.dumps(recent_blocks), 'application/json'))
-        web_root.putChild('uptime', WebInterface(get_uptime, 'application/json'))
-        if draw is not None:
-            web_root.putChild('chain_img', WebInterface(lambda: draw.get(tracker, current_work.value['best_share_hash']), 'image/png'))
-        
-        new_root = resource.Resource()
-        web_root.putChild('web', new_root)
-        
-        stat_log = []
-        if os.path.exists(os.path.join(datadir_path, 'stats')):
-            try:
-                with open(os.path.join(datadir_path, 'stats'), 'rb') as f:
-                    stat_log = json.loads(f.read())
-            except:
-                log.err(None, 'Error loading stats:')
-        def update_stat_log():
-            while stat_log and stat_log[0]['time'] < time.time() - 24*60*60:
-                stat_log.pop(0)
-            
-            lookbehind = 3600//net.SHARE_PERIOD
-            if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
-                return None
-            
-            global_stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
-            (stale_orphan_shares, stale_doa_shares), shares, _ = get_stale_counts()
-            
-            miner_hash_rates = {}
-            miner_dead_hash_rates = {}
-            datums, dt = local_rate_monitor.get_datums_in_last()
-            for datum in datums:
-                miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt
-                if datum['dead']:
-                    miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
-            
-            stat_log.append(dict(
-                time=time.time(),
-                pool_hash_rate=p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], lookbehind)/(1-global_stale_prop),
-                pool_stale_prop=global_stale_prop,
-                local_hash_rates=miner_hash_rates,
-                local_dead_hash_rates=miner_dead_hash_rates,
-                shares=shares,
-                stale_shares=stale_orphan_shares + stale_doa_shares,
-                stale_shares_breakdown=dict(orphan=stale_orphan_shares, doa=stale_doa_shares),
-                current_payout=get_current_txouts().get(bitcoin_data.pubkey_hash_to_script2(my_pubkey_hash), 0)*1e-8,
-                peers=dict(
-                    incoming=sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming),
-                    outgoing=sum(1 for peer in p2p_node.peers.itervalues() if not peer.incoming),
-                ),
-                attempts_to_share=bitcoin_data.target_to_average_attempts(tracker.shares[current_work.value['best_share_hash']].max_target),
-                attempts_to_block=bitcoin_data.target_to_average_attempts(current_work.value['bits'].target),
-                block_value=current_work2.value['subsidy']*1e-8,
-            ))
-            
-            with open(os.path.join(datadir_path, 'stats'), 'wb') as f:
-                f.write(json.dumps(stat_log))
-        task.LoopingCall(update_stat_log).start(5*60)
-        new_root.putChild('log', WebInterface(lambda: json.dumps(stat_log), 'application/json'))
-        
-        grapher = graphs.Grapher(os.path.join(datadir_path, 'rrd'))
-        web_root.putChild('graphs', grapher.get_resource())
-        def add_point():
-            if tracker.get_height(current_work.value['best_share_hash']) < 720:
-                return
-            nonstalerate = p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], 720)
-            poolrate = nonstalerate / (1 - p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], 720))
-            grapher.add_poolrate_point(poolrate, poolrate - nonstalerate)
-        task.LoopingCall(add_point).start(100)
+        web_root = web.get_web_root(tracker, current_work, current_work2, get_current_txouts, datadir_path, net, get_stale_counts, my_pubkey_hash, local_rate_monitor, args.worker_fee, p2p_node, my_share_hashes, recent_blocks)
+        worker_interface.WorkerInterface(WorkerBridge()).attach_to(web_root)
         
         def attempt_listen():
             try:
diff --git a/p2pool/web.py b/p2pool/web.py
new file mode 100644 (file)
index 0000000..35d111c
--- /dev/null
@@ -0,0 +1,234 @@
+import json
+import os
+import time
+
+from twisted.internet import task
+from twisted.python import log
+from twisted.web import resource
+
+from bitcoin import data as bitcoin_data
+from . import data as p2pool_data, graphs
+from util import math
+
+def get_web_root(tracker, current_work, current_work2, get_current_txouts, datadir_path, net, get_stale_counts, my_pubkey_hash, local_rate_monitor, worker_fee, p2p_node, my_share_hashes, recent_blocks):
+    start_time = time.time()
+    
+    web_root = resource.Resource()
+    
+    def get_rate():
+        if tracker.get_height(current_work.value['best_share_hash']) < 720:
+            return json.dumps(None)
+        return json.dumps(p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], 720)
+            / (1 - p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], 720)))
+    
+    def get_users():
+        height, last = tracker.get_height_and_last(current_work.value['best_share_hash'])
+        weights, total_weight, donation_weight = tracker.get_cumulative_weights(current_work.value['best_share_hash'], min(height, 720), 65535*2**256)
+        res = {}
+        for script in sorted(weights, key=lambda s: weights[s]):
+            res[bitcoin_data.script2_to_human(script, net.PARENT)] = weights[script]/total_weight
+        return json.dumps(res)
+    
+    def get_current_scaled_txouts(scale, trunc=0):
+        txouts = get_current_txouts()
+        total = sum(txouts.itervalues())
+        results = dict((script, value*scale//total) for script, value in txouts.iteritems())
+        if trunc > 0:
+            total_random = 0
+            random_set = set()
+            for s in sorted(results, key=results.__getitem__):
+                if results[s] >= trunc:
+                    break
+                total_random += results[s]
+                random_set.add(s)
+            if total_random:
+                winner = math.weighted_choice((script, results[script]) for script in random_set)
+                for script in random_set:
+                    del results[script]
+                results[winner] = total_random
+        if sum(results.itervalues()) < int(scale):
+            results[math.weighted_choice(results.iteritems())] += int(scale) - sum(results.itervalues())
+        return results
+    
+    def get_current_payouts():
+        return json.dumps(dict((bitcoin_data.script2_to_human(script, net.PARENT), value/1e8) for script, value in get_current_txouts().iteritems()))
+    
+    def get_patron_sendmany(this):
+        try:
+            if '/' in this:
+                this, trunc = this.split('/', 1)
+            else:
+                trunc = '0.01'
+            return json.dumps(dict(
+                (bitcoin_data.script2_to_address(script, net.PARENT), value/1e8)
+                for script, value in get_current_scaled_txouts(scale=int(float(this)*1e8), trunc=int(float(trunc)*1e8)).iteritems()
+                if bitcoin_data.script2_to_address(script, net.PARENT) is not None
+            ))
+        except:
+            log.err()
+            return json.dumps(None)
+    
+    def get_global_stats():
+        # averaged over last hour
+        lookbehind = 3600//net.SHARE_PERIOD
+        if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
+            return None
+        
+        nonstale_hash_rate = p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], lookbehind)
+        stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
+        return json.dumps(dict(
+            pool_nonstale_hash_rate=nonstale_hash_rate,
+            pool_hash_rate=nonstale_hash_rate/(1 - stale_prop),
+            pool_stale_prop=stale_prop,
+        ))
+    
+    def get_local_stats():
+        lookbehind = 3600//net.SHARE_PERIOD
+        if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
+            return None
+        
+        global_stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
+        
+        my_unstale_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes)
+        my_orphan_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes and share.share_data['stale_info'] == 253)
+        my_doa_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes and share.share_data['stale_info'] == 254)
+        my_share_count = my_unstale_count + my_orphan_count + my_doa_count
+        my_stale_count = my_orphan_count + my_doa_count
+        
+        my_stale_prop = my_stale_count/my_share_count if my_share_count != 0 else None
+        
+        my_work = sum(bitcoin_data.target_to_average_attempts(share.target)
+            for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind - 1)
+            if share.hash in my_share_hashes)
+        actual_time = (tracker.shares[current_work.value['best_share_hash']].timestamp -
+            tracker.shares[tracker.get_nth_parent_hash(current_work.value['best_share_hash'], lookbehind - 1)].timestamp)
+        share_att_s = my_work / actual_time
+        
+        miner_hash_rates = {}
+        miner_dead_hash_rates = {}
+        datums, dt = local_rate_monitor.get_datums_in_last()
+        for datum in datums:
+            miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt
+            if datum['dead']:
+                miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
+        
+        return json.dumps(dict(
+            my_hash_rates_in_last_hour=dict(
+                note="DEPRECATED",
+                nonstale=share_att_s,
+                rewarded=share_att_s/(1 - global_stale_prop),
+                actual=share_att_s/(1 - my_stale_prop) if my_stale_prop is not None else 0, # 0 because we don't have any shares anyway
+            ),
+            my_share_counts_in_last_hour=dict(
+                shares=my_share_count,
+                unstale_shares=my_unstale_count,
+                stale_shares=my_stale_count,
+                orphan_stale_shares=my_orphan_count,
+                doa_stale_shares=my_doa_count,
+            ),
+            my_stale_proportions_in_last_hour=dict(
+                stale=my_stale_prop,
+                orphan_stale=my_orphan_count/my_share_count if my_share_count != 0 else None,
+                dead_stale=my_doa_count/my_share_count if my_share_count != 0 else None,
+            ),
+            miner_hash_rates=miner_hash_rates,
+            miner_dead_hash_rates=miner_dead_hash_rates,
+        ))
+    
+    def get_peer_addresses():
+        return ' '.join(peer.transport.getPeer().host + (':' + str(peer.transport.getPeer().port) if peer.transport.getPeer().port != net.P2P_PORT else '') for peer in p2p_node.peers.itervalues())
+    
+    def get_uptime():
+        return json.dumps(time.time() - start_time)
+    
+    class WebInterface(resource.Resource):
+        def __init__(self, func, mime_type, *fields):
+            self.func, self.mime_type, self.fields = func, mime_type, fields
+        
+        def render_GET(self, request):
+            request.setHeader('Content-Type', self.mime_type)
+            request.setHeader('Access-Control-Allow-Origin', '*')
+            return self.func(*(request.args[field][0] for field in self.fields))
+    
+    web_root.putChild('rate', WebInterface(get_rate, 'application/json'))
+    web_root.putChild('users', WebInterface(get_users, 'application/json'))
+    web_root.putChild('fee', WebInterface(lambda: json.dumps(worker_fee), 'application/json'))
+    web_root.putChild('current_payouts', WebInterface(get_current_payouts, 'application/json'))
+    web_root.putChild('patron_sendmany', WebInterface(get_patron_sendmany, 'text/plain', 'total'))
+    web_root.putChild('global_stats', WebInterface(get_global_stats, 'application/json'))
+    web_root.putChild('local_stats', WebInterface(get_local_stats, 'application/json'))
+    web_root.putChild('peer_addresses', WebInterface(get_peer_addresses, 'text/plain'))
+    web_root.putChild('payout_addr', WebInterface(lambda: json.dumps(bitcoin_data.pubkey_hash_to_address(my_pubkey_hash, net.PARENT)), 'application/json'))
+    web_root.putChild('recent_blocks', WebInterface(lambda: json.dumps(recent_blocks), 'application/json'))
+    web_root.putChild('uptime', WebInterface(get_uptime, 'application/json'))
+    
+    try:
+        from . import draw
+        web_root.putChild('chain_img', WebInterface(lambda: draw.get(tracker, current_work.value['best_share_hash']), 'image/png'))
+    except ImportError:
+        print "Install Pygame and PIL to enable visualizations! Visualizations disabled."
+    
+    new_root = resource.Resource()
+    web_root.putChild('web', new_root)
+    
+    stat_log = []
+    if os.path.exists(os.path.join(datadir_path, 'stats')):
+        try:
+            with open(os.path.join(datadir_path, 'stats'), 'rb') as f:
+                stat_log = json.loads(f.read())
+        except:
+            log.err(None, 'Error loading stats:')
+    def update_stat_log():
+        while stat_log and stat_log[0]['time'] < time.time() - 24*60*60:
+            stat_log.pop(0)
+        
+        lookbehind = 3600//net.SHARE_PERIOD
+        if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
+            return None
+        
+        global_stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
+        (stale_orphan_shares, stale_doa_shares), shares, _ = get_stale_counts()
+        
+        miner_hash_rates = {}
+        miner_dead_hash_rates = {}
+        datums, dt = local_rate_monitor.get_datums_in_last()
+        for datum in datums:
+            miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt
+            if datum['dead']:
+                miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
+        
+        stat_log.append(dict(
+            time=time.time(),
+            pool_hash_rate=p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], lookbehind)/(1-global_stale_prop),
+            pool_stale_prop=global_stale_prop,
+            local_hash_rates=miner_hash_rates,
+            local_dead_hash_rates=miner_dead_hash_rates,
+            shares=shares,
+            stale_shares=stale_orphan_shares + stale_doa_shares,
+            stale_shares_breakdown=dict(orphan=stale_orphan_shares, doa=stale_doa_shares),
+            current_payout=get_current_txouts().get(bitcoin_data.pubkey_hash_to_script2(my_pubkey_hash), 0)*1e-8,
+            peers=dict(
+                incoming=sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming),
+                outgoing=sum(1 for peer in p2p_node.peers.itervalues() if not peer.incoming),
+            ),
+            attempts_to_share=bitcoin_data.target_to_average_attempts(tracker.shares[current_work.value['best_share_hash']].max_target),
+            attempts_to_block=bitcoin_data.target_to_average_attempts(current_work.value['bits'].target),
+            block_value=current_work2.value['subsidy']*1e-8,
+        ))
+        
+        with open(os.path.join(datadir_path, 'stats'), 'wb') as f:
+            f.write(json.dumps(stat_log))
+    task.LoopingCall(update_stat_log).start(5*60)
+    new_root.putChild('log', WebInterface(lambda: json.dumps(stat_log), 'application/json'))
+    
+    grapher = graphs.Grapher(os.path.join(datadir_path, 'rrd'))
+    web_root.putChild('graphs', grapher.get_resource())
+    def add_point():
+        if tracker.get_height(current_work.value['best_share_hash']) < 720:
+            return
+        nonstalerate = p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], 720)
+        poolrate = nonstalerate / (1 - p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], 720))
+        grapher.add_poolrate_point(poolrate, poolrate - nonstalerate)
+    task.LoopingCall(add_point).start(100)
+    
+    return web_root