X-Git-Url: https://git.novaco.in/?a=blobdiff_plain;f=p2pool%2Fweb.py;h=4400e7d1f1de2eea2787cc7fbfa8d4c34a31038e;hb=391787fd6661af8ec0e24ff61c8c6364da9708b9;hp=1cbd6690ba039db73ca7d54bcb0f333afcb0fa78;hpb=6c5ca49d46cea44fa23115fa04f5612053221ccb;p=p2pool.git
diff --git a/p2pool/web.py b/p2pool/web.py
index 1cbd669..4400e7d 100644
--- a/p2pool/web.py
+++ b/p2pool/web.py
@@ -1,40 +1,66 @@
from __future__ import division
-import cgi
+import errno
import json
import os
+import sys
import time
-import types
+import traceback
-from twisted.internet import reactor, task
+from twisted.internet import defer, reactor
from twisted.python import log
-from twisted.web import resource
+from twisted.web import resource, static
+import p2pool
from bitcoin import data as bitcoin_data
-from . import data as p2pool_data, graphs
-from util import math
+from . import data as p2pool_data, p2p
+from util import deferral, deferred_resource, graph, math, memory, pack, variable
-def get_web_root(tracker, current_work, current_work2, get_current_txouts, datadir_path, net, get_stale_counts, my_pubkey_hash, local_rate_monitor, worker_fee, p2p_node, my_share_hashes, recent_blocks, pseudoshare_received):
+def _atomic_read(filename):
+ try:
+ with open(filename, 'rb') as f:
+ return f.read()
+ except IOError, e:
+ if e.errno != errno.ENOENT:
+ raise
+ try:
+ with open(filename + '.new', 'rb') as f:
+ return f.read()
+ except IOError, e:
+ if e.errno != errno.ENOENT:
+ raise
+ return None
+
+def _atomic_write(filename, data):
+ with open(filename + '.new', 'wb') as f:
+ f.write(data)
+ f.flush()
+ try:
+ os.fsync(f.fileno())
+ except:
+ pass
+ try:
+ os.rename(filename + '.new', filename)
+ except: # XXX windows can't overwrite
+ os.remove(filename)
+ os.rename(filename + '.new', filename)
+
+def get_web_root(wb, datadir_path, bitcoind_getinfo_var, stop_event=variable.Event()):
+ node = wb.node
start_time = time.time()
web_root = resource.Resource()
- def get_rate():
- if tracker.get_height(current_work.value['best_share_hash']) < 720:
- return json.dumps(None)
- return json.dumps(p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], 720)
- / (1 - p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], 720)))
-
def get_users():
- height, last = tracker.get_height_and_last(current_work.value['best_share_hash'])
- weights, total_weight, donation_weight = tracker.get_cumulative_weights(current_work.value['best_share_hash'], min(height, 720), 65535*2**256)
+ height, last = node.tracker.get_height_and_last(node.best_share_var.value)
+ weights, total_weight, donation_weight = node.tracker.get_cumulative_weights(node.best_share_var.value, min(height, 720), 65535*2**256)
res = {}
for script in sorted(weights, key=lambda s: weights[s]):
- res[bitcoin_data.script2_to_human(script, net.PARENT)] = weights[script]/total_weight
- return json.dumps(res)
+ res[bitcoin_data.script2_to_address(script, node.net.PARENT)] = weights[script]/total_weight
+ return res
def get_current_scaled_txouts(scale, trunc=0):
- txouts = get_current_txouts()
+ txouts = node.get_current_txouts()
total = sum(txouts.itervalues())
results = dict((script, value*scale//total) for script, value in txouts.iteritems())
if trunc > 0:
@@ -54,69 +80,58 @@ def get_web_root(tracker, current_work, current_work2, get_current_txouts, datad
results[math.weighted_choice(results.iteritems())] += int(scale) - sum(results.itervalues())
return results
- def get_current_payouts():
- return json.dumps(dict((bitcoin_data.script2_to_human(script, net.PARENT), value/1e8) for script, value in get_current_txouts().iteritems()))
-
- def get_patron_sendmany(this):
- try:
- if '/' in this:
- this, trunc = this.split('/', 1)
- else:
- trunc = '0.01'
- return json.dumps(dict(
- (bitcoin_data.script2_to_address(script, net.PARENT), value/1e8)
- for script, value in get_current_scaled_txouts(scale=int(float(this)*1e8), trunc=int(float(trunc)*1e8)).iteritems()
- if bitcoin_data.script2_to_address(script, net.PARENT) is not None
- ))
- except:
- log.err()
- return json.dumps(None)
+ def get_patron_sendmany(total=None, trunc='0.01'):
+ if total is None:
+ return 'need total argument. go to patron_sendmany/'
+ total = int(float(total)*1e8)
+ trunc = int(float(trunc)*1e8)
+ return json.dumps(dict(
+ (bitcoin_data.script2_to_address(script, node.net.PARENT), value/1e8)
+ for script, value in get_current_scaled_txouts(total, trunc).iteritems()
+ if bitcoin_data.script2_to_address(script, node.net.PARENT) is not None
+ ))
def get_global_stats():
# averaged over last hour
- lookbehind = 3600//net.SHARE_PERIOD
- if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
+ if node.tracker.get_height(node.best_share_var.value) < 10:
return None
+ lookbehind = min(node.tracker.get_height(node.best_share_var.value), 3600//node.net.SHARE_PERIOD)
- nonstale_hash_rate = p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], lookbehind)
- stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
- return json.dumps(dict(
+ nonstale_hash_rate = p2pool_data.get_pool_attempts_per_second(node.tracker, node.best_share_var.value, lookbehind)
+ stale_prop = p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, lookbehind)
+ return dict(
pool_nonstale_hash_rate=nonstale_hash_rate,
pool_hash_rate=nonstale_hash_rate/(1 - stale_prop),
pool_stale_prop=stale_prop,
- ))
+ min_difficulty=bitcoin_data.target_to_difficulty(node.tracker.items[node.best_share_var.value].max_target),
+ )
def get_local_stats():
- lookbehind = 3600//net.SHARE_PERIOD
- if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
+ if node.tracker.get_height(node.best_share_var.value) < 10:
return None
+ lookbehind = min(node.tracker.get_height(node.best_share_var.value), 3600//node.net.SHARE_PERIOD)
- global_stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
+ global_stale_prop = p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, lookbehind)
- my_unstale_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes)
- my_orphan_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes and share.share_data['stale_info'] == 253)
- my_doa_count = sum(1 for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind) if share.hash in my_share_hashes and share.share_data['stale_info'] == 254)
+ my_unstale_count = sum(1 for share in node.tracker.get_chain(node.best_share_var.value, lookbehind) if share.hash in wb.my_share_hashes)
+ my_orphan_count = sum(1 for share in node.tracker.get_chain(node.best_share_var.value, lookbehind) if share.hash in wb.my_share_hashes and share.share_data['stale_info'] == 'orphan')
+ my_doa_count = sum(1 for share in node.tracker.get_chain(node.best_share_var.value, lookbehind) if share.hash in wb.my_share_hashes and share.share_data['stale_info'] == 'doa')
my_share_count = my_unstale_count + my_orphan_count + my_doa_count
my_stale_count = my_orphan_count + my_doa_count
my_stale_prop = my_stale_count/my_share_count if my_share_count != 0 else None
my_work = sum(bitcoin_data.target_to_average_attempts(share.target)
- for share in tracker.get_chain(current_work.value['best_share_hash'], lookbehind - 1)
- if share.hash in my_share_hashes)
- actual_time = (tracker.shares[current_work.value['best_share_hash']].timestamp -
- tracker.shares[tracker.get_nth_parent_hash(current_work.value['best_share_hash'], lookbehind - 1)].timestamp)
+ for share in node.tracker.get_chain(node.best_share_var.value, lookbehind - 1)
+ if share.hash in wb.my_share_hashes)
+ actual_time = (node.tracker.items[node.best_share_var.value].timestamp -
+ node.tracker.items[node.tracker.get_nth_parent_hash(node.best_share_var.value, lookbehind - 1)].timestamp)
share_att_s = my_work / actual_time
- miner_hash_rates = {}
- miner_dead_hash_rates = {}
- datums, dt = local_rate_monitor.get_datums_in_last()
- for datum in datums:
- miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt
- if datum['dead']:
- miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
+ miner_hash_rates, miner_dead_hash_rates = wb.get_local_rates()
+ (stale_orphan_shares, stale_doa_shares), shares, _ = wb.get_stale_counts()
- return json.dumps(dict(
+ return dict(
my_hash_rates_in_last_hour=dict(
note="DEPRECATED",
nonstale=share_att_s,
@@ -137,40 +152,77 @@ def get_web_root(tracker, current_work, current_work2, get_current_txouts, datad
),
miner_hash_rates=miner_hash_rates,
miner_dead_hash_rates=miner_dead_hash_rates,
- ))
-
- def get_peer_addresses():
- return ' '.join(peer.transport.getPeer().host + (':' + str(peer.transport.getPeer().port) if peer.transport.getPeer().port != net.P2P_PORT else '') for peer in p2p_node.peers.itervalues())
-
- def get_uptime():
- return json.dumps(time.time() - start_time)
+ efficiency_if_miner_perfect=(1 - stale_orphan_shares/shares)/(1 - global_stale_prop) if shares else None, # ignores dead shares because those are miner's fault and indicated by pseudoshare rejection
+ efficiency=(1 - (stale_orphan_shares+stale_doa_shares)/shares)/(1 - global_stale_prop) if shares else None,
+ peers=dict(
+ incoming=sum(1 for peer in node.p2p_node.peers.itervalues() if peer.incoming),
+ outgoing=sum(1 for peer in node.p2p_node.peers.itervalues() if not peer.incoming),
+ ),
+ shares=dict(
+ total=shares,
+ orphan=stale_orphan_shares,
+ dead=stale_doa_shares,
+ ),
+ uptime=time.time() - start_time,
+ attempts_to_share=bitcoin_data.target_to_average_attempts(node.tracker.items[node.best_share_var.value].max_target),
+ attempts_to_block=bitcoin_data.target_to_average_attempts(node.bitcoind_work.value['bits'].target),
+ block_value=node.bitcoind_work.value['subsidy']*1e-8,
+ warnings=p2pool_data.get_warnings(node.tracker, node.best_share_var.value, node.net, bitcoind_getinfo_var.value, node.bitcoind_work.value),
+ donation_proportion=wb.donation_percentage/100,
+ version=p2pool.__version__,
+ protocol_version=p2p.Protocol.VERSION,
+ fee=wb.worker_fee,
+ )
- class WebInterface(resource.Resource):
- def __init__(self, func, mime_type, *fields):
- self.func, self.mime_type, self.fields = func, mime_type, fields
+ class WebInterface(deferred_resource.DeferredResource):
+ def __init__(self, func, mime_type='application/json', args=()):
+ deferred_resource.DeferredResource.__init__(self)
+ self.func, self.mime_type, self.args = func, mime_type, args
+ def getChild(self, child, request):
+ return WebInterface(self.func, self.mime_type, self.args + (child,))
+
+ @defer.inlineCallbacks
def render_GET(self, request):
request.setHeader('Content-Type', self.mime_type)
request.setHeader('Access-Control-Allow-Origin', '*')
- return self.func(*(request.args[field][0] for field in self.fields))
+ res = yield self.func(*self.args)
+ defer.returnValue(json.dumps(res) if self.mime_type == 'application/json' else res)
- web_root.putChild('rate', WebInterface(get_rate, 'application/json'))
- web_root.putChild('users', WebInterface(get_users, 'application/json'))
- web_root.putChild('fee', WebInterface(lambda: json.dumps(worker_fee), 'application/json'))
- web_root.putChild('current_payouts', WebInterface(get_current_payouts, 'application/json'))
- web_root.putChild('patron_sendmany', WebInterface(get_patron_sendmany, 'text/plain', 'total'))
- web_root.putChild('global_stats', WebInterface(get_global_stats, 'application/json'))
- web_root.putChild('local_stats', WebInterface(get_local_stats, 'application/json'))
- web_root.putChild('peer_addresses', WebInterface(get_peer_addresses, 'text/plain'))
- web_root.putChild('payout_addr', WebInterface(lambda: json.dumps(bitcoin_data.pubkey_hash_to_address(my_pubkey_hash, net.PARENT)), 'application/json'))
- web_root.putChild('recent_blocks', WebInterface(lambda: json.dumps(recent_blocks), 'application/json'))
- web_root.putChild('uptime', WebInterface(get_uptime, 'application/json'))
-
- try:
- from . import draw
- web_root.putChild('chain_img', WebInterface(lambda: draw.get(tracker, current_work.value['best_share_hash']), 'image/png'))
- except ImportError:
- print "Install Pygame and PIL to enable visualizations! Visualizations disabled."
+ def decent_height():
+ return min(node.tracker.get_height(node.best_share_var.value), 720)
+ web_root.putChild('rate', WebInterface(lambda: p2pool_data.get_pool_attempts_per_second(node.tracker, node.best_share_var.value, decent_height())/(1-p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, decent_height()))))
+ web_root.putChild('difficulty', WebInterface(lambda: bitcoin_data.target_to_difficulty(node.tracker.items[node.best_share_var.value].max_target)))
+ web_root.putChild('users', WebInterface(get_users))
+ web_root.putChild('user_stales', WebInterface(lambda: dict((bitcoin_data.pubkey_hash_to_address(ph, node.net.PARENT), prop) for ph, prop in
+ p2pool_data.get_user_stale_props(node.tracker, node.best_share_var.value, node.tracker.get_height(node.best_share_var.value)).iteritems())))
+ web_root.putChild('fee', WebInterface(lambda: wb.worker_fee))
+ web_root.putChild('current_payouts', WebInterface(lambda: dict((bitcoin_data.script2_to_address(script, node.net.PARENT), value/1e8) for script, value in node.get_current_txouts().iteritems())))
+ web_root.putChild('patron_sendmany', WebInterface(get_patron_sendmany, 'text/plain'))
+ web_root.putChild('global_stats', WebInterface(get_global_stats))
+ web_root.putChild('local_stats', WebInterface(get_local_stats))
+ web_root.putChild('peer_addresses', WebInterface(lambda: ' '.join('%s%s' % (peer.transport.getPeer().host, ':'+str(peer.transport.getPeer().port) if peer.transport.getPeer().port != node.net.P2P_PORT else '') for peer in node.p2p_node.peers.itervalues())))
+ web_root.putChild('peer_txpool_sizes', WebInterface(lambda: dict(('%s:%i' % (peer.transport.getPeer().host, peer.transport.getPeer().port), peer.remembered_txs_size) for peer in node.p2p_node.peers.itervalues())))
+ web_root.putChild('pings', WebInterface(defer.inlineCallbacks(lambda: defer.returnValue(
+ dict([(a, (yield b)) for a, b in
+ [(
+ '%s:%i' % (peer.transport.getPeer().host, peer.transport.getPeer().port),
+ defer.inlineCallbacks(lambda peer=peer: defer.returnValue(
+ min([(yield peer.do_ping().addCallback(lambda x: x/0.001).addErrback(lambda fail: None)) for i in xrange(3)])
+ ))()
+ ) for peer in list(node.p2p_node.peers.itervalues())]
+ ])
+ ))))
+ web_root.putChild('peer_versions', WebInterface(lambda: dict(('%s:%i' % peer.addr, peer.other_sub_version) for peer in node.p2p_node.peers.itervalues())))
+ web_root.putChild('payout_addr', WebInterface(lambda: bitcoin_data.pubkey_hash_to_address(wb.my_pubkey_hash, node.net.PARENT)))
+ web_root.putChild('recent_blocks', WebInterface(lambda: [dict(
+ ts=s.timestamp,
+ hash='%064x' % s.header_hash,
+ number=pack.IntType(24).unpack(s.share_data['coinbase'][1:4]) if len(s.share_data['coinbase']) >= 4 else None,
+ share='%064x' % s.hash,
+ ) for s in node.tracker.get_chain(node.best_share_var.value, min(node.tracker.get_height(node.best_share_var.value), 24*60*60//node.net.SHARE_PERIOD)) if s.pow_hash <= s.header['bits'].target]))
+ web_root.putChild('uptime', WebInterface(lambda: time.time() - start_time))
+ web_root.putChild('stale_rates', WebInterface(lambda: p2pool_data.get_stale_counts(node.tracker, node.best_share_var.value, decent_height(), rates=True)))
new_root = resource.Resource()
web_root.putChild('web', new_root)
@@ -186,147 +238,211 @@ def get_web_root(tracker, current_work, current_work2, get_current_txouts, datad
while stat_log and stat_log[0]['time'] < time.time() - 24*60*60:
stat_log.pop(0)
- lookbehind = 3600//net.SHARE_PERIOD
- if tracker.get_height(current_work.value['best_share_hash']) < lookbehind:
+ lookbehind = 3600//node.net.SHARE_PERIOD
+ if node.tracker.get_height(node.best_share_var.value) < lookbehind:
return None
- global_stale_prop = p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], lookbehind)
- (stale_orphan_shares, stale_doa_shares), shares, _ = get_stale_counts()
-
- miner_hash_rates = {}
- miner_dead_hash_rates = {}
- datums, dt = local_rate_monitor.get_datums_in_last()
- for datum in datums:
- miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt
- if datum['dead']:
- miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
+ global_stale_prop = p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, lookbehind)
+ (stale_orphan_shares, stale_doa_shares), shares, _ = wb.get_stale_counts()
+ miner_hash_rates, miner_dead_hash_rates = wb.get_local_rates()
stat_log.append(dict(
time=time.time(),
- pool_hash_rate=p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], lookbehind)/(1-global_stale_prop),
+ pool_hash_rate=p2pool_data.get_pool_attempts_per_second(node.tracker, node.best_share_var.value, lookbehind)/(1-global_stale_prop),
pool_stale_prop=global_stale_prop,
local_hash_rates=miner_hash_rates,
local_dead_hash_rates=miner_dead_hash_rates,
shares=shares,
stale_shares=stale_orphan_shares + stale_doa_shares,
stale_shares_breakdown=dict(orphan=stale_orphan_shares, doa=stale_doa_shares),
- current_payout=get_current_txouts().get(bitcoin_data.pubkey_hash_to_script2(my_pubkey_hash), 0)*1e-8,
+ current_payout=node.get_current_txouts().get(bitcoin_data.pubkey_hash_to_script2(wb.my_pubkey_hash), 0)*1e-8,
peers=dict(
- incoming=sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming),
- outgoing=sum(1 for peer in p2p_node.peers.itervalues() if not peer.incoming),
+ incoming=sum(1 for peer in node.p2p_node.peers.itervalues() if peer.incoming),
+ outgoing=sum(1 for peer in node.p2p_node.peers.itervalues() if not peer.incoming),
),
- attempts_to_share=bitcoin_data.target_to_average_attempts(tracker.shares[current_work.value['best_share_hash']].max_target),
- attempts_to_block=bitcoin_data.target_to_average_attempts(current_work.value['bits'].target),
- block_value=current_work2.value['subsidy']*1e-8,
+ attempts_to_share=bitcoin_data.target_to_average_attempts(node.tracker.items[node.best_share_var.value].max_target),
+ attempts_to_block=bitcoin_data.target_to_average_attempts(node.bitcoind_work.value['bits'].target),
+ block_value=node.bitcoind_work.value['subsidy']*1e-8,
))
with open(os.path.join(datadir_path, 'stats'), 'wb') as f:
f.write(json.dumps(stat_log))
- task.LoopingCall(update_stat_log).start(5*60)
- new_root.putChild('log', WebInterface(lambda: json.dumps(stat_log), 'application/json'))
+ x = deferral.RobustLoopingCall(update_stat_log)
+ x.start(5*60)
+ stop_event.watch(x.stop)
+ new_root.putChild('log', WebInterface(lambda: stat_log))
- class ShareExplorer(resource.Resource):
- def __init__(self, share_hash):
- self.share_hash = share_hash
- def render_GET(self, request):
- request.setHeader('Content-Type', 'text/html')
- if self.share_hash not in tracker.shares:
- return 'share not known'
- share = tracker.shares[self.share_hash]
-
- format_bits = lambda bits: '%f (bits=%#8x) Work required: %sH
' % (bitcoin_data.target_to_difficulty(bits.target), bits.bits, math.format(bitcoin_data.target_to_average_attempts(bits.target)))
-
- request.write('Share %s
' % (share.hash, p2pool_data.format_hash(share.hash)))
- request.write('Previous: %s
' % (share.previous_hash, p2pool_data.format_hash(share.previous_hash)))
- request.write('Next: %s
' % (', '.join('%s' % (next, p2pool_data.format_hash(next)) for next in tracker.reverse_shares.get(share.hash, set())),))
- request.write('Verified: %s
' % (share.hash in tracker.verified.shares,))
- request.write('Time first seen: %s
' % (time.ctime(start_time if share.time_seen == 0 else share.time_seen),))
- request.write('Peer first received from: %s
' % ('%s:%i' % share.peer.addr if share.peer is not None else 'self or cache',))
-
- request.write('Share data
')
- request.write('Timestamp: %s
' % (time.ctime(share.timestamp),))
- request.write('Difficulty: %s
' % (format_bits(share.share_info['bits']),))
- request.write('Minimum difficulty: %s
' % (format_bits(share.share_info.get('max_bits', share.share_info['bits'])),))
- request.write('Payout script: %s
' % (bitcoin_data.script2_to_human(share.new_script, share.net.PARENT),))
- request.write('Donation: %.2f%%
' % (share.share_data['donation']/65535*100,))
- request.write('Stale info: %s
' % ({0: 'none', 253: 'had an orphan', 254: 'had a dead'}.get(share.share_data['stale_info'], 'unknown %i' % (share.share_data['stale_info'],)),))
- request.write('Nonce: %s
' % (cgi.escape(repr(share.share_data['nonce'])),))
-
- request.write('Block header
')
- request.write('Hash: %064x
' % (net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header_hash, share.header_hash))
- request.write('Version: %i
' % (share.header['version'],))
- request.write('Previous block: %064x
' % (net.PARENT.BLOCK_EXPLORER_URL_PREFIX, share.header['previous_block'], share.header['previous_block']))
- request.write('Timestamp: %s (%i)
' % (time.ctime(share.header['timestamp']), share.header['timestamp']))
- request.write('Difficulty: %f (bits=%#8x) Work: %sH
' % (bitcoin_data.target_to_difficulty(share.header['bits'].target), share.header['bits'].bits, math.format(bitcoin_data.target_to_average_attempts(share.header['bits'].target))))
- request.write('Nonce: %i
' % (share.header['nonce'],))
- if share.other_txs is not None:
- tx_count = len(share.other_txs)
- elif len(share.merkle_branch) == 0:
- tx_count = 1
- else:
- tx_count = 'between %i and %i' % (2**len(share.merkle_branch)//2+1, 2**len(share.merkle_branch))
- request.write('Transactions: %s
' % (tx_count,))
- coinbase = share.share_data['coinbase'].ljust(2, '\x00')
- request.write('Coinbase: %s %s
' % (cgi.escape(repr(coinbase)), coinbase.encode('hex')))
- request.write('Generation value: %.8f %s
' % (share.share_data['subsidy']*1e-8, net.PARENT.SYMBOL))
- #request.write('Generation txn: %32x
' % (share.gentx_hash,))
-
- return ''
- class Explorer(resource.Resource):
- def render_GET(self, request):
- if not request.path.endswith('/'):
- request.redirect(request.path + '/')
- return ''
- request.setHeader('Content-Type', 'text/html')
- request.write('P2Pool share explorer
')
-
- request.write('Verified heads
')
- request.write('')
- for h in tracker.verified.heads:
- request.write('- %s%s
' % (h, p2pool_data.format_hash(h), ' BEST' if h == current_work.value['best_share_hash'] else ''))
- request.write('
')
-
- request.write('Verified tails
')
- request.write('')
- for tail in tracker.verified.tails:
- for h in tracker.reverse_shares.get(tail, set()):
- request.write('- %s%s
' % (h, p2pool_data.format_hash(h), ' BEST' if h == current_work.value['best_share_hash'] else ''))
- request.write('
')
-
- request.write('Heads
')
- request.write('')
- for h in tracker.heads:
- request.write('- %s%s
' % (h, p2pool_data.format_hash(h), ' BEST' if h == current_work.value['best_share_hash'] else ''))
- request.write('
')
-
- request.write('Tails
')
- request.write('')
- for tail in tracker.tails:
- for h in tracker.reverse_shares.get(tail, set()):
- request.write('- %s%s
' % (h, p2pool_data.format_hash(h), ' BEST' if h == current_work.value['best_share_hash'] else ''))
- request.write('
')
-
+ def get_share(share_hash_str):
+ if int(share_hash_str, 16) not in node.tracker.items:
+ return None
+ share = node.tracker.items[int(share_hash_str, 16)]
+
+ return dict(
+ parent='%064x' % share.previous_hash,
+ children=['%064x' % x for x in sorted(node.tracker.reverse.get(share.hash, set()), key=lambda sh: -len(node.tracker.reverse.get(sh, set())))], # sorted from most children to least children
+ type_name=type(share).__name__,
+ local=dict(
+ verified=share.hash in node.tracker.verified.items,
+ time_first_seen=start_time if share.time_seen == 0 else share.time_seen,
+ peer_first_received_from=share.peer_addr,
+ ),
+ share_data=dict(
+ timestamp=share.timestamp,
+ target=share.target,
+ max_target=share.max_target,
+ payout_address=bitcoin_data.script2_to_address(share.new_script, node.net.PARENT),
+ donation=share.share_data['donation']/65535,
+ stale_info=share.share_data['stale_info'],
+ nonce=share.share_data['nonce'],
+ desired_version=share.share_data['desired_version'],
+ absheight=share.absheight,
+ abswork=share.abswork,
+ ),
+ block=dict(
+ hash='%064x' % share.header_hash,
+ header=dict(
+ version=share.header['version'],
+ previous_block='%064x' % share.header['previous_block'],
+ merkle_root='%064x' % share.header['merkle_root'],
+ timestamp=share.header['timestamp'],
+ target=share.header['bits'].target,
+ nonce=share.header['nonce'],
+ ),
+ gentx=dict(
+ hash='%064x' % share.gentx_hash,
+ coinbase=share.share_data['coinbase'].ljust(2, '\x00').encode('hex'),
+ value=share.share_data['subsidy']*1e-8,
+ last_txout_nonce='%016x' % share.contents['last_txout_nonce'],
+ ),
+ other_transaction_hashes=['%064x' % x for x in share.get_other_tx_hashes(node.tracker)],
+ ),
+ )
+ new_root.putChild('share', WebInterface(lambda share_hash_str: get_share(share_hash_str)))
+ new_root.putChild('heads', WebInterface(lambda: ['%064x' % x for x in node.tracker.heads]))
+ new_root.putChild('verified_heads', WebInterface(lambda: ['%064x' % x for x in node.tracker.verified.heads]))
+ new_root.putChild('tails', WebInterface(lambda: ['%064x' % x for t in node.tracker.tails for x in node.tracker.reverse.get(t, set())]))
+ new_root.putChild('verified_tails', WebInterface(lambda: ['%064x' % x for t in node.tracker.verified.tails for x in node.tracker.verified.reverse.get(t, set())]))
+ new_root.putChild('best_share_hash', WebInterface(lambda: '%064x' % node.best_share_var.value))
+ new_root.putChild('my_share_hashes', WebInterface(lambda: ['%064x' % my_share_hash for my_share_hash in wb.my_share_hashes]))
+ def get_share_data(share_hash_str):
+ if int(share_hash_str, 16) not in node.tracker.items:
return ''
- def getChild(self, child, request):
- if not child:
- return self
- return ShareExplorer(int(child, 16))
- new_root.putChild('explorer', Explorer())
+ share = node.tracker.items[int(share_hash_str, 16)]
+ return p2pool_data.share_type.pack(share.as_share1a())
+ new_root.putChild('share_data', WebInterface(lambda share_hash_str: get_share_data(share_hash_str), 'application/octet-stream'))
+ new_root.putChild('currency_info', WebInterface(lambda: dict(
+ symbol=node.net.PARENT.SYMBOL,
+ block_explorer_url_prefix=node.net.PARENT.BLOCK_EXPLORER_URL_PREFIX,
+ address_explorer_url_prefix=node.net.PARENT.ADDRESS_EXPLORER_URL_PREFIX,
+ tx_explorer_url_prefix=node.net.PARENT.TX_EXPLORER_URL_PREFIX,
+ )))
+ new_root.putChild('version', WebInterface(lambda: p2pool.__version__))
- grapher = graphs.Grapher(os.path.join(datadir_path, 'rrd'))
- web_root.putChild('graphs', grapher.get_resource())
- def add_point():
- if tracker.get_height(current_work.value['best_share_hash']) < 720:
- return
- nonstalerate = p2pool_data.get_pool_attempts_per_second(tracker, current_work.value['best_share_hash'], 720)
- poolrate = nonstalerate / (1 - p2pool_data.get_average_stale_prop(tracker, current_work.value['best_share_hash'], 720))
- grapher.add_poolrate_point(poolrate, poolrate - nonstalerate)
- task.LoopingCall(add_point).start(100)
- @pseudoshare_received.watch
+ hd_path = os.path.join(datadir_path, 'graph_db')
+ hd_data = _atomic_read(hd_path)
+ hd_obj = {}
+ if hd_data is not None:
+ try:
+ hd_obj = json.loads(hd_data)
+ except Exception:
+ log.err(None, 'Error reading graph database:')
+ dataview_descriptions = {
+ 'last_hour': graph.DataViewDescription(150, 60*60),
+ 'last_day': graph.DataViewDescription(300, 60*60*24),
+ 'last_week': graph.DataViewDescription(300, 60*60*24*7),
+ 'last_month': graph.DataViewDescription(300, 60*60*24*30),
+ 'last_year': graph.DataViewDescription(300, 60*60*24*365.25),
+ }
+ hd = graph.HistoryDatabase.from_obj({
+ 'local_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False),
+ 'local_dead_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False),
+ 'local_share_hash_rates': graph.DataStreamDescription(dataview_descriptions, is_gauge=False,
+ multivalues=True, multivalue_undefined_means_0=True,
+ default_func=graph.make_multivalue_migrator(dict(good='local_share_hash_rate', dead='local_dead_share_hash_rate', orphan='local_orphan_share_hash_rate'),
+ post_func=lambda bins: [dict((k, (v[0] - (sum(bin.get(rem_k, (0, 0))[0] for rem_k in ['dead', 'orphan']) if k == 'good' else 0), v[1])) for k, v in bin.iteritems()) for bin in bins])),
+ 'pool_rates': graph.DataStreamDescription(dataview_descriptions, multivalues=True,
+ multivalue_undefined_means_0=True),
+ 'current_payout': graph.DataStreamDescription(dataview_descriptions),
+ 'current_payouts': graph.DataStreamDescription(dataview_descriptions, multivalues=True),
+ 'peers': graph.DataStreamDescription(dataview_descriptions, multivalues=True, default_func=graph.make_multivalue_migrator(dict(incoming='incoming_peers', outgoing='outgoing_peers'))),
+ 'miner_hash_rates': graph.DataStreamDescription(dataview_descriptions, is_gauge=False, multivalues=True),
+ 'miner_dead_hash_rates': graph.DataStreamDescription(dataview_descriptions, is_gauge=False, multivalues=True),
+ 'desired_version_rates': graph.DataStreamDescription(dataview_descriptions, multivalues=True,
+ multivalue_undefined_means_0=True),
+ 'traffic_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False, multivalues=True),
+ 'getwork_latency': graph.DataStreamDescription(dataview_descriptions),
+ 'memory_usage': graph.DataStreamDescription(dataview_descriptions),
+ }, hd_obj)
+ x = deferral.RobustLoopingCall(lambda: _atomic_write(hd_path, json.dumps(hd.to_obj())))
+ x.start(100)
+ stop_event.watch(x.stop)
+ @wb.pseudoshare_received.watch
def _(work, dead, user):
- reactor.callLater(1, grapher.add_localrate_point, work, dead)
+ t = time.time()
+ hd.datastreams['local_hash_rate'].add_datum(t, work)
+ if dead:
+ hd.datastreams['local_dead_hash_rate'].add_datum(t, work)
if user is not None:
- reactor.callLater(1, grapher.add_localminer_point, user, work, dead)
-
+ hd.datastreams['miner_hash_rates'].add_datum(t, {user: work})
+ if dead:
+ hd.datastreams['miner_dead_hash_rates'].add_datum(t, {user: work})
+ @wb.share_received.watch
+ def _(work, dead, share_hash):
+ t = time.time()
+ if not dead:
+ hd.datastreams['local_share_hash_rates'].add_datum(t, dict(good=work))
+ else:
+ hd.datastreams['local_share_hash_rates'].add_datum(t, dict(dead=work))
+ def later():
+ res = node.tracker.is_child_of(share_hash, node.best_share_var.value)
+ if res is None: res = False # share isn't connected to sharechain? assume orphaned
+ if res and dead: # share was DOA, but is now in sharechain
+ # move from dead to good
+ hd.datastreams['local_share_hash_rates'].add_datum(t, dict(dead=-work, good=work))
+ elif not res and not dead: # share wasn't DOA, and isn't in sharechain
+ # move from good to orphan
+ hd.datastreams['local_share_hash_rates'].add_datum(t, dict(good=-work, orphan=work))
+ reactor.callLater(200, later)
+ @node.p2p_node.traffic_happened.watch
+ def _(name, bytes):
+ hd.datastreams['traffic_rate'].add_datum(time.time(), {name: bytes})
+ def add_point():
+ if node.tracker.get_height(node.best_share_var.value) < 10:
+ return None
+ lookbehind = min(node.net.CHAIN_LENGTH, 60*60//node.net.SHARE_PERIOD, node.tracker.get_height(node.best_share_var.value))
+ t = time.time()
+
+ pool_rates = p2pool_data.get_stale_counts(node.tracker, node.best_share_var.value, lookbehind, rates=True)
+ pool_total = sum(pool_rates.itervalues())
+ hd.datastreams['pool_rates'].add_datum(t, pool_rates)
+
+ current_txouts = node.get_current_txouts()
+ hd.datastreams['current_payout'].add_datum(t, current_txouts.get(bitcoin_data.pubkey_hash_to_script2(wb.my_pubkey_hash), 0)*1e-8)
+ miner_hash_rates, miner_dead_hash_rates = wb.get_local_rates()
+ current_txouts_by_address = dict((bitcoin_data.script2_to_address(script, node.net.PARENT), amount) for script, amount in current_txouts.iteritems())
+ hd.datastreams['current_payouts'].add_datum(t, dict((user, current_txouts_by_address[user]*1e-8) for user in miner_hash_rates if user in current_txouts_by_address))
+
+ hd.datastreams['peers'].add_datum(t, dict(
+ incoming=sum(1 for peer in node.p2p_node.peers.itervalues() if peer.incoming),
+ outgoing=sum(1 for peer in node.p2p_node.peers.itervalues() if not peer.incoming),
+ ))
+
+ vs = p2pool_data.get_desired_version_counts(node.tracker, node.best_share_var.value, lookbehind)
+ vs_total = sum(vs.itervalues())
+ hd.datastreams['desired_version_rates'].add_datum(t, dict((str(k), v/vs_total*pool_total) for k, v in vs.iteritems()))
+ try:
+ hd.datastreams['memory_usage'].add_datum(t, memory.resident())
+ except:
+ if p2pool.DEBUG:
+ traceback.print_exc()
+ x = deferral.RobustLoopingCall(add_point)
+ x.start(5)
+ stop_event.watch(x.stop)
+ @node.bitcoind_work.changed.watch
+ def _(new_work):
+ hd.datastreams['getwork_latency'].add_datum(time.time(), new_work['latency'])
+ new_root.putChild('graph_data', WebInterface(lambda source, view: hd.datastreams[source].dataviews[view].get_data(time.time())))
+
+ web_root.putChild('static', static.File(os.path.join(os.path.dirname(sys.argv[0]), 'web-static')))
return web_root