X-Git-Url: https://git.novaco.in/?a=blobdiff_plain;f=p2pool%2Fweb.py;h=6f72550373eafcc10dad957cc30864b151e43c2f;hb=b142f305e6c8cfbe68e5bb5c8e2f907d5f679741;hp=fdd149643b57d97dccba101831556ec6d60a13c6;hpb=9328e4738faa7388b33f474c46d0d09c6c2723b2;p=p2pool.git diff --git a/p2pool/web.py b/p2pool/web.py index fdd1496..6f72550 100644 --- a/p2pool/web.py +++ b/p2pool/web.py @@ -5,15 +5,16 @@ import json import os import sys import time +import traceback -from twisted.internet import defer, task +from twisted.internet import defer, reactor from twisted.python import log from twisted.web import resource, static import p2pool from bitcoin import data as bitcoin_data -from . import data as p2pool_data -from util import deferred_resource, graph, math, pack +from . import data as p2pool_data, p2p +from util import deferral, deferred_resource, graph, math, memory, pack, variable def _atomic_read(filename): try: @@ -44,7 +45,7 @@ def _atomic_write(filename, data): os.remove(filename) os.rename(filename + '.new', filename) -def get_web_root(wb, datadir_path, bitcoind_warning_var): +def get_web_root(wb, datadir_path, bitcoind_getinfo_var, stop_event=variable.Event()): node = wb.node start_time = time.time() @@ -90,16 +91,6 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): if bitcoin_data.script2_to_address(script, node.net.PARENT) is not None )) - def get_local_rates(): - miner_hash_rates = {} - miner_dead_hash_rates = {} - datums, dt = wb.local_rate_monitor.get_datums_in_last() - for datum in datums: - miner_hash_rates[datum['user']] = miner_hash_rates.get(datum['user'], 0) + datum['work']/dt - if datum['dead']: - miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt - return miner_hash_rates, miner_dead_hash_rates - def get_global_stats(): # averaged over last hour if node.tracker.get_height(node.best_share_var.value) < 10: @@ -137,7 +128,7 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): node.tracker.items[node.tracker.get_nth_parent_hash(node.best_share_var.value, lookbehind - 1)].timestamp) share_att_s = my_work / actual_time - miner_hash_rates, miner_dead_hash_rates = get_local_rates() + miner_hash_rates, miner_dead_hash_rates = wb.get_local_rates() (stale_orphan_shares, stale_doa_shares), shares, _ = wb.get_stale_counts() return dict( @@ -176,8 +167,11 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): attempts_to_share=bitcoin_data.target_to_average_attempts(node.tracker.items[node.best_share_var.value].max_target), attempts_to_block=bitcoin_data.target_to_average_attempts(node.bitcoind_work.value['bits'].target), block_value=node.bitcoind_work.value['subsidy']*1e-8, - warnings=p2pool_data.get_warnings(node.tracker, node.best_share_var.value, node.net, bitcoind_warning_var.value, node.bitcoind_work.value), + warnings=p2pool_data.get_warnings(node.tracker, node.best_share_var.value, node.net, bitcoind_getinfo_var.value, node.bitcoind_work.value), donation_proportion=wb.donation_percentage/100, + version=p2pool.__version__, + protocol_version=p2p.Protocol.VERSION, + fee=wb.worker_fee, ) class WebInterface(deferred_resource.DeferredResource): @@ -195,7 +189,9 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): res = yield self.func(*self.args) defer.returnValue(json.dumps(res) if self.mime_type == 'application/json' else res) - web_root.putChild('rate', WebInterface(lambda: p2pool_data.get_pool_attempts_per_second(node.tracker, node.best_share_var.value, 720)/(1-p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, 720)))) + def decent_height(): + return min(node.tracker.get_height(node.best_share_var.value), 720) + web_root.putChild('rate', WebInterface(lambda: p2pool_data.get_pool_attempts_per_second(node.tracker, node.best_share_var.value, decent_height())/(1-p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, decent_height())))) web_root.putChild('difficulty', WebInterface(lambda: bitcoin_data.target_to_difficulty(node.tracker.items[node.best_share_var.value].max_target))) web_root.putChild('users', WebInterface(get_users)) web_root.putChild('user_stales', WebInterface(lambda: dict((bitcoin_data.pubkey_hash_to_address(ph, node.net.PARENT), prop) for ph, prop in @@ -205,7 +201,7 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): web_root.putChild('patron_sendmany', WebInterface(get_patron_sendmany, 'text/plain')) web_root.putChild('global_stats', WebInterface(get_global_stats)) web_root.putChild('local_stats', WebInterface(get_local_stats)) - web_root.putChild('peer_addresses', WebInterface(lambda: ['%s:%i' % (peer.transport.getPeer().host, peer.transport.getPeer().port) for peer in node.p2p_node.peers.itervalues()])) + web_root.putChild('peer_addresses', WebInterface(lambda: ' '.join('%s%s' % (peer.transport.getPeer().host, ':'+str(peer.transport.getPeer().port) if peer.transport.getPeer().port != node.net.P2P_PORT else '') for peer in node.p2p_node.peers.itervalues()))) web_root.putChild('peer_txpool_sizes', WebInterface(lambda: dict(('%s:%i' % (peer.transport.getPeer().host, peer.transport.getPeer().port), peer.remembered_txs_size) for peer in node.p2p_node.peers.itervalues()))) web_root.putChild('pings', WebInterface(defer.inlineCallbacks(lambda: defer.returnValue( dict([(a, (yield b)) for a, b in @@ -222,11 +218,11 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): web_root.putChild('recent_blocks', WebInterface(lambda: [dict( ts=s.timestamp, hash='%064x' % s.header_hash, - number=pack.IntType(24).unpack(s.share_data['coinbase'][1:4]), + number=pack.IntType(24).unpack(s.share_data['coinbase'][1:4]) if len(s.share_data['coinbase']) >= 4 else None, share='%064x' % s.hash, ) for s in node.tracker.get_chain(node.best_share_var.value, min(node.tracker.get_height(node.best_share_var.value), 24*60*60//node.net.SHARE_PERIOD)) if s.pow_hash <= s.header['bits'].target])) web_root.putChild('uptime', WebInterface(lambda: time.time() - start_time)) - web_root.putChild('stale_rates', WebInterface(lambda: p2pool_data.get_stale_counts(node.tracker, node.best_share_var.value, 720, rates=True))) + web_root.putChild('stale_rates', WebInterface(lambda: p2pool_data.get_stale_counts(node.tracker, node.best_share_var.value, decent_height(), rates=True))) new_root = resource.Resource() web_root.putChild('web', new_root) @@ -248,7 +244,7 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): global_stale_prop = p2pool_data.get_average_stale_prop(node.tracker, node.best_share_var.value, lookbehind) (stale_orphan_shares, stale_doa_shares), shares, _ = wb.get_stale_counts() - miner_hash_rates, miner_dead_hash_rates = get_local_rates() + miner_hash_rates, miner_dead_hash_rates = wb.get_local_rates() stat_log.append(dict( time=time.time(), @@ -271,7 +267,9 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): with open(os.path.join(datadir_path, 'stats'), 'wb') as f: f.write(json.dumps(stat_log)) - task.LoopingCall(update_stat_log).start(5*60) + x = deferral.RobustLoopingCall(update_stat_log) + x.start(5*60) + stop_event.watch(x.stop) new_root.putChild('log', WebInterface(lambda: stat_log)) def get_share(share_hash_str): @@ -286,7 +284,7 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): local=dict( verified=share.hash in node.tracker.verified.items, time_first_seen=start_time if share.time_seen == 0 else share.time_seen, - peer_first_received_from=share.peer.addr if share.peer is not None else None, + peer_first_received_from=share.peer_addr, ), share_data=dict( timestamp=share.timestamp, @@ -297,6 +295,8 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): stale_info=share.share_data['stale_info'], nonce=share.share_data['nonce'], desired_version=share.share_data['desired_version'], + absheight=share.absheight, + abswork=share.abswork, ), block=dict( hash='%064x' % share.header_hash, @@ -312,8 +312,9 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): hash='%064x' % share.gentx_hash, coinbase=share.share_data['coinbase'].ljust(2, '\x00').encode('hex'), value=share.share_data['subsidy']*1e-8, + last_txout_nonce='%016x' % share.contents['last_txout_nonce'], ), - txn_count_range=[len(share.other_txs), len(share.other_txs)] if share.other_txs is not None else 1 if len(share.merkle_link['branch']) == 0 else [2**len(share.merkle_link['branch'])//2+1, 2**len(share.merkle_link['branch'])], + other_transaction_hashes=['%064x' % x for x in share.get_other_tx_hashes(node.tracker)], ), ) new_root.putChild('share', WebInterface(lambda share_hash_str: get_share(share_hash_str))) @@ -332,6 +333,7 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): symbol=node.net.PARENT.SYMBOL, block_explorer_url_prefix=node.net.PARENT.BLOCK_EXPLORER_URL_PREFIX, address_explorer_url_prefix=node.net.PARENT.ADDRESS_EXPLORER_URL_PREFIX, + tx_explorer_url_prefix=node.net.PARENT.TX_EXPLORER_URL_PREFIX, ))) new_root.putChild('version', WebInterface(lambda: p2pool.__version__)) @@ -350,46 +352,42 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): 'last_month': graph.DataViewDescription(300, 60*60*24*30), 'last_year': graph.DataViewDescription(300, 60*60*24*365.25), } - def build_desired_rates(ds_name, ds_desc, dv_name, dv_desc, obj): + def build_peers(ds_name, ds_desc, dv_name, dv_desc, obj): if not obj: last_bin_end = 0 bins = dv_desc.bin_count*[{}] else: - pool_rates = obj['pool_rates'][dv_name] - desired_versions = obj['desired_versions'][dv_name] - def get_total_pool_rate(t): - n = int((pool_rates['last_bin_end'] - t)/dv_desc.bin_width) - if n < 0 or n >= dv_desc.bin_count: - return None - total = sum(x[0] for x in pool_rates['bins'][n].values()) - count = math.mean(x[1] for x in pool_rates['bins'][n].values()) - if count == 0: - return None - return total/count - last_bin_end = desired_versions['last_bin_end'] - bins = [dict((name, (total*get_total_pool_rate(last_bin_end - (i+1/2)*dv_desc.bin_width), count)) for name, (total, count) in desired_versions['bins'][i].iteritems()) for i in xrange(dv_desc.bin_count)] + incoming_peers = obj['incoming_peers'][dv_name] + outgoing_peers = obj['outgoing_peers'][dv_name] + assert incoming_peers['last_bin_end'] == outgoing_peers['last_bin_end'] + last_bin_end = incoming_peers['last_bin_end'] + assert len(incoming_peers['bins']) == len(outgoing_peers['bins']) == dv_desc.bin_count + bins = [dict(incoming=inc.get('null', (0, 0)), outgoing=out.get('null', (0, 0))) for inc, out in zip(incoming_peers['bins'], outgoing_peers['bins'])] return graph.DataView(dv_desc, ds_desc, last_bin_end, bins) hd = graph.HistoryDatabase.from_obj({ 'local_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False), 'local_dead_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False), 'local_share_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False), 'local_dead_share_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False), + 'local_orphan_share_hash_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False), 'pool_rates': graph.DataStreamDescription(dataview_descriptions, multivalues=True, multivalue_undefined_means_0=True), 'current_payout': graph.DataStreamDescription(dataview_descriptions), 'current_payouts': graph.DataStreamDescription(dataview_descriptions, multivalues=True), - 'incoming_peers': graph.DataStreamDescription(dataview_descriptions), - 'outgoing_peers': graph.DataStreamDescription(dataview_descriptions), + 'peers': graph.DataStreamDescription(dataview_descriptions, multivalues=True, default_func=build_peers), 'miner_hash_rates': graph.DataStreamDescription(dataview_descriptions, is_gauge=False, multivalues=True), 'miner_dead_hash_rates': graph.DataStreamDescription(dataview_descriptions, is_gauge=False, multivalues=True), 'desired_versions': graph.DataStreamDescription(dataview_descriptions, multivalues=True, multivalue_undefined_means_0=True), 'desired_version_rates': graph.DataStreamDescription(dataview_descriptions, multivalues=True, - multivalue_undefined_means_0=True, default_func=build_desired_rates), + multivalue_undefined_means_0=True), 'traffic_rate': graph.DataStreamDescription(dataview_descriptions, is_gauge=False, multivalues=True), 'getwork_latency': graph.DataStreamDescription(dataview_descriptions), + 'memory_usage': graph.DataStreamDescription(dataview_descriptions), }, hd_obj) - task.LoopingCall(lambda: _atomic_write(hd_path, json.dumps(hd.to_obj()))).start(100) + x = deferral.RobustLoopingCall(lambda: _atomic_write(hd_path, json.dumps(hd.to_obj()))) + x.start(100) + stop_event.watch(x.stop) @wb.pseudoshare_received.watch def _(work, dead, user): t = time.time() @@ -401,11 +399,21 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): if dead: hd.datastreams['miner_dead_hash_rates'].add_datum(t, {user: work}) @wb.share_received.watch - def _(work, dead): + def _(work, dead, share_hash): t = time.time() hd.datastreams['local_share_hash_rate'].add_datum(t, work) if dead: hd.datastreams['local_dead_share_hash_rate'].add_datum(t, work) + def later(): + res = node.tracker.is_child_of(share_hash, node.best_share_var.value) + if res is None: return # share isn't connected to sharechain? + if res and dead: # share was DOA, but is now in sharechain + # remove from DOA graph + hd.datastreams['local_dead_share_hash_rate'].add_datum(t, -work) + elif not res and not dead: # share wasn't DOA, and isn't in sharechain + # add to orphan graph + hd.datastreams['local_orphan_share_hash_rate'].add_datum(t, work) + reactor.callLater(200, later) @node.p2p_node.traffic_happened.watch def _(name, bytes): hd.datastreams['traffic_rate'].add_datum(time.time(), {name: bytes}) @@ -421,18 +429,27 @@ def get_web_root(wb, datadir_path, bitcoind_warning_var): current_txouts = node.get_current_txouts() hd.datastreams['current_payout'].add_datum(t, current_txouts.get(bitcoin_data.pubkey_hash_to_script2(wb.my_pubkey_hash), 0)*1e-8) - miner_hash_rates, miner_dead_hash_rates = get_local_rates() + miner_hash_rates, miner_dead_hash_rates = wb.get_local_rates() current_txouts_by_address = dict((bitcoin_data.script2_to_address(script, node.net.PARENT), amount) for script, amount in current_txouts.iteritems()) hd.datastreams['current_payouts'].add_datum(t, dict((user, current_txouts_by_address[user]*1e-8) for user in miner_hash_rates if user in current_txouts_by_address)) - hd.datastreams['incoming_peers'].add_datum(t, sum(1 for peer in node.p2p_node.peers.itervalues() if peer.incoming)) - hd.datastreams['outgoing_peers'].add_datum(t, sum(1 for peer in node.p2p_node.peers.itervalues() if not peer.incoming)) + hd.datastreams['peers'].add_datum(t, dict( + incoming=sum(1 for peer in node.p2p_node.peers.itervalues() if peer.incoming), + outgoing=sum(1 for peer in node.p2p_node.peers.itervalues() if not peer.incoming), + )) vs = p2pool_data.get_desired_version_counts(node.tracker, node.best_share_var.value, lookbehind) vs_total = sum(vs.itervalues()) hd.datastreams['desired_versions'].add_datum(t, dict((str(k), v/vs_total) for k, v in vs.iteritems())) hd.datastreams['desired_version_rates'].add_datum(t, dict((str(k), v/vs_total*pool_total) for k, v in vs.iteritems())) - task.LoopingCall(add_point).start(5) + try: + hd.datastreams['memory_usage'].add_datum(t, memory.resident()) + except: + if p2pool.DEBUG: + traceback.print_exc() + x = deferral.RobustLoopingCall(add_point) + x.start(5) + stop_event.watch(x.stop) @node.bitcoind_work.changed.watch def _(new_work): hd.datastreams['getwork_latency'].add_datum(time.time(), new_work['latency'])