os.remove(filename)
os.rename(filename + '.new', filename)
-def get_web_root(tracker, current_work, current_work2, get_current_txouts, datadir_path, net, get_stale_counts, my_pubkey_hash, local_rate_monitor, worker_fee, p2p_node, my_share_hashes, recent_blocks, pseudoshare_received):
+def get_web_root(tracker, current_work, current_work2, get_current_txouts, datadir_path, net, get_stale_counts, my_pubkey_hash, local_rate_monitor, worker_fee, p2p_node, my_share_hashes, recent_blocks, pseudoshare_received, share_received):
start_time = time.time()
web_root = resource.Resource()
if datum['dead']:
miner_dead_hash_rates[datum['user']] = miner_dead_hash_rates.get(datum['user'], 0) + datum['work']/dt
+ (stale_orphan_shares, stale_doa_shares), shares, _ = get_stale_counts()
+
return json.dumps(dict(
my_hash_rates_in_last_hour=dict(
note="DEPRECATED",
),
miner_hash_rates=miner_hash_rates,
miner_dead_hash_rates=miner_dead_hash_rates,
+ efficiency_if_miner_perfect=(1 - stale_orphan_shares/shares)/(1 - global_stale_prop) if shares else None, # ignores dead shares because those are miner's fault and indicated by pseudoshare rejection
))
def get_peer_addresses():
format_bits = lambda bits: '%f (bits=%#8x) Work required: %sH</p>' % (bitcoin_data.target_to_difficulty(bits.target), bits.bits, math.format(bitcoin_data.target_to_average_attempts(bits.target)))
- request.write('<h1>%s <a href="%x">%s</a></h1>' % (share.__class__, share.hash, p2pool_data.format_hash(share.hash)))
+ request.write('<h1>%s <a href="%x">%s</a></h1>' % (share.__class__.__name__, share.hash, p2pool_data.format_hash(share.hash)))
if share.previous_hash is not None:
request.write('<p>Previous: <a href="%x">%s</a>' % (share.previous_hash, p2pool_data.format_hash(share.previous_hash)))
if tracker.get_height(share.hash) >= 100:
request.write('<p>Nonce: %i</p>' % (share.header['nonce'],))
if share.other_txs is not None:
tx_count = len(share.other_txs)
- elif len(share.merkle_branch) == 0:
+ elif len(share.merkle_link['branch']) == 0:
tx_count = 1
else:
- tx_count = 'between %i and %i' % (2**len(share.merkle_branch)//2+1, 2**len(share.merkle_branch))
+ tx_count = 'between %i and %i' % (2**len(share.merkle_link['branch'])//2+1, 2**len(share.merkle_link['branch']))
request.write('<p>Transactions: %s</p>' % (tx_count,))
coinbase = share.share_data['coinbase'].ljust(2, '\x00')
request.write('<p>Coinbase: %s %s</p>' % (cgi.escape(repr(coinbase)), coinbase.encode('hex')))
grapher.add_poolrate_point(poolrate, poolrate - nonstalerate)
task.LoopingCall(add_point).start(100)
@pseudoshare_received.watch
- def _(work, dead, user):
+ def _(work, dead, user, had_vip_pass):
reactor.callLater(1, grapher.add_localrate_point, work, dead)
- if user is not None:
+ if user is not None and had_vip_pass:
reactor.callLater(1, grapher.add_localminer_point, user, work, dead)
hd_path = os.path.join(datadir_path, 'graph_db')
'last_month': graph.DataViewDescription(300, 60*60*24*30),
'last_year': graph.DataViewDescription(300, 60*60*24*365.25),
}
+ def combine_and_keep_largest(*dicts):
+ res = {}
+ for d in dicts:
+ for k, v in d.iteritems():
+ res[k] = res.get(k, 0) + v
+ return dict((k, v) for k, v in sorted(res.iteritems(), key=lambda (k, v): v)[-30:] if v)
hd = graph.HistoryDatabase.from_obj({
'local_hash_rate': graph.DataStreamDescription(False, dataview_descriptions),
'local_dead_hash_rate': graph.DataStreamDescription(False, dataview_descriptions),
+ 'local_share_hash_rate': graph.DataStreamDescription(False, dataview_descriptions),
+ 'local_dead_share_hash_rate': graph.DataStreamDescription(False, dataview_descriptions),
'pool_rate': graph.DataStreamDescription(True, dataview_descriptions),
'pool_stale_rate': graph.DataStreamDescription(True, dataview_descriptions),
'current_payout': graph.DataStreamDescription(True, dataview_descriptions),
+ 'incoming_peers': graph.DataStreamDescription(True, dataview_descriptions),
+ 'outgoing_peers': graph.DataStreamDescription(True, dataview_descriptions),
+ 'miner_hash_rates': graph.DataStreamDescription(False, dataview_descriptions, {}, combine_and_keep_largest, math.mult_dict),
+ 'miner_dead_hash_rates': graph.DataStreamDescription(False, dataview_descriptions, {}, combine_and_keep_largest, math.mult_dict),
}, hd_obj)
task.LoopingCall(lambda: _atomic_write(hd_path, json.dumps(hd.to_obj()))).start(100)
@pseudoshare_received.watch
- def _(work, dead, user):
+ def _(work, dead, user, had_vip_pass):
t = time.time()
hd.datastreams['local_hash_rate'].add_datum(t, work)
if dead:
hd.datastreams['local_dead_hash_rate'].add_datum(t, work)
+ if user is not None:
+ hd.datastreams['miner_hash_rates'].add_datum(t, {user: work})
+ if dead:
+ hd.datastreams['miner_dead_hash_rates'].add_datum(t, {user: work})
+ @share_received.watch
+ def _(work, dead):
+ t = time.time()
+ hd.datastreams['local_share_hash_rate'].add_datum(t, work)
+ if dead:
+ hd.datastreams['local_dead_share_hash_rate'].add_datum(t, work)
def add_point():
if tracker.get_height(current_work.value['best_share_hash']) < 720:
return
hd.datastreams['pool_rate'].add_datum(t, poolrate)
hd.datastreams['pool_stale_rate'].add_datum(t, poolrate - nonstalerate)
hd.datastreams['current_payout'].add_datum(t, get_current_txouts().get(bitcoin_data.pubkey_hash_to_script2(my_pubkey_hash), 0)*1e-8)
+ hd.datastreams['incoming_peers'].add_datum(t, sum(1 for peer in p2p_node.peers.itervalues() if peer.incoming))
+ hd.datastreams['outgoing_peers'].add_datum(t, sum(1 for peer in p2p_node.peers.itervalues() if not peer.incoming))
task.LoopingCall(add_point).start(5)
new_root.putChild('graph_data', WebInterface(lambda source, view: json.dumps(hd.datastreams[source].dataviews[view].get_data(time.time())), 'application/json'))