1 from __future__ import division
8 from twisted.internet import defer
9 from twisted.python import log
11 import bitcoin.getwork as bitcoin_getwork, bitcoin.data as bitcoin_data
12 from bitcoin import script, worker_interface
13 from util import jsonrpc, variable, deferral, math, pack
14 import p2pool, p2pool.data as p2pool_data
16 class WorkerBridge(worker_interface.WorkerBridge):
17 def __init__(self, my_pubkey_hash, net, donation_percentage, bitcoind_work, best_block_header, merged_urls, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, worker_fee, p2p_node, submit_block, set_best_share, broadcast_share):
18 worker_interface.WorkerBridge.__init__(self)
19 self.recent_shares_ts_work = []
21 self.my_pubkey_hash = my_pubkey_hash
23 self.donation_percentage = donation_percentage
24 self.bitcoind_work = bitcoind_work
25 self.best_block_header = best_block_header
26 self.best_share_var = best_share_var
27 self.tracker = tracker
28 self.my_share_hashes = my_share_hashes
29 self.my_doa_share_hashes = my_doa_share_hashes
30 self.worker_fee = worker_fee
31 self.p2p_node = p2p_node
32 self.submit_block = submit_block
33 self.set_best_share = set_best_share
34 self.broadcast_share = broadcast_share
36 self.pseudoshare_received = variable.Event()
37 self.share_received = variable.Event()
38 self.local_rate_monitor = math.RateMonitor(10*60)
40 self.removed_unstales_var = variable.Variable((0, 0, 0))
41 self.removed_doa_unstales_var = variable.Variable(0)
43 @tracker.verified.removed.watch
45 if share.hash in self.my_share_hashes and tracker.is_child_of(share.hash, self.best_share_var.value):
46 assert share.share_data['stale_info'] in [None, 'orphan', 'doa'] # we made these shares in this instance
47 self.removed_unstales_var.set((
48 self.removed_unstales_var.value[0] + 1,
49 self.removed_unstales_var.value[1] + (1 if share.share_data['stale_info'] == 'orphan' else 0),
50 self.removed_unstales_var.value[2] + (1 if share.share_data['stale_info'] == 'doa' else 0),
52 if share.hash in self.my_doa_share_hashes and self.tracker.is_child_of(share.hash, self.best_share_var.value):
53 self.removed_doa_unstales_var.set(self.removed_doa_unstales_var.value + 1)
57 self.merged_work = variable.Variable({})
59 @defer.inlineCallbacks
60 def set_merged_work(merged_url, merged_userpass):
61 merged_proxy = jsonrpc.Proxy(merged_url, dict(Authorization='Basic ' + base64.b64encode(merged_userpass)))
63 auxblock = yield deferral.retry('Error while calling merged getauxblock:', 30)(merged_proxy.rpc_getauxblock)()
64 self.merged_work.set(dict(self.merged_work.value, **{auxblock['chainid']: dict(
65 hash=int(auxblock['hash'], 16),
66 target='p2pool' if auxblock['target'] == 'p2pool' else pack.IntType(256).unpack(auxblock['target'].decode('hex')),
67 merged_proxy=merged_proxy,
69 yield deferral.sleep(1)
70 for merged_url, merged_userpass in merged_urls:
71 set_merged_work(merged_url, merged_userpass)
73 @self.merged_work.changed.watch
74 def _(new_merged_work):
75 print 'Got new merged mining work!'
79 self.current_work = variable.Variable(None)
81 t = self.bitcoind_work.value
82 bb = self.best_block_header.value
83 if bb is not None and bb['previous_block'] == t['previous_block'] and net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(bb)) <= t['bits'].target:
84 print 'Skipping from block %x to block %x!' % (bb['previous_block'],
85 bitcoin_data.hash256(bitcoin_data.block_header_type.pack(bb)))
87 version=bb['version'],
88 previous_block=bitcoin_data.hash256(bitcoin_data.block_header_type.pack(bb)),
89 bits=bb['bits'], # not always true
91 height=t['height'] + 1,
92 time=bb['timestamp'] + 600, # better way?
94 merkle_link=bitcoin_data.calculate_merkle_link([None], 0),
95 subsidy=net.PARENT.SUBSIDY_FUNC(self.bitcoind_work.value['height']),
96 last_update=self.bitcoind_work.value['last_update'],
99 self.current_work.set(t)
100 self.bitcoind_work.changed.watch(lambda _: compute_work())
101 self.best_block_header.changed.watch(lambda _: compute_work())
104 self.new_work_event = variable.Event()
105 @self.current_work.transitioned.watch
106 def _(before, after):
107 # trigger LP if version/previous_block/bits changed or transactions changed from nothing
108 if any(before[x] != after[x] for x in ['version', 'previous_block', 'bits']) or (not before['transactions'] and after['transactions']):
109 self.new_work_event.happened()
110 self.merged_work.changed.watch(lambda _: self.new_work_event.happened())
111 self.best_share_var.changed.watch(lambda _: self.new_work_event.happened())
113 def get_stale_counts(self):
114 '''Returns (orphans, doas), total, (orphans_recorded_in_chain, doas_recorded_in_chain)'''
115 my_shares = len(self.my_share_hashes)
116 my_doa_shares = len(self.my_doa_share_hashes)
117 delta = self.tracker.verified.get_delta_to_last(self.best_share_var.value)
118 my_shares_in_chain = delta.my_count + self.removed_unstales_var.value[0]
119 my_doa_shares_in_chain = delta.my_doa_count + self.removed_doa_unstales_var.value
120 orphans_recorded_in_chain = delta.my_orphan_announce_count + self.removed_unstales_var.value[1]
121 doas_recorded_in_chain = delta.my_dead_announce_count + self.removed_unstales_var.value[2]
123 my_shares_not_in_chain = my_shares - my_shares_in_chain
124 my_doa_shares_not_in_chain = my_doa_shares - my_doa_shares_in_chain
126 return (my_shares_not_in_chain - my_doa_shares_not_in_chain, my_doa_shares_not_in_chain), my_shares, (orphans_recorded_in_chain, doas_recorded_in_chain)
128 def get_user_details(self, request):
129 user = request.getUser() if request.getUser() is not None else ''
131 desired_pseudoshare_target = None
133 user, desired_pseudoshare_difficulty_str = user.rsplit('+', 1)
135 desired_pseudoshare_target = bitcoin_data.difficulty_to_target(float(desired_pseudoshare_difficulty_str))
139 desired_share_target = 2**256 - 1
141 user, min_diff_str = user.rsplit('/', 1)
143 desired_share_target = bitcoin_data.difficulty_to_target(float(min_diff_str))
147 if random.uniform(0, 100) < self.worker_fee:
148 pubkey_hash = self.my_pubkey_hash
151 pubkey_hash = bitcoin_data.address_to_pubkey_hash(user, self.net.PARENT)
153 pubkey_hash = self.my_pubkey_hash
155 return user, pubkey_hash, desired_share_target, desired_pseudoshare_target
157 def preprocess_request(self, request):
158 user, pubkey_hash, desired_share_target, desired_pseudoshare_target = self.get_user_details(request)
159 return pubkey_hash, desired_share_target, desired_pseudoshare_target
161 def get_work(self, pubkey_hash, desired_share_target, desired_pseudoshare_target):
162 if len(self.p2p_node.peers) == 0 and self.net.PERSIST:
163 raise jsonrpc.Error_for_code(-12345)(u'p2pool is not connected to any peers')
164 if self.best_share_var.value is None and self.net.PERSIST:
165 raise jsonrpc.Error_for_code(-12345)(u'p2pool is downloading shares')
166 if time.time() > self.current_work.value['last_update'] + 60:
167 raise jsonrpc.Error_for_code(-12345)(u'lost contact with bitcoind')
169 if self.merged_work.value:
170 tree, size = bitcoin_data.make_auxpow_tree(self.merged_work.value)
171 mm_hashes = [self.merged_work.value.get(tree.get(i), dict(hash=0))['hash'] for i in xrange(size)]
172 mm_data = '\xfa\xbemm' + bitcoin_data.aux_pow_coinbase_type.pack(dict(
173 merkle_root=bitcoin_data.merkle_hash(mm_hashes),
177 mm_later = [(aux_work, mm_hashes.index(aux_work['hash']), mm_hashes) for chain_id, aux_work in self.merged_work.value.iteritems()]
182 tx_hashes = [bitcoin_data.hash256(bitcoin_data.tx_type.pack(tx)) for tx in self.current_work.value['transactions']]
183 tx_map = dict(zip(tx_hashes, self.current_work.value['transactions']))
185 share_type = p2pool_data.NewShare
186 if self.best_share_var.value is not None:
187 previous_share = self.tracker.items[self.best_share_var.value]
188 if isinstance(previous_share, p2pool_data.Share):
189 # Share -> NewShare only valid if 85% of hashes in [net.CHAIN_LENGTH*9//10, net.CHAIN_LENGTH] for new version
190 if self.tracker.get_height(previous_share.hash) < self.net.CHAIN_LENGTH:
191 share_type = p2pool_data.Share
192 elif time.time() < 1351383661 and self.net.NAME == 'bitcoin':
193 share_type = p2pool_data.Share
195 counts = p2pool_data.get_desired_version_counts(self.tracker,
196 self.tracker.get_nth_parent_hash(previous_share.hash, self.net.CHAIN_LENGTH*9//10), self.net.CHAIN_LENGTH//10)
197 if counts.get(p2pool_data.NewShare.VERSION, 0) < sum(counts.itervalues())*95//100:
198 share_type = p2pool_data.Share
201 share_info, gentx, other_transaction_hashes, get_share = share_type.generate_transaction(
202 tracker=self.tracker,
204 previous_share_hash=self.best_share_var.value,
205 coinbase=(script.create_push_script([
206 self.current_work.value['height'],
207 ] + ([mm_data] if mm_data else []) + [
208 ]) + self.current_work.value['coinbaseflags'])[:100],
209 nonce=random.randrange(2**32),
210 pubkey_hash=pubkey_hash,
211 subsidy=self.current_work.value['subsidy'],
212 donation=math.perfect_round(65535*self.donation_percentage/100),
213 stale_info=(lambda (orphans, doas), total, (orphans_recorded_in_chain, doas_recorded_in_chain):
214 'orphan' if orphans > orphans_recorded_in_chain else
215 'doa' if doas > doas_recorded_in_chain else
217 )(*self.get_stale_counts()),
218 desired_version=p2pool_data.NewShare.VERSION,
220 block_target=self.current_work.value['bits'].target,
221 desired_timestamp=int(time.time() + 0.5),
222 desired_target=desired_share_target,
223 ref_merkle_link=dict(branch=[], index=0),
224 desired_other_transaction_hashes=tx_hashes,
229 transactions = [gentx] + [tx_map[tx_hash] for tx_hash in other_transaction_hashes]
231 mm_later = [(dict(aux_work, target=aux_work['target'] if aux_work['target'] != 'p2pool' else share_info['bits'].target), index, hashes) for aux_work, index, hashes in mm_later]
233 if desired_pseudoshare_target is None:
235 if len(self.recent_shares_ts_work) == 50:
236 hash_rate = sum(work for ts, work in self.recent_shares_ts_work[1:])//(self.recent_shares_ts_work[-1][0] - self.recent_shares_ts_work[0][0])
238 target = min(target, int(2**256/hash_rate))
240 target = desired_pseudoshare_target
241 target = max(target, share_info['bits'].target)
242 for aux_work, index, hashes in mm_later:
243 target = max(target, aux_work['target'])
244 target = math.clip(target, self.net.PARENT.SANE_TARGET_RANGE)
246 getwork_time = time.time()
247 lp_count = self.new_work_event.times
248 merkle_link = bitcoin_data.calculate_merkle_link([bitcoin_data.hash256(bitcoin_data.tx_type.pack(tx)) for tx in transactions], 0)
250 print 'New work for worker! Difficulty: %.06f Share difficulty: %.06f Total block value: %.6f %s including %i transactions' % (
251 bitcoin_data.target_to_difficulty(target),
252 bitcoin_data.target_to_difficulty(share_info['bits'].target),
253 self.current_work.value['subsidy']*1e-8, self.net.PARENT.SYMBOL,
254 len(self.current_work.value['transactions']),
257 ba = bitcoin_getwork.BlockAttempt(
258 version=min(self.current_work.value['version'], 2),
259 previous_block=self.current_work.value['previous_block'],
260 merkle_root=bitcoin_data.check_merkle_link(bitcoin_data.hash256(bitcoin_data.tx_type.pack(transactions[0])), merkle_link),
261 timestamp=self.current_work.value['time'],
262 bits=self.current_work.value['bits'],
266 received_header_hashes = set()
268 def got_response(header, request):
269 header_hash = bitcoin_data.hash256(bitcoin_data.block_header_type.pack(header))
270 pow_hash = self.net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(header))
272 if pow_hash <= header['bits'].target or p2pool.DEBUG:
273 self.submit_block(dict(header=header, txs=transactions), ignore_failure=False)
274 if pow_hash <= header['bits'].target:
276 print 'GOT BLOCK FROM MINER! Passing to bitcoind! %s%064x' % (self.net.PARENT.BLOCK_EXPLORER_URL_PREFIX, header_hash)
279 log.err(None, 'Error while processing potential block:')
281 user, _, _, _ = self.get_user_details(request)
282 assert header['previous_block'] == ba.previous_block
283 assert header['merkle_root'] == ba.merkle_root
284 assert header['bits'] == ba.bits
286 on_time = self.new_work_event.times == lp_count
288 for aux_work, index, hashes in mm_later:
290 if pow_hash <= aux_work['target'] or p2pool.DEBUG:
291 df = deferral.retry('Error submitting merged block: (will retry)', 10, 10)(aux_work['merged_proxy'].rpc_getauxblock)(
292 pack.IntType(256, 'big').pack(aux_work['hash']).encode('hex'),
293 bitcoin_data.aux_pow_type.pack(dict(
296 block_hash=header_hash,
297 merkle_link=merkle_link,
299 merkle_link=bitcoin_data.calculate_merkle_link(hashes, index),
300 parent_block_header=header,
304 def _(result, aux_work=aux_work):
305 if result != (pow_hash <= aux_work['target']):
306 print >>sys.stderr, 'Merged block submittal result: %s Expected: %s' % (result, pow_hash <= aux_work['target'])
308 print 'Merged block submittal result: %s' % (result,)
311 log.err(err, 'Error submitting merged block:')
313 log.err(None, 'Error while processing merged mining POW:')
315 if pow_hash <= share_info['bits'].target and header_hash not in received_header_hashes:
316 share = get_share(header, transactions)
318 print 'GOT SHARE! %s %s prev %s age %.2fs%s' % (
320 p2pool_data.format_hash(share.hash),
321 p2pool_data.format_hash(share.previous_hash),
322 time.time() - getwork_time,
323 ' DEAD ON ARRIVAL' if not on_time else '',
325 self.my_share_hashes.add(share.hash)
327 self.my_doa_share_hashes.add(share.hash)
329 self.tracker.add(share)
331 self.tracker.verified.add(share)
332 self.set_best_share()
335 if pow_hash <= header['bits'].target or p2pool.DEBUG:
336 self.broadcast_share(share.hash)
338 log.err(None, 'Error forwarding block solution:')
340 self.share_received.happened(bitcoin_data.target_to_average_attempts(share.target), not on_time)
342 if pow_hash > target:
343 print 'Worker %s submitted share with hash > target:' % (request.getUser(),)
344 print ' Hash: %56x' % (pow_hash,)
345 print ' Target: %56x' % (target,)
346 elif header_hash in received_header_hashes:
347 print >>sys.stderr, 'Worker %s @ %s submitted share more than once!' % (request.getUser(), request.getClientIP())
349 received_header_hashes.add(header_hash)
351 self.pseudoshare_received.happened(bitcoin_data.target_to_average_attempts(target), not on_time, user)
352 self.recent_shares_ts_work.append((time.time(), bitcoin_data.target_to_average_attempts(target)))
353 while len(self.recent_shares_ts_work) > 50:
354 self.recent_shares_ts_work.pop(0)
355 self.local_rate_monitor.add_datum(dict(work=bitcoin_data.target_to_average_attempts(target), dead=not on_time, user=user))
359 return ba, got_response