always produce blocks with nVersion=1 instead of passing through value received from...
[p2pool.git] / p2pool / work.py
1 from __future__ import division
2
3 import base64
4 import random
5 import sys
6 import time
7
8 from twisted.internet import defer
9 from twisted.python import log
10
11 import bitcoin.getwork as bitcoin_getwork, bitcoin.data as bitcoin_data
12 from bitcoin import worker_interface
13 from util import jsonrpc, variable, deferral, math, pack
14 import p2pool, p2pool.data as p2pool_data
15
16 class WorkerBridge(worker_interface.WorkerBridge):
17     def __init__(self, my_pubkey_hash, net, donation_percentage, bitcoind_work, best_block_header, merged_urls, best_share_var, tracker, my_share_hashes, my_doa_share_hashes, worker_fee, p2p_node, submit_block, set_best_share, broadcast_share, block_height_var):
18         worker_interface.WorkerBridge.__init__(self)
19         self.recent_shares_ts_work = []
20         
21         self.my_pubkey_hash = my_pubkey_hash
22         self.net = net
23         self.donation_percentage = donation_percentage
24         self.bitcoind_work = bitcoind_work
25         self.best_block_header = best_block_header
26         self.best_share_var = best_share_var
27         self.tracker = tracker
28         self.my_share_hashes = my_share_hashes
29         self.my_doa_share_hashes = my_doa_share_hashes
30         self.worker_fee = worker_fee
31         self.p2p_node = p2p_node
32         self.submit_block = submit_block
33         self.set_best_share = set_best_share
34         self.broadcast_share = broadcast_share
35         self.block_height_var = block_height_var
36         
37         self.pseudoshare_received = variable.Event()
38         self.share_received = variable.Event()
39         self.local_rate_monitor = math.RateMonitor(10*60)
40         
41         self.removed_unstales_var = variable.Variable((0, 0, 0))
42         self.removed_doa_unstales_var = variable.Variable(0)
43         
44         @tracker.verified.removed.watch
45         def _(share):
46             if share.hash in self.my_share_hashes and tracker.is_child_of(share.hash, self.best_share_var.value):
47                 assert share.share_data['stale_info'] in [None, 'orphan', 'doa'] # we made these shares in this instance
48                 self.removed_unstales_var.set((
49                     self.removed_unstales_var.value[0] + 1,
50                     self.removed_unstales_var.value[1] + (1 if share.share_data['stale_info'] == 'orphan' else 0),
51                     self.removed_unstales_var.value[2] + (1 if share.share_data['stale_info'] == 'doa' else 0),
52                 ))
53             if share.hash in self.my_doa_share_hashes and self.tracker.is_child_of(share.hash, self.best_share_var.value):
54                 self.removed_doa_unstales_var.set(self.removed_doa_unstales_var.value + 1)
55         
56         # MERGED WORK
57         
58         self.merged_work = variable.Variable({})
59         
60         @defer.inlineCallbacks
61         def set_merged_work(merged_url, merged_userpass):
62             merged_proxy = jsonrpc.Proxy(merged_url, dict(Authorization='Basic ' + base64.b64encode(merged_userpass)))
63             while True:
64                 auxblock = yield deferral.retry('Error while calling merged getauxblock:', 30)(merged_proxy.rpc_getauxblock)()
65                 self.merged_work.set(dict(self.merged_work.value, **{auxblock['chainid']: dict(
66                     hash=int(auxblock['hash'], 16),
67                     target='p2pool' if auxblock['target'] == 'p2pool' else pack.IntType(256).unpack(auxblock['target'].decode('hex')),
68                     merged_proxy=merged_proxy,
69                 )}))
70                 yield deferral.sleep(1)
71         for merged_url, merged_userpass in merged_urls:
72             set_merged_work(merged_url, merged_userpass)
73         
74         @self.merged_work.changed.watch
75         def _(new_merged_work):
76             print 'Got new merged mining work!'
77         
78         # COMBINE WORK
79         
80         self.current_work = variable.Variable(None)
81         def compute_work():
82             t = dict(self.bitcoind_work.value)
83             
84             bb = self.best_block_header.value
85             if bb is not None and bb['previous_block'] == t['previous_block'] and net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(bb)) <= t['bits'].target:
86                 print 'Skipping from block %x to block %x!' % (bb['previous_block'],
87                     bitcoin_data.hash256(bitcoin_data.block_header_type.pack(bb)))
88                 t = dict(
89                     previous_block=bitcoin_data.hash256(bitcoin_data.block_header_type.pack(bb)),
90                     bits=bb['bits'], # not always true
91                     coinbaseflags='',
92                     time=bb['timestamp'] + 600, # better way?
93                     transactions=[],
94                     merkle_link=bitcoin_data.calculate_merkle_link([None], 0),
95                     subsidy=net.PARENT.SUBSIDY_FUNC(self.block_height_var.value),
96                     clock_offset=self.bitcoind_work.value['clock_offset'],
97                     last_update=self.bitcoind_work.value['last_update'],
98                 )
99             
100             self.current_work.set(t)
101         self.bitcoind_work.changed.watch(lambda _: compute_work())
102         self.best_block_header.changed.watch(lambda _: compute_work())
103         compute_work()
104         
105         self.new_work_event = variable.Event()
106         @self.current_work.transitioned.watch
107         def _(before, after):
108             # trigger LP if previous_block/bits changed or transactions changed from nothing
109             if any(before[x] != after[x] for x in ['previous_block', 'bits']) or (not before['transactions'] and after['transactions']):
110                 self.new_work_event.happened()
111         self.merged_work.changed.watch(lambda _: self.new_work_event.happened())
112         self.best_share_var.changed.watch(lambda _: self.new_work_event.happened())
113     
114     def get_stale_counts(self):
115         '''Returns (orphans, doas), total, (orphans_recorded_in_chain, doas_recorded_in_chain)'''
116         my_shares = len(self.my_share_hashes)
117         my_doa_shares = len(self.my_doa_share_hashes)
118         delta = self.tracker.verified.get_delta_to_last(self.best_share_var.value)
119         my_shares_in_chain = delta.my_count + self.removed_unstales_var.value[0]
120         my_doa_shares_in_chain = delta.my_doa_count + self.removed_doa_unstales_var.value
121         orphans_recorded_in_chain = delta.my_orphan_announce_count + self.removed_unstales_var.value[1]
122         doas_recorded_in_chain = delta.my_dead_announce_count + self.removed_unstales_var.value[2]
123         
124         my_shares_not_in_chain = my_shares - my_shares_in_chain
125         my_doa_shares_not_in_chain = my_doa_shares - my_doa_shares_in_chain
126         
127         return (my_shares_not_in_chain - my_doa_shares_not_in_chain, my_doa_shares_not_in_chain), my_shares, (orphans_recorded_in_chain, doas_recorded_in_chain)
128     
129     def get_user_details(self, request):
130         user = request.getUser() if request.getUser() is not None else ''
131         
132         desired_pseudoshare_target = None
133         if '+' in user:
134             user, desired_pseudoshare_difficulty_str = user.rsplit('+', 1)
135             try:
136                 desired_pseudoshare_target = bitcoin_data.difficulty_to_target(float(desired_pseudoshare_difficulty_str))
137             except:
138                 pass
139         
140         desired_share_target = 2**256 - 1
141         if '/' in user:
142             user, min_diff_str = user.rsplit('/', 1)
143             try:
144                 desired_share_target = bitcoin_data.difficulty_to_target(float(min_diff_str))
145             except:
146                 pass
147         
148         if random.uniform(0, 100) < self.worker_fee:
149             pubkey_hash = self.my_pubkey_hash
150         else:
151             try:
152                 pubkey_hash = bitcoin_data.address_to_pubkey_hash(user, self.net.PARENT)
153             except: # XXX blah
154                 pubkey_hash = self.my_pubkey_hash
155         
156         return user, pubkey_hash, desired_share_target, desired_pseudoshare_target
157     
158     def preprocess_request(self, request):
159         user, pubkey_hash, desired_share_target, desired_pseudoshare_target = self.get_user_details(request)
160         return pubkey_hash, desired_share_target, desired_pseudoshare_target
161     
162     def get_work(self, pubkey_hash, desired_share_target, desired_pseudoshare_target):
163         if len(self.p2p_node.peers) == 0 and self.net.PERSIST:
164             raise jsonrpc.Error(-12345, u'p2pool is not connected to any peers')
165         if self.best_share_var.value is None and self.net.PERSIST:
166             raise jsonrpc.Error(-12345, u'p2pool is downloading shares')
167         if time.time() > self.current_work.value['last_update'] + 60:
168             raise jsonrpc.Error(-12345, u'lost contact with bitcoind')
169         
170         if self.merged_work.value:
171             tree, size = bitcoin_data.make_auxpow_tree(self.merged_work.value)
172             mm_hashes = [self.merged_work.value.get(tree.get(i), dict(hash=0))['hash'] for i in xrange(size)]
173             mm_data = '\xfa\xbemm' + bitcoin_data.aux_pow_coinbase_type.pack(dict(
174                 merkle_root=bitcoin_data.merkle_hash(mm_hashes),
175                 size=size,
176                 nonce=0,
177             ))
178             mm_later = [(aux_work, mm_hashes.index(aux_work['hash']), mm_hashes) for chain_id, aux_work in self.merged_work.value.iteritems()]
179         else:
180             mm_data = ''
181             mm_later = []
182         
183         if True:
184             share_info, generate_tx = p2pool_data.Share.generate_transaction(
185                 tracker=self.tracker,
186                 share_data=dict(
187                     previous_share_hash=self.best_share_var.value,
188                     coinbase=(mm_data + self.current_work.value['coinbaseflags'])[:100],
189                     nonce=random.randrange(2**32),
190                     pubkey_hash=pubkey_hash,
191                     subsidy=self.current_work.value['subsidy'],
192                     donation=math.perfect_round(65535*self.donation_percentage/100),
193                     stale_info=(lambda (orphans, doas), total, (orphans_recorded_in_chain, doas_recorded_in_chain):
194                         'orphan' if orphans > orphans_recorded_in_chain else
195                         'doa' if doas > doas_recorded_in_chain else
196                         None
197                     )(*self.get_stale_counts()),
198                     desired_version=3,
199                 ),
200                 block_target=self.current_work.value['bits'].target,
201                 desired_timestamp=int(time.time() - self.current_work.value['clock_offset']),
202                 desired_target=desired_share_target,
203                 ref_merkle_link=dict(branch=[], index=0),
204                 net=self.net,
205             )
206         
207         mm_later = [(dict(aux_work, target=aux_work['target'] if aux_work['target'] != 'p2pool' else share_info['bits'].target), index, hashes) for aux_work, index, hashes in mm_later]
208         
209         if desired_pseudoshare_target is None:
210             target = 2**256-1
211             if len(self.recent_shares_ts_work) == 50:
212                 hash_rate = sum(work for ts, work in self.recent_shares_ts_work[1:])//(self.recent_shares_ts_work[-1][0] - self.recent_shares_ts_work[0][0])
213                 if hash_rate:
214                     target = min(target, int(2**256/hash_rate))
215         else:
216             target = desired_pseudoshare_target
217         target = max(target, share_info['bits'].target)
218         for aux_work, index, hashes in mm_later:
219             target = max(target, aux_work['target'])
220         target = math.clip(target, self.net.PARENT.SANE_TARGET_RANGE)
221         
222         transactions = [generate_tx] + list(self.current_work.value['transactions'])
223         packed_generate_tx = bitcoin_data.tx_type.pack(generate_tx)
224         merkle_root = bitcoin_data.check_merkle_link(bitcoin_data.hash256(packed_generate_tx), self.current_work.value['merkle_link'])
225         
226         getwork_time = time.time()
227         lp_count = self.new_work_event.times
228         merkle_link = self.current_work.value['merkle_link']
229         
230         print 'New work for worker! Difficulty: %.06f Share difficulty: %.06f Total block value: %.6f %s including %i transactions' % (
231             bitcoin_data.target_to_difficulty(target),
232             bitcoin_data.target_to_difficulty(share_info['bits'].target),
233             self.current_work.value['subsidy']*1e-8, self.net.PARENT.SYMBOL,
234             len(self.current_work.value['transactions']),
235         )
236         
237         bits = self.current_work.value['bits']
238         previous_block = self.current_work.value['previous_block']
239         ba = bitcoin_getwork.BlockAttempt(
240             version=1,
241             previous_block=self.current_work.value['previous_block'],
242             merkle_root=merkle_root,
243             timestamp=self.current_work.value['time'],
244             bits=self.current_work.value['bits'],
245             share_target=target,
246         )
247         
248         received_header_hashes = set()
249         
250         def got_response(header, request):
251             header_hash = bitcoin_data.hash256(bitcoin_data.block_header_type.pack(header))
252             pow_hash = self.net.PARENT.POW_FUNC(bitcoin_data.block_header_type.pack(header))
253             try:
254                 if pow_hash <= header['bits'].target or p2pool.DEBUG:
255                     self.submit_block(dict(header=header, txs=transactions), ignore_failure=False)
256                     if pow_hash <= header['bits'].target:
257                         print
258                         print 'GOT BLOCK FROM MINER! Passing to bitcoind! %s%064x' % (self.net.PARENT.BLOCK_EXPLORER_URL_PREFIX, header_hash)
259                         print
260             except:
261                 log.err(None, 'Error while processing potential block:')
262             
263             user, _, _, _ = self.get_user_details(request)
264             assert header['merkle_root'] == merkle_root
265             assert header['previous_block'] == previous_block
266             assert header['bits'] == bits
267             
268             on_time = self.new_work_event.times == lp_count
269             
270             for aux_work, index, hashes in mm_later:
271                 try:
272                     if pow_hash <= aux_work['target'] or p2pool.DEBUG:
273                         df = deferral.retry('Error submitting merged block: (will retry)', 10, 10)(aux_work['merged_proxy'].rpc_getauxblock)(
274                             pack.IntType(256, 'big').pack(aux_work['hash']).encode('hex'),
275                             bitcoin_data.aux_pow_type.pack(dict(
276                                 merkle_tx=dict(
277                                     tx=transactions[0],
278                                     block_hash=header_hash,
279                                     merkle_link=merkle_link,
280                                 ),
281                                 merkle_link=bitcoin_data.calculate_merkle_link(hashes, index),
282                                 parent_block_header=header,
283                             )).encode('hex'),
284                         )
285                         @df.addCallback
286                         def _(result, aux_work=aux_work):
287                             if result != (pow_hash <= aux_work['target']):
288                                 print >>sys.stderr, 'Merged block submittal result: %s Expected: %s' % (result, pow_hash <= aux_work['target'])
289                             else:
290                                 print 'Merged block submittal result: %s' % (result,)
291                         @df.addErrback
292                         def _(err):
293                             log.err(err, 'Error submitting merged block:')
294                 except:
295                     log.err(None, 'Error while processing merged mining POW:')
296             
297             if pow_hash <= share_info['bits'].target and header_hash not in received_header_hashes:
298                 min_header = dict(header);del min_header['merkle_root']
299                 hash_link = p2pool_data.prefix_to_hash_link(packed_generate_tx[:-32-4], p2pool_data.Share.gentx_before_refhash)
300                 share = p2pool_data.Share(self.net, None, dict(
301                     min_header=min_header, share_info=share_info, hash_link=hash_link,
302                     ref_merkle_link=dict(branch=[], index=0),
303                 ), merkle_link=merkle_link, other_txs=transactions[1:] if pow_hash <= header['bits'].target else None)
304                 
305                 print 'GOT SHARE! %s %s prev %s age %.2fs%s' % (
306                     request.getUser(),
307                     p2pool_data.format_hash(share.hash),
308                     p2pool_data.format_hash(share.previous_hash),
309                     time.time() - getwork_time,
310                     ' DEAD ON ARRIVAL' if not on_time else '',
311                 )
312                 self.my_share_hashes.add(share.hash)
313                 if not on_time:
314                     self.my_doa_share_hashes.add(share.hash)
315                 
316                 self.tracker.add(share)
317                 if not p2pool.DEBUG:
318                     self.tracker.verified.add(share)
319                 self.set_best_share()
320                 
321                 try:
322                     if pow_hash <= header['bits'].target or p2pool.DEBUG:
323                         self.broadcast_share(share.hash)
324                 except:
325                     log.err(None, 'Error forwarding block solution:')
326                 
327                 self.share_received.happened(bitcoin_data.target_to_average_attempts(share.target), not on_time)
328             
329             if pow_hash > target:
330                 print 'Worker %s submitted share with hash > target:' % (request.getUser(),)
331                 print '    Hash:   %56x' % (pow_hash,)
332                 print '    Target: %56x' % (target,)
333             elif header_hash in received_header_hashes:
334                 print >>sys.stderr, 'Worker %s @ %s submitted share more than once!' % (request.getUser(), request.getClientIP())
335             else:
336                 received_header_hashes.add(header_hash)
337                 
338                 self.pseudoshare_received.happened(bitcoin_data.target_to_average_attempts(target), not on_time, user)
339                 self.recent_shares_ts_work.append((time.time(), bitcoin_data.target_to_average_attempts(target)))
340                 while len(self.recent_shares_ts_work) > 50:
341                     self.recent_shares_ts_work.pop(0)
342                 self.local_rate_monitor.add_datum(dict(work=bitcoin_data.target_to_average_attempts(target), dead=not on_time, user=user))
343             
344             return on_time
345         
346         return ba, got_response