from __future__ import division
import random
+import tempfile
from twisted.internet import defer, reactor
from twisted.trial import unittest
-from twisted.web import resource, server
+from twisted.web import client, resource, server
from p2pool import data, node, work
from p2pool.bitcoin import data as bitcoin_data, networks, worker_interface
def send_block(self, block):
pass
+ def send_tx(self, tx):
+ pass
+
def get_block_header(self, hash):
if hash == 0x16c169477c25421250ec5d32cf9c6d38538b5de970a2355fd89:
return defer.succeed({
return {
"version" : 2,
"previousblockhash" : "000000000000016c169477c25421250ec5d32cf9c6d38538b5de970a2355fd89",
- "transactions" : [
+ "transactions" : ([
{
"data" : "01000000014c6ade5af7e7803fdf2ecae88d939ec044fdb3de84bc70168723969666b30e38010000008b483045022100e1fce6361811e24d57b494c3d71a9e653e54b9489dd5a1889affdef8a1e912b002204079a4720f25b55a0f092bcd70a5824e38a85072bb8e58477df8eb6a66b967ae01410426952de5ee7e5fea3c2065ffceada913d7d643f5631f0d714d667a0b81b599aada24f6e0b46d4bd4051b8111be95cf460fbd1977eadb3f2adc68b4018f8b5ba6ffffffff020065cd1d000000001976a9144799fc9c1b2cfb2c0187551be50f6ea41ed37ed888ac80969800000000001976a914ac2092a73378e8b80a127748a10274c962579f5188ac00000000",
"hash" : "9665fece00aceffd175a28912707889c9b2039be004c77e1497b59d8f5132966",
"fee" : 0,
"sigops" : 2
}
- ],
+ ] if random.randrange(2) else []),
"coinbaseaux" : {
"flags" : "062f503253482f"
},
self.n.p2p_node.start()
wb = work.WorkerBridge(node=self.n, my_pubkey_hash=random.randrange(2**160), donation_percentage=random.uniform(0, 10), merged_urls=merged_urls, worker_fee=3)
+ self.wb = wb
web_root = resource.Resource()
worker_interface.WorkerInterface(wb).attach_to(web_root)
self.web_port = reactor.listenTCP(0, server.Site(web_root))
blah = yield proxy.rpc_getwork()
yield proxy.rpc_getwork(blah['data'])
+
yield deferral.sleep(3)
assert len(n.tracker.items) == 100
proxy = jsonrpc.Proxy('http://127.0.0.1:' + str(random.choice(nodes).web_port.getHost().port))
blah = yield proxy.rpc_getwork()
yield proxy.rpc_getwork(blah['data'])
- yield deferral.sleep(random.expovariate(1/.1))
+ yield deferral.sleep(.02)
print i
print type(nodes[0].n.tracker.items[nodes[0].n.best_share_var.value])
-
+
+ # crawl web pages
+ from p2pool import web
+ stop_event = variable.Event()
+ web2_root = web.get_web_root(nodes[0].wb, tempfile.mkdtemp(), variable.Variable(None), stop_event)
+ web2_port = reactor.listenTCP(0, server.Site(web2_root))
+ for name in web2_root.listNames() + ['web/' + x for x in web2_root.getChildWithDefault('web', None).listNames()]:
+ print
+ print name
+ try:
+ res = yield client.getPage('http://127.0.0.1:%i/%s' % (web2_port.getHost().port, name))
+ except:
+ import traceback
+ traceback.print_exc()
+ else:
+ print repr(res)[:100]
+ print
+ yield web2_port.stopListening()
+ stop_event.happened()
+ del web2_root
+
yield deferral.sleep(3)
for i, n in enumerate(nodes):
assert len(n.n.tracker.items) == SHARES, (i, len(n.n.tracker.items))
assert n.n.tracker.verified.get_height(n.n.best_share_var.value) == SHARES, (i, n.n.tracker.verified.get_height(n.n.best_share_var.value))
- assert type(n.n.tracker.items[nodes[0].n.best_share_var.value]) is data.NewNewShare
+ assert type(n.n.tracker.items[nodes[0].n.best_share_var.value]) is data.Share
assert type(n.n.tracker.items[n.n.tracker.get_nth_parent_hash(nodes[0].n.best_share_var.value, SHARES - 5)]) is data.Share
for n in nodes:
gc.collect()
yield deferral.sleep(20) # waiting for work_poller to exit
+ test_nodes.timeout = 300