X-Git-Url: https://git.novaco.in/?a=blobdiff_plain;f=server.py;h=9caf3530ef277f13c6ef64618ab193952a381f1c;hb=45062a27b514a45237d640ec7de133c98d3e475b;hp=7177764612e951fefa97ce90b368aba5f6492227;hpb=56e26122fbf0e0adb5ed1518c795794f4b890eb1;p=electrum-server.git diff --git a/server.py b/server.py index 7177764..9caf353 100755 --- a/server.py +++ b/server.py @@ -25,7 +25,7 @@ Todo: """ -import time, socket, operator, thread, ast, sys,re +import time, json, socket, operator, thread, ast, sys,re import psycopg2, binascii from Abe.abe import hash_to_address, decode_check_address @@ -36,6 +36,12 @@ import ConfigParser from json import dumps, loads import urllib +# we need to import electrum +sys.path.append('../client/') +from wallet import Wallet +from interface import Interface + + config = ConfigParser.ConfigParser() # set some defaults, which will be overwritten by the config file config.add_section('server') @@ -44,7 +50,6 @@ config.set('server', 'host', 'localhost') config.set('server', 'port', 50000) config.set('server', 'password', '') config.set('server', 'irc', 'yes') -config.set('server', 'cache', 'no') config.set('server', 'ircname', 'Electrum server') config.add_section('database') config.set('database', 'type', 'psycopg2') @@ -57,22 +62,48 @@ try: except: print "Could not read electrum.conf. I will use the default values." +try: + f = open('/etc/electrum.banner','r') + config.set('server','banner', f.read()) + f.close() +except: + pass + password = config.get('server','password') bitcoind_url = 'http://%s:%s@%s:%s/' % ( config.get('bitcoind','user'), config.get('bitcoind','password'), config.get('bitcoind','host'), config.get('bitcoind','port')) stopping = False block_number = -1 +old_block_number = -1 sessions = {} +sessions_sub_numblocks = {} # sessions that have subscribed to the service + dblock = thread.allocate_lock() peer_list = {} wallets = {} # for ultra-light clients such as bccapi +from Queue import Queue +input_queue = Queue() +output_queue = Queue() +address_queue = Queue() + class MyStore(Datastore_class): - def import_tx(self, tx, is_coinbase): - tx_id = super(MyStore, self).import_tx(tx, is_coinbase) - if config.get('server', 'cache') == 'yes': self.update_tx_cache(tx_id) + def import_block(self, b, chain_ids=frozenset()): + block_id = super(MyStore, self).import_block(b, chain_ids) + #print "block", block_id + for pos in xrange(len(b['transactions'])): + tx = b['transactions'][pos] + if 'hash' not in tx: + tx['hash'] = util.double_sha256(tx['tx']) + tx_id = store.tx_find_id_and_value(tx) + if tx_id: + self.update_tx_cache(tx_id) + else: + print "error: import_block: no tx_id" + return block_id + def update_tx_cache(self, txid): inrows = self.get_tx_inputs(txid, False) @@ -82,6 +113,8 @@ class MyStore(Datastore_class): if self.tx_cache.has_key(address): print "cache: invalidating", address self.tx_cache.pop(address) + address_queue.put(address) + outrows = self.get_tx_outputs(txid, False) for row in outrows: _hash = store.binout(row[6]) @@ -89,6 +122,7 @@ class MyStore(Datastore_class): if self.tx_cache.has_key(address): print "cache: invalidating", address self.tx_cache.pop(address) + address_queue.put(address) def safe_sql(self,sql, params=(), lock=True): try: @@ -203,10 +237,9 @@ class MyStore(Datastore_class): def get_history(self, addr): - if config.get('server','cache') == 'yes': - cached_version = self.tx_cache.get( addr ) - if cached_version is not None: - return cached_version + cached_version = self.tx_cache.get( addr ) + if cached_version is not None: + return cached_version version, binaddr = decode_check_address(addr) if binaddr is None: @@ -229,7 +262,6 @@ class MyStore(Datastore_class): tx_hash = self.hashout_hex(tx_hash) txpoint = { "nTime": int(nTime), - #"chain_id": int(chain_id), "height": int(height), "is_in": int(is_in), "blk_hash": self.hashout_hex(blk_hash), @@ -268,7 +300,6 @@ class MyStore(Datastore_class): #print "mempool", tx_hash txpoint = { "nTime": 0, - #"chain_id": 1, "height": 0, "is_in": int(is_in), "blk_hash": 'mempool', @@ -312,22 +343,38 @@ class MyStore(Datastore_class): if not row[4]: txpoint['raw_scriptPubKey'] = row[1] # cache result - if config.get('server','cache') == 'yes' and not address_has_mempool: + if not address_has_mempool: self.tx_cache[addr] = txpoints return txpoints +class Direct_Interface(Interface): + def __init__(self): + pass + + def handler(self, method, params = ''): + cmds = {'session.new':new_session, + 'session.poll':poll_session, + 'session.update':update_session, + 'blockchain.transaction.broadcast':send_tx, + 'blockchain.address.get_history':store.get_history + } + func = cmds[method] + return func( params ) + + def send_tx(tx): postdata = dumps({"method": 'importtransaction', 'params': [tx], 'id':'jsonrpc'}) respdata = urllib.urlopen(bitcoind_url, postdata).read() - try: - v = loads(respdata)['result'] - except: - v = "error: transaction rejected by memorypool" - return v + r = loads(respdata) + if r['error'] != None: + out = "error: transaction rejected by memorypool\n"+tx + else: + out = r['result'] + return out @@ -366,7 +413,7 @@ def get_cache(pw,addr): return 'wrong password' -def cmd_poll(session_id): +def poll_session(session_id): session = sessions.get(session_id) if session is None: print time.asctime(), "session not found", session_id @@ -379,18 +426,7 @@ def cmd_poll(session_id): k = 0 for addr in addresses: if store.tx_cache.get( addr ) is not None: k += 1 - - # get addtess status, i.e. the last block for that address. - tx_points = store.get_history(addr) - if not tx_points: - status = None - else: - lastpoint = tx_points[-1] - status = lastpoint['blk_hash'] - # this is a temporary hack; move it up once old clients have disappeared - if status == 'mempool' and session['version'] != "old": - status = status + ':%d'% len(tx_points) - + status = get_address_status( addr ) last_status = addresses.get( addr ) if last_status != status: addresses[addr] = status @@ -405,39 +441,88 @@ def cmd_poll(session_id): return out -def new_session(addresses, version, ipaddr): - session_id = random_string(10) +def do_update_address(addr): + # an address was involved in a transaction; we check if it was subscribed to in a session + # the address can be subscribed in several sessions; the cache should ensure that we don't do redundant requests + for session_id in sessions.keys(): + session = sessions[session_id] + if session.get('type') != 'subscribe': continue + addresses = session['addresses'].keys() - print time.strftime("[%d/%m/%Y-%H:%M:%S]"), "new session", ipaddr, addresses[0] if addresses else addresses, len(addresses), version + if addr in addresses: + status = get_address_status( addr ) + message_id, last_status = session['addresses'][addr] + if last_status != status: + #print "sending new status for %s:"%addr, status + send_status(session_id,message_id,addr,status) + sessions[session_id]['addresses'][addr] = (message_id,status) - sessions[session_id] = { 'addresses':{}, 'version':version, 'ip':ipaddr } + +def get_address_status(addr): + # get address status, i.e. the last block for that address. + tx_points = store.get_history(addr) + if not tx_points: + status = None + else: + lastpoint = tx_points[-1] + status = lastpoint['blk_hash'] + # this is a temporary hack; move it up once old clients have disappeared + if status == 'mempool': # and session['version'] != "old": + status = status + ':%d'% len(tx_points) + return status + + +def send_numblocks(session_id): + message_id = sessions_sub_numblocks[session_id] + out = json.dumps( {'id':message_id, 'result':block_number} ) + output_queue.put((session_id, out)) + +def send_status(session_id, message_id, address, status): + out = json.dumps( { 'id':message_id, 'result':status } ) + output_queue.put((session_id, out)) + +def subscribe_to_numblocks(session_id, message_id): + sessions_sub_numblocks[session_id] = message_id + send_numblocks(session_id) + +def subscribe_to_address(session_id, message_id, address): + status = get_address_status(address) + sessions[session_id]['type'] = 'subscribe' + sessions[session_id]['addresses'][address] = (message_id, status) + sessions[session_id]['last_time'] = time.time() + send_status(session_id, message_id, address, status) + +def new_session(version, addresses): + session_id = random_string(10) + sessions[session_id] = { 'addresses':{}, 'version':version } for a in addresses: sessions[session_id]['addresses'][a] = '' out = repr( (session_id, config.get('server','banner').replace('\\n','\n') ) ) sessions[session_id]['last_time'] = time.time() return out -def update_session(session_id,addresses,ipaddr): - print time.strftime("[%d/%m/%Y-%H:%M:%S]"), "update session", ipaddr, addresses[0] if addresses else addresses, len(addresses) +def update_session(session_id,addresses): sessions[session_id]['addresses'] = {} for a in addresses: sessions[session_id]['addresses'][a] = '' - out = 'ok' sessions[session_id]['last_time'] = time.time() + return 'ok' - -def listen_thread(store): +def native_server_thread(): s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) s.bind((config.get('server','host'), config.getint('server','port'))) s.listen(1) while not stopping: conn, addr = s.accept() - thread.start_new_thread(client_thread, (addr, conn,)) - + try: + thread.start_new_thread(native_client_thread, (addr, conn,)) + except: + # can't start new thread if there is no memory.. + traceback.print_exc(file=sys.stdout) -def client_thread(ipaddr,conn): +def native_client_thread(ipaddr,conn): #print "client thread", ipaddr try: ipaddr = ipaddr[0] @@ -469,6 +554,10 @@ def client_thread(ipaddr,conn): conn.close() +def timestr(): + return time.strftime("[%d/%m/%Y-%H:%M:%S]") + +# used by the native handler def do_command(cmd, data, ipaddr): if cmd=='b': @@ -485,7 +574,8 @@ def do_command(cmd, data, ipaddr): except: print "error", data return None - out = new_session(addresses, version, ipaddr) + print timestr(), "new session", ipaddr, addresses[0] if addresses else addresses, len(addresses), version + out = new_session(version, addresses) elif cmd=='update_session': try: @@ -493,8 +583,8 @@ def do_command(cmd, data, ipaddr): except: print "error" return None - out = update_session(session_id,addresses,ipaddr) - + print timestr(), "update session", ipaddr, addresses[0] if addresses else addresses, len(addresses) + out = update_session(session_id,addresses) elif cmd == 'bccapi_login': import electrum @@ -503,7 +593,7 @@ def do_command(cmd, data, ipaddr): master_public_key = k.decode('hex') # todo: sanitize. no need to decode twice... print master_public_key wallet_id = random_string(10) - w = electrum.Wallet() + w = Wallet( Direct_Interface() ) w.master_public_key = master_public_key.decode('hex') w.synchronize() wallets[wallet_id] = w @@ -511,7 +601,7 @@ def do_command(cmd, data, ipaddr): print "wallets", wallets elif cmd == 'bccapi_getAccountInfo': - from electrum import int_to_hex + from wallet import int_to_hex v, wallet_id = ast.literal_eval(data) w = wallets.get(wallet_id) if w is not None: @@ -524,7 +614,7 @@ def do_command(cmd, data, ipaddr): out = "error" elif cmd == 'bccapi_getAccountStatement': - from electrum import int_to_hex + from wallet import int_to_hex v, wallet_id = ast.literal_eval(data) w = wallets.get(wallet_id) if w is not None: @@ -544,7 +634,7 @@ def do_command(cmd, data, ipaddr): out = '' elif cmd=='poll': - out = cmd_poll(data) + out = poll_session(data) elif cmd == 'h': # history @@ -556,7 +646,7 @@ def do_command(cmd, data, ipaddr): elif cmd =='tx': out = send_tx(data) - print "sent tx:", out + print timestr(), "sent tx:", ipaddr, out elif cmd == 'stop': out = cmd_stop(data) @@ -571,42 +661,174 @@ def do_command(cmd, data, ipaddr): +#################################################################### + +def tcp_server_thread(): + thread.start_new_thread(process_input_queue, ()) + thread.start_new_thread(process_output_queue, ()) + + s = socket.socket(socket.AF_INET, socket.SOCK_STREAM) + s.setsockopt(socket.SOL_SOCKET, socket.SO_REUSEADDR, 1) + s.bind((config.get('server','host'), 50001)) + s.listen(1) + while not stopping: + conn, addr = s.accept() + try: + thread.start_new_thread(tcp_client_thread, (addr, conn,)) + except: + # can't start new thread if there is no memory.. + traceback.print_exc(file=sys.stdout) + + +def close_session(session_id): + print "lost connection", session_id + sessions.pop(session_id) + if session_id in sessions_sub_numblocks: + sessions_sub_numblocks.pop(session_id) + + +# one thread per client. put requests in a queue. +def tcp_client_thread(ipaddr,conn): + """ use a persistent connection. put commands in a queue.""" + + print timestr(), "TCP session", ipaddr + global sessions + + session_id = random_string(10) + sessions[session_id] = { 'conn':conn, 'addresses':{}, 'version':'unknown' } + + ipaddr = ipaddr[0] + msg = '' + + while not stopping: + try: + d = conn.recv(1024) + except socket.error: + d = '' + if not d: + close_session(session_id) + break + + msg += d + while True: + s = msg.find('\n') + if s ==-1: + break + else: + c = msg[0:s].strip() + msg = msg[s+1:] + if c == 'quit': + conn.close() + close_session(session_id) + return + try: + c = json.loads(c) + except: + print "json error", repr(c) + continue + try: + message_id = c.get('id') + method = c.get('method') + params = c.get('params') + except: + print "syntax error", repr(c), ipaddr + continue + + # add to queue + input_queue.put((session_id, message_id, method, params)) + + + +# read commands from the input queue. perform requests, etc. this should be called from the main thread. +def process_input_queue(): + while not stopping: + session_id, message_id, method, data = input_queue.get() + if session_id not in sessions.keys(): + continue + out = None + if method == 'address.subscribe': + address = data[0] + subscribe_to_address(session_id,message_id,address) + elif method == 'numblocks.subscribe': + subscribe_to_numblocks(session_id,message_id) + elif method == 'client.version': + sessions[session_id]['version'] = data[0] + elif method == 'server.banner': + out = { 'result':config.get('server','banner').replace('\\n','\n') } + elif method == 'server.peers': + out = { 'result':peer_list.values() } + elif method == 'address.get_history': + address = data[0] + out = { 'result':store.get_history( address ) } + elif method == 'transaction.broadcast': + txo = send_tx(data) + print "sent tx:", txo + out = { 'result':txo } + else: + print "unknown command", method + if out: + out['id'] = message_id + out = json.dumps( out ) + output_queue.put((session_id, out)) + +# this is a separate thread +def process_output_queue(): + while not stopping: + session_id, out = output_queue.get() + session = sessions.get(session_id) + if session: + try: + conn = session.get('conn') + conn.send(out+'\n') + except: + close_session(session_id) + + + + +#################################################################### + def memorypool_update(store): + ds = BCDataStream.BCDataStream() + previous_transactions = store.mempool_keys store.mempool_keys = [] postdata = dumps({"method": 'getmemorypool', 'params': [], 'id':'jsonrpc'}) respdata = urllib.urlopen(bitcoind_url, postdata).read() - v = loads(respdata)['result'] + r = loads(respdata) + if r['error'] != None: + return - - v = v['transactions'] + v = r['result'].get('transactions') for hextx in v: ds.clear() ds.write(hextx.decode('hex')) tx = deserialize.parse_Transaction(ds) tx['hash'] = util.double_sha256(tx['tx']) - tx_hash = tx['hash'][::-1].encode('hex') + tx_hash = store.hashin(tx['hash']) + store.mempool_keys.append(tx_hash) if store.tx_find_id_and_value(tx): pass else: - store.import_tx(tx, False) + tx_id = store.import_tx(tx, False) + store.update_tx_cache(tx_id) store.commit() - def clean_session_thread(): while not stopping: time.sleep(30) t = time.time() for k,s in sessions.items(): + if s.get('type') == 'subscribe': continue t0 = s['last_time'] if t - t0 > 5*60: - print time.strftime("[%d/%m/%Y-%H:%M:%S]"), "end session", s['ip'] sessions.pop(k) + print "lost session", k def irc_thread(): @@ -652,12 +874,12 @@ def irc_thread(): -def jsonrpc_thread(store): +def http_server_thread(store): # see http://code.google.com/p/jsonrpclib/ from SocketServer import ThreadingMixIn from jsonrpclib.SimpleJSONRPCServer import SimpleJSONRPCServer class SimpleThreadedJSONRPCServer(ThreadingMixIn, SimpleJSONRPCServer): pass - server = SimpleThreadedJSONRPCServer(('localhost', 8080)) + server = SimpleThreadedJSONRPCServer(( config.get('server','host'), 8081)) server.register_function(lambda : peer_list.values(), 'peers') server.register_function(cmd_stop, 'stop') server.register_function(cmd_load, 'load') @@ -666,6 +888,9 @@ def jsonrpc_thread(store): server.register_function(get_cache, 'get_cache') server.register_function(send_tx, 'blockchain.transaction.broadcast') server.register_function(store.get_history, 'blockchain.address.get_history') + server.register_function(new_session, 'session.new') + server.register_function(update_session, 'session.update') + server.register_function(poll_session, 'session.poll') server.serve_forever() @@ -676,7 +901,7 @@ if __name__ == '__main__': if len(sys.argv)>1: import jsonrpclib - server = jsonrpclib.Server('http://localhost:8080') + server = jsonrpclib.Server('http://%s:8081'%config.get('server','host')) cmd = sys.argv[1] if cmd == 'load': out = server.load(password) @@ -694,12 +919,13 @@ if __name__ == '__main__': out = server.blockchain.transaction.broadcast(sys.argv[2]) elif cmd == 'b': out = server.blocks() + else: + out = "Unknown command: '%s'" % cmd print out sys.exit(0) print "starting Electrum server" - print "cache:", config.get('server', 'cache') conf = DataStore.CONFIG_DEFAULTS args, argv = readconf.parse_argv( [], conf) @@ -714,9 +940,13 @@ if __name__ == '__main__': store.tx_cache = {} store.mempool_keys = {} - thread.start_new_thread(listen_thread, (store,)) - thread.start_new_thread(jsonrpc_thread, (store,)) + # supported protocols + thread.start_new_thread(native_server_thread, ()) + thread.start_new_thread(tcp_server_thread, ()) + thread.start_new_thread(http_server_thread, (store,)) + thread.start_new_thread(clean_session_thread, ()) + if (config.get('server','irc') == 'yes' ): thread.start_new_thread(irc_thread, ()) @@ -725,10 +955,30 @@ if __name__ == '__main__': dblock.acquire() store.catch_up() memorypool_update(store) + block_number = store.get_block_number(1) - dblock.release() + if block_number != old_block_number: + old_block_number = block_number + for session_id in sessions_sub_numblocks.keys(): + send_numblocks(session_id) + + except IOError: + print "IOError: cannot reach bitcoind" + block_number = 0 except: traceback.print_exc(file=sys.stdout) + block_number = 0 + finally: + dblock.release() + + # do addresses + while True: + try: + addr = address_queue.get(False) + except: + break + do_update_address(addr) + time.sleep(10) print "server stopped"