mirror of
https://github.com/LBRYFoundation/LBRY-Vault.git
synced 2025-08-30 08:51:32 +00:00
fix for memcache
This commit is contained in:
parent
fdc08e700c
commit
3b11b1d0ff
1 changed files with 22 additions and 11 deletions
|
@ -59,11 +59,9 @@ block_number = -1
|
||||||
sessions = {}
|
sessions = {}
|
||||||
sessions_last_time = {}
|
sessions_last_time = {}
|
||||||
dblock = thread.allocate_lock()
|
dblock = thread.allocate_lock()
|
||||||
|
|
||||||
peer_list = {}
|
peer_list = {}
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
class MyStore(Datastore_class):
|
class MyStore(Datastore_class):
|
||||||
|
|
||||||
def import_tx(self, tx, is_coinbase):
|
def import_tx(self, tx, is_coinbase):
|
||||||
|
@ -76,14 +74,14 @@ class MyStore(Datastore_class):
|
||||||
_hash = store.binout(row[6])
|
_hash = store.binout(row[6])
|
||||||
address = hash_to_address(chr(0), _hash)
|
address = hash_to_address(chr(0), _hash)
|
||||||
if self.tx_cache.has_key(address):
|
if self.tx_cache.has_key(address):
|
||||||
#print "cache: invalidating", address, self.ismempool
|
print "cache: invalidating", address, self.ismempool
|
||||||
self.tx_cache.pop(address)
|
self.tx_cache.pop(address)
|
||||||
outrows = self.get_tx_outputs(txid, False)
|
outrows = self.get_tx_outputs(txid, False)
|
||||||
for row in outrows:
|
for row in outrows:
|
||||||
_hash = store.binout(row[6])
|
_hash = store.binout(row[6])
|
||||||
address = hash_to_address(chr(0), _hash)
|
address = hash_to_address(chr(0), _hash)
|
||||||
if self.tx_cache.has_key(address):
|
if self.tx_cache.has_key(address):
|
||||||
#print "cache: invalidating", address, self.ismempool
|
print "cache: invalidating", address, self.ismempool
|
||||||
self.tx_cache.pop(address)
|
self.tx_cache.pop(address)
|
||||||
|
|
||||||
def safe_sql(self,sql, params=(), lock=True):
|
def safe_sql(self,sql, params=(), lock=True):
|
||||||
|
@ -245,12 +243,15 @@ class MyStore(Datastore_class):
|
||||||
rows = []
|
rows = []
|
||||||
rows += self.get_address_in_rows_memorypool( dbhash )
|
rows += self.get_address_in_rows_memorypool( dbhash )
|
||||||
rows += self.get_address_out_rows_memorypool( dbhash )
|
rows += self.get_address_out_rows_memorypool( dbhash )
|
||||||
address_has_no_mempool = (rows == [])
|
address_has_mempool = False
|
||||||
|
|
||||||
for row in rows:
|
for row in rows:
|
||||||
is_in, tx_hash, tx_id, pos, value = row
|
is_in, tx_hash, tx_id, pos, value = row
|
||||||
tx_hash = self.hashout_hex(tx_hash)
|
tx_hash = self.hashout_hex(tx_hash)
|
||||||
if tx_hash in known_tx:
|
if tx_hash in known_tx:
|
||||||
continue
|
continue
|
||||||
|
|
||||||
|
address_has_mempool = True
|
||||||
#print "mempool", tx_hash
|
#print "mempool", tx_hash
|
||||||
txpoint = {
|
txpoint = {
|
||||||
"nTime": 0,
|
"nTime": 0,
|
||||||
|
@ -298,7 +299,7 @@ class MyStore(Datastore_class):
|
||||||
if not row[4]: txpoint['raw_scriptPubKey'] = row[1]
|
if not row[4]: txpoint['raw_scriptPubKey'] = row[1]
|
||||||
|
|
||||||
# cache result
|
# cache result
|
||||||
if config.get('server','cache') == 'yes' and address_has_no_mempool:
|
if config.get('server','cache') == 'yes' and not address_has_mempool:
|
||||||
self.tx_cache[addr] = txpoints
|
self.tx_cache[addr] = txpoints
|
||||||
|
|
||||||
return txpoints
|
return txpoints
|
||||||
|
@ -378,19 +379,27 @@ def client_thread(ipaddr,conn):
|
||||||
elif cmd=='poll':
|
elif cmd=='poll':
|
||||||
session_id = data
|
session_id = data
|
||||||
addresses = sessions.get(session_id)
|
addresses = sessions.get(session_id)
|
||||||
if not addresses:
|
if addresses is None:
|
||||||
print "session not found", ipaddr
|
print time.asctime(), " Session not found", session_id, ipaddr
|
||||||
out = repr( (-1, {}))
|
out = repr( (-1, {}))
|
||||||
else:
|
else:
|
||||||
|
t1 = time.time()
|
||||||
sessions_last_time[session_id] = time.time()
|
sessions_last_time[session_id] = time.time()
|
||||||
ret = {}
|
ret = {}
|
||||||
|
k = 0
|
||||||
for addr in addresses:
|
for addr in addresses:
|
||||||
|
if store.tx_cache.get( addr ) is not None: k += 1
|
||||||
status = store.get_status( addr )
|
status = store.get_status( addr )
|
||||||
last_status = addresses.get( addr )
|
last_status = addresses.get( addr )
|
||||||
if last_status != status:
|
if last_status != status:
|
||||||
sessions[session_id][addr] = status
|
addresses[addr] = status
|
||||||
ret[addr] = status
|
ret[addr] = status
|
||||||
|
if ret:
|
||||||
|
sessions[session_id] = addresses
|
||||||
out = repr( (block_number, ret ) )
|
out = repr( (block_number, ret ) )
|
||||||
|
t2 = time.time() - t1
|
||||||
|
if t2 > 10:
|
||||||
|
print "high load:", session_id, "%d/%d"%(k,len(addresses)), t2
|
||||||
|
|
||||||
elif cmd == 'h':
|
elif cmd == 'h':
|
||||||
# history
|
# history
|
||||||
|
@ -467,8 +476,8 @@ def clean_session_thread():
|
||||||
time.sleep(30)
|
time.sleep(30)
|
||||||
t = time.time()
|
t = time.time()
|
||||||
for k,t0 in sessions_last_time.items():
|
for k,t0 in sessions_last_time.items():
|
||||||
if t - t0 > 60:
|
if t - t0 > 5*60:
|
||||||
print "lost session",k
|
print time.asctime(), "lost session",k
|
||||||
sessions.pop(k)
|
sessions.pop(k)
|
||||||
sessions_last_time.pop(k)
|
sessions_last_time.pop(k)
|
||||||
|
|
||||||
|
@ -547,6 +556,8 @@ if __name__ == '__main__':
|
||||||
|
|
||||||
|
|
||||||
print "starting Electrum server"
|
print "starting Electrum server"
|
||||||
|
print "cache:", config.get('server', 'cache')
|
||||||
|
|
||||||
conf = DataStore.CONFIG_DEFAULTS
|
conf = DataStore.CONFIG_DEFAULTS
|
||||||
args, argv = readconf.parse_argv( [], conf)
|
args, argv = readconf.parse_argv( [], conf)
|
||||||
args.dbtype= config.get('database','type')
|
args.dbtype= config.get('database','type')
|
||||||
|
|
Loading…
Add table
Reference in a new issue