electrum

Electrum Bitcoin wallet
git clone https://git.parazyd.org/electrum
Log | Files | Refs | Submodules

commit 46aa5c19584a1118deabb9dc2a6eba6012d38509
parent 9f188c087c379078443cfb589a1b5fbd0146dd21
Author: ThomasV <thomasv@electrum.org>
Date:   Tue,  5 Mar 2019 12:20:56 +0100

lnrouter: perform SQL requests in a separate thread. persist database.

Diffstat:
Melectrum/lnrouter.py | 81+++++++++++++++++++++++++++++++++++++++++++++++++++++---------------------------
1 file changed, 54 insertions(+), 27 deletions(-)

diff --git a/electrum/lnrouter.py b/electrum/lnrouter.py @@ -29,11 +29,11 @@ import queue import os import json import threading +import concurrent from collections import defaultdict from typing import Sequence, List, Tuple, Optional, Dict, NamedTuple, TYPE_CHECKING, Set import binascii import base64 -import asyncio from sqlalchemy import create_engine, Column, ForeignKey, Integer, String, DateTime, Boolean from sqlalchemy.pool import StaticPool @@ -212,43 +212,59 @@ class Address(Base): port = Column(Integer, primary_key=True) last_connected_date = Column(DateTime(), nullable=False) -class ChannelDB: + +class ChannelDB(PrintError): NUM_MAX_RECENT_PEERS = 20 def __init__(self, network: 'Network'): self.network = network - self.num_nodes = 0 self.num_channels = 0 - self.path = os.path.join(get_headers_dir(network.config), 'channel_db.sqlite3') - - # (intentionally not persisted) self._channel_updates_for_private_channels = {} # type: Dict[Tuple[bytes, bytes], dict] - self.ca_verifier = LNChannelVerifier(network, self) + self.db_requests = queue.Queue() + threading.Thread(target=self.sql_thread).start() - self.network.run_from_another_thread(self.sqlinit()) - - async def sqlinit(self): - """ - this has to run on the async thread since that is where - the lnpeer loop is running from, which will do call in here - """ + def sql_thread(self): engine = create_engine('sqlite:///' + self.path, pool_reset_on_return=None, poolclass=StaticPool)#, echo=True) self.DBSession = scoped_session(session_factory) self.DBSession.remove() self.DBSession.configure(bind=engine, autoflush=False) + if not os.path.exists(self.path): + Base.metadata.create_all(engine) + self._update_counts() + while self.network.asyncio_loop.is_running(): + try: + future, func, args, kwargs = self.db_requests.get(timeout=0.1) + except queue.Empty: + continue + try: + result = func(self, *args, **kwargs) + except BaseException as e: + future.set_exception(e) + continue + future.set_result(result) + # write + self.DBSession.commit() + self.DBSession.remove() + self.print_error("SQL thread terminated") - Base.metadata.drop_all(engine) - Base.metadata.create_all(engine) + def sql(func): + def wrapper(self, *args, **kwargs): + f = concurrent.futures.Future() + self.db_requests.put((f, func, args, kwargs)) + return f.result(timeout=10) + return wrapper - def update_counts(self): + # not @sql + def _update_counts(self): self.num_channels = self.DBSession.query(ChannelInfo).count() self.num_nodes = self.DBSession.query(NodeInfo).count() - def add_recent_peer(self, peer : LNPeerAddr): + @sql + def add_recent_peer(self, peer: LNPeerAddr): addr = self.DBSession.query(Address).filter_by(node_id = peer.pubkey.hex()).one_or_none() if addr is None: addr = Address(node_id = peer.pubkey.hex(), host = peer.host, port = peer.port, last_connected_date = datetime.datetime.now()) @@ -257,6 +273,7 @@ class ChannelDB: self.DBSession.add(addr) self.DBSession.commit() + @sql def get_200_randomly_sorted_nodes_not_in(self, node_ids_bytes): unshuffled = self.DBSession \ .query(NodeInfo) \ @@ -265,15 +282,14 @@ class ChannelDB: .all() return random.sample(unshuffled, len(unshuffled)) + @sql def nodes_get(self, node_id): - return self.network.run_from_another_thread(self._nodes_get(node_id)) - - async def _nodes_get(self, node_id): return self.DBSession \ .query(NodeInfo) \ .filter_by(node_id = node_id.hex()) \ .one_or_none() + @sql def get_last_good_address(self, node_id) -> Optional[LNPeerAddr]: adr_db = self.DBSession \ .query(Address) \ @@ -284,6 +300,7 @@ class ChannelDB: return None return LNPeerAddr(adr_db.host, adr_db.port, bytes.fromhex(adr_db.node_id)) + @sql def get_recent_peers(self): return [LNPeerAddr(x.host, x.port, bytes.fromhex(x.node_id)) for x in self.DBSession \ .query(Address) \ @@ -291,9 +308,11 @@ class ChannelDB: .order_by(Address.last_connected_date.desc()) \ .limit(self.NUM_MAX_RECENT_PEERS)] + @sql def get_channel_info(self, channel_id: bytes): - return self.chan_query_for_id(channel_id).one_or_none() + return self._chan_query_for_id(channel_id).one_or_none() + @sql def get_channels_for_node(self, node_id): """Returns the set of channels that have node_id as one of the endpoints.""" condition = or_( @@ -302,6 +321,7 @@ class ChannelDB: rows = self.DBSession.query(ChannelInfo).filter(condition).all() return [bytes.fromhex(x.short_channel_id) for x in rows] + @sql def missing_short_chan_ids(self) -> Set[int]: expr = not_(Policy.short_channel_id.in_(self.DBSession.query(ChannelInfo.short_channel_id))) chan_ids_from_policy = set(x[0] for x in self.DBSession.query(Policy.short_channel_id).filter(expr).all()) @@ -318,13 +338,15 @@ class ChannelDB: return chan_ids_from_id2 return set() + @sql def add_verified_channel_info(self, short_id, capacity): # called from lnchannelverifier - channel_info = self.get_channel_info(short_id) + channel_info = self._chan_query_for_id(short_id).one_or_none() channel_info.trusted = True channel_info.capacity = capacity self.DBSession.commit() + @sql @profiler def on_channel_announcement(self, msg_payloads, trusted=False): if type(msg_payloads) is dict: @@ -344,9 +366,10 @@ class ChannelDB: self.DBSession.add(channel_info) if not trusted: self.ca_verifier.add_new_channel_info(channel_info.short_channel_id, channel_info.msg_payload) self.DBSession.commit() + self._update_counts() self.network.trigger_callback('ln_status') - self.update_counts() + @sql @profiler def on_channel_update(self, msg_payloads, trusted=False): if type(msg_payloads) is dict: @@ -364,6 +387,7 @@ class ChannelDB: self._update_channel_info(channel_info, msg_payload, trusted=trusted) self.DBSession.commit() + @sql @profiler def on_node_announcement(self, msg_payloads): if type(msg_payloads) is dict: @@ -411,8 +435,8 @@ class ChannelDB: if old_addr: del old_addr self.DBSession.commit() + self._update_counts() self.network.trigger_callback('ln_status') - self.update_counts() def get_routing_policy_for_channel(self, start_node_id: bytes, short_channel_id: bytes) -> Optional[bytes]: @@ -431,11 +455,12 @@ class ChannelDB: short_channel_id = msg_payload['short_channel_id'] self._channel_updates_for_private_channels[(start_node_id, short_channel_id)] = msg_payload + @sql def remove_channel(self, short_channel_id): - self.chan_query_for_id(short_channel_id).delete('evaluate') + self._chan_query_for_id(short_channel_id).delete('evaluate') self.DBSession.commit() - def chan_query_for_id(self, short_channel_id) -> Query: + def _chan_query_for_id(self, short_channel_id) -> Query: return self.DBSession.query(ChannelInfo).filter_by(short_channel_id = short_channel_id.hex()) def print_graph(self, full_ids=False): @@ -495,6 +520,7 @@ class ChannelDB: old_policy.channel_flags = new_policy.channel_flags old_policy.timestamp = new_policy.timestamp + @sql def get_policy_for_node(self, node) -> Optional['Policy']: """ raises when initiator/non-initiator both unequal node @@ -507,6 +533,7 @@ class ChannelDB: n2 = self.DBSession.query(Policy).filter_by(short_channel_id = self.short_channel_id, start_node = self.node2_id).one_or_none() return n2 + @sql def get_node_addresses(self, node_info): return self.DBSession.query(Address).join(NodeInfo).filter_by(node_id = node_info.node_id).all()