electrum

Electrum Bitcoin wallet
git clone https://git.parazyd.org/electrum
Log | Files | Refs | Submodules

commit ed234d3444e6095e749e5f272b769dd5df62516b
parent c8260249b0d9fc5ffb8a97a817ec9489563e4a7c
Author: SomberNight <somber.night@protonmail.com>
Date:   Thu, 27 Feb 2020 19:00:59 +0100

rename all TaskGroup() fields to "taskgroup"

for consistency

Diffstat:
Melectrum/interface.py | 8++++----
Melectrum/lnpeer.py | 4++--
Melectrum/lnverifier.py | 6+++---
Melectrum/lnworker.py | 2+-
Melectrum/network.py | 24++++++++++++------------
Melectrum/synchronizer.py | 6+++---
Melectrum/tests/test_network.py | 2+-
Melectrum/util.py | 8++++----
Melectrum/verifier.py | 6+++---
9 files changed, 33 insertions(+), 33 deletions(-)

diff --git a/electrum/interface.py b/electrum/interface.py @@ -254,8 +254,8 @@ class Interface(Logger): self.debug = False asyncio.run_coroutine_threadsafe( - self.network.main_taskgroup.spawn(self.run()), self.network.asyncio_loop) - self.group = SilentTaskGroup() + self.network.taskgroup.spawn(self.run()), self.network.asyncio_loop) + self.taskgroup = SilentTaskGroup() def diagnostic_name(self): return str(NetAddress(self.host, self.port)) @@ -370,7 +370,7 @@ class Interface(Logger): self.ready.cancel() return wrapper_func - @ignore_exceptions # do not kill main_taskgroup + @ignore_exceptions # do not kill network.taskgroup @log_exceptions @handle_disconnect async def run(self): @@ -489,7 +489,7 @@ class Interface(Logger): self.logger.info(f"connection established. version: {ver}") try: - async with self.group as group: + async with self.taskgroup as group: await group.spawn(self.ping) await group.spawn(self.run_fetch_blocks) await group.spawn(self.monitor_connection) diff --git a/electrum/lnpeer.py b/electrum/lnpeer.py @@ -94,7 +94,7 @@ class Peer(Logger): self._local_changed_events = defaultdict(asyncio.Event) self._remote_changed_events = defaultdict(asyncio.Event) Logger.__init__(self) - self.group = SilentTaskGroup() + self.taskgroup = SilentTaskGroup() def send_message(self, message_name: str, **kwargs): assert type(message_name) is str @@ -242,7 +242,7 @@ class Peer(Logger): @log_exceptions @handle_disconnect async def main_loop(self): - async with self.group as group: + async with self.taskgroup as group: await group.spawn(self._message_loop()) await group.spawn(self.query_gossip()) await group.spawn(self.process_gossip()) diff --git a/electrum/lnverifier.py b/electrum/lnverifier.py @@ -75,7 +75,7 @@ class LNChannelVerifier(NetworkJobOnDefaultServer): return True async def _start_tasks(self): - async with self.group as group: + async with self.taskgroup as group: await group.spawn(self.main) async def main(self): @@ -100,10 +100,10 @@ class LNChannelVerifier(NetworkJobOnDefaultServer): header = blockchain.read_header(block_height) if header is None: if block_height < constants.net.max_checkpoint(): - await self.group.spawn(self.network.request_chunk(block_height, None, can_return_early=True)) + await self.taskgroup.spawn(self.network.request_chunk(block_height, None, can_return_early=True)) continue self.started_verifying_channel.add(short_channel_id) - await self.group.spawn(self.verify_channel(block_height, short_channel_id)) + await self.taskgroup.spawn(self.verify_channel(block_height, short_channel_id)) #self.logger.info(f'requested short_channel_id {bh2u(short_channel_id)}') async def verify_channel(self, block_height: int, short_channel_id: ShortChannelID): diff --git a/electrum/lnworker.py b/electrum/lnworker.py @@ -1294,7 +1294,7 @@ class LNWallet(LNWorker): continue peer = self.peers.get(chan.node_id, None) if peer: - await peer.group.spawn(peer.reestablish_channel(chan)) + await peer.taskgroup.spawn(peer.reestablish_channel(chan)) else: await self.taskgroup.spawn(self.reestablish_peer_for_given_channel(chan)) diff --git a/electrum/network.py b/electrum/network.py @@ -268,7 +268,7 @@ class Network(Logger): if not self.default_server: self.default_server = pick_random_server() - self.main_taskgroup = None # type: TaskGroup + self.taskgroup = None # type: TaskGroup # locks self.restart_lock = asyncio.Lock() @@ -661,7 +661,7 @@ class Network(Logger): old_server = old_interface.server if old_interface else None # Stop any current interface in order to terminate subscriptions, - # and to cancel tasks in interface.group. + # and to cancel tasks in interface.taskgroup. # However, for headers sub, give preference to this interface # over unknown ones, i.e. start it again right away. if old_server and old_server != server: @@ -680,7 +680,7 @@ class Network(Logger): assert i.ready.done(), "interface we are switching to is not ready yet" blockchain_updated = i.blockchain != self.blockchain() self.interface = i - await i.group.spawn(self._request_server_info(i)) + await i.taskgroup.spawn(self._request_server_info(i)) self.trigger_callback('default_server_changed') self.default_server_changed_event.set() self.default_server_changed_event.clear() @@ -1118,8 +1118,8 @@ class Network(Logger): f.write(json.dumps(cp, indent=4)) async def _start(self): - assert not self.main_taskgroup - self.main_taskgroup = main_taskgroup = SilentTaskGroup() + assert not self.taskgroup + self.taskgroup = taskgroup = SilentTaskGroup() assert not self.interface and not self.interfaces assert not self.connecting and not self.server_queue self.logger.info('starting network') @@ -1135,11 +1135,11 @@ class Network(Logger): await self._init_headers_file() # note: if a task finishes with CancelledError, that # will NOT raise, and the group will keep the other tasks running - async with main_taskgroup as group: + async with taskgroup as group: await group.spawn(self._maintain_sessions()) [await group.spawn(job) for job in self._jobs] except BaseException as e: - self.logger.exception('main_taskgroup died.') + self.logger.exception('taskgroup died.') raise e asyncio.run_coroutine_threadsafe(main(), self.asyncio_loop) @@ -1158,10 +1158,10 @@ class Network(Logger): async def _stop(self, full_shutdown=False): self.logger.info("stopping network") try: - await asyncio.wait_for(self.main_taskgroup.cancel_remaining(), timeout=2) + await asyncio.wait_for(self.taskgroup.cancel_remaining(), timeout=2) except (asyncio.TimeoutError, asyncio.CancelledError) as e: self.logger.info(f"exc during main_taskgroup cancellation: {repr(e)}") - self.main_taskgroup = None # type: TaskGroup + self.taskgroup = None # type: TaskGroup self.interface = None # type: Interface self.interfaces = {} # type: Dict[str, Interface] self.connecting.clear() @@ -1196,7 +1196,7 @@ class Network(Logger): async def launch_already_queued_up_new_interfaces(): while self.server_queue.qsize() > 0: server = self.server_queue.get() - await self.main_taskgroup.spawn(self._run_new_interface(server)) + await self.taskgroup.spawn(self._run_new_interface(server)) async def maybe_queue_new_interfaces_to_be_launched_later(): now = time.time() for i in range(self.num_server - len(self.interfaces) - len(self.connecting)): @@ -1218,7 +1218,7 @@ class Network(Logger): await self._ensure_there_is_a_main_interface() if self.is_connected(): if self.config.is_fee_estimates_update_required(): - await self.interface.group.spawn(self._request_fee_estimates, self.interface) + await self.interface.taskgroup.spawn(self._request_fee_estimates, self.interface) while True: try: @@ -1228,7 +1228,7 @@ class Network(Logger): await maintain_main_interface() except asyncio.CancelledError: # suppress spurious cancellations - group = self.main_taskgroup + group = self.taskgroup if not group or group.closed(): raise await asyncio.sleep(0.1) diff --git a/electrum/synchronizer.py b/electrum/synchronizer.py @@ -75,7 +75,7 @@ class SynchronizerBase(NetworkJobOnDefaultServer): async def _start_tasks(self): try: - async with self.group as group: + async with self.taskgroup as group: await group.spawn(self.send_subscriptions()) await group.spawn(self.handle_status()) await group.spawn(self.main()) @@ -116,13 +116,13 @@ class SynchronizerBase(NetworkJobOnDefaultServer): while True: addr = await self.add_queue.get() - await self.group.spawn(subscribe_to_address, addr) + await self.taskgroup.spawn(subscribe_to_address, addr) async def handle_status(self): while True: h, status = await self.status_queue.get() addr = self.scripthash_to_address[h] - await self.group.spawn(self._on_address_status, addr, status) + await self.taskgroup.spawn(self._on_address_status, addr, status) self._processed_some_notifications = True def num_requests_sent_and_answered(self) -> Tuple[int, int]: diff --git a/electrum/tests/test_network.py b/electrum/tests/test_network.py @@ -16,7 +16,7 @@ class MockTaskGroup: async def spawn(self, x): return class MockNetwork: - main_taskgroup = MockTaskGroup() + taskgroup = MockTaskGroup() asyncio_loop = asyncio.get_event_loop() class MockInterface(Interface): diff --git a/electrum/util.py b/electrum/util.py @@ -1124,14 +1124,14 @@ class NetworkJobOnDefaultServer(Logger): """Initialise fields. Called every time the underlying server connection changes. """ - self.group = SilentTaskGroup() + self.taskgroup = SilentTaskGroup() async def _start(self, interface: 'Interface'): self.interface = interface - await interface.group.spawn(self._start_tasks) + await interface.taskgroup.spawn(self._start_tasks) async def _start_tasks(self): - """Start tasks in self.group. Called every time the underlying + """Start tasks in self.taskgroup. Called every time the underlying server connection changes. """ raise NotImplementedError() # implemented by subclasses @@ -1141,7 +1141,7 @@ class NetworkJobOnDefaultServer(Logger): await self._stop() async def _stop(self): - await self.group.cancel_remaining() + await self.taskgroup.cancel_remaining() @log_exceptions async def _restart(self, *args): diff --git a/electrum/verifier.py b/electrum/verifier.py @@ -59,7 +59,7 @@ class SPV(NetworkJobOnDefaultServer): self.requested_merkle = set() # txid set of pending requests async def _start_tasks(self): - async with self.group as group: + async with self.taskgroup as group: await group.spawn(self.main) def diagnostic_name(self): @@ -87,12 +87,12 @@ class SPV(NetworkJobOnDefaultServer): header = self.blockchain.read_header(tx_height) if header is None: if tx_height < constants.net.max_checkpoint(): - await self.group.spawn(self.network.request_chunk(tx_height, None, can_return_early=True)) + await self.taskgroup.spawn(self.network.request_chunk(tx_height, None, can_return_early=True)) continue # request now self.logger.info(f'requested merkle {tx_hash}') self.requested_merkle.add(tx_hash) - await self.group.spawn(self._request_and_verify_single_proof, tx_hash, tx_height) + await self.taskgroup.spawn(self._request_and_verify_single_proof, tx_hash, tx_height) async def _request_and_verify_single_proof(self, tx_hash, tx_height): try: