Compare commits

..

No commits in common. "3a31c9d3388b28277bc9af7f3aca22217431072a" and "22c14e235e3390674796dfc3c462863395d2bfda" have entirely different histories.

7 changed files with 57 additions and 182 deletions

View File

@ -41,9 +41,6 @@ setup(
],
install_requires=[
# discovery subsys
'bidict',
# trio related
# proper range spec:
# https://packaging.python.org/en/latest/discussions/install-requires-vs-requirements/#id5

View File

@ -219,8 +219,7 @@ def daemon(
arb_addr: tuple[str, int],
):
'''
Run a daemon actor as a "remote registrar" and/or plain ol
separate actor (service) tree.
Run a daemon actor as a "remote arbiter".
'''
if loglevel in ('trace', 'debug'):

View File

@ -1,7 +1,6 @@
'''
Discovery subsystem via a "registrar" actor scenarios.
'''
"""
Actor "discovery" testing
"""
import os
import signal
import platform
@ -128,10 +127,7 @@ async def unpack_reg(actor_or_portal):
else:
msg = await actor_or_portal.run_from_ns('self', 'get_registry')
return {
tuple(key.split('.')): val
for key, val in msg.items()
}
return {tuple(key.split('.')): val for key, val in msg.items()}
async def spawn_and_check_registry(
@ -287,41 +283,37 @@ async def close_chans_before_nursery(
async with tractor.open_nursery() as tn:
portal1 = await tn.start_actor(
name='consumer1',
enable_modules=[__name__],
)
name='consumer1', enable_modules=[__name__])
portal2 = await tn.start_actor(
'consumer2',
enable_modules=[__name__],
)
'consumer2', enable_modules=[__name__])
async with (
portal1.open_stream_from(
# TODO: compact this back as was in last commit once
# 3.9+, see https://github.com/goodboy/tractor/issues/207
async with portal1.open_stream_from(
stream_forever
) as agen1:
async with portal2.open_stream_from(
stream_forever
) as agen1,
portal2.open_stream_from(
stream_forever
) as agen2,
):
async with trio.open_nursery() as n:
n.start_soon(streamer, agen1)
n.start_soon(cancel, use_signal, .5)
try:
await streamer(agen2)
finally:
# Kill the root nursery thus resulting in
# normal arbiter channel ops to fail during
# teardown. It doesn't seem like this is
# reliably triggered by an external SIGINT.
# tractor.current_actor()._root_nursery.cancel_scope.cancel()
) as agen2:
async with trio.open_nursery() as n:
n.start_soon(streamer, agen1)
n.start_soon(cancel, use_signal, .5)
try:
await streamer(agen2)
finally:
# Kill the root nursery thus resulting in
# normal arbiter channel ops to fail during
# teardown. It doesn't seem like this is
# reliably triggered by an external SIGINT.
# tractor.current_actor()._root_nursery.cancel_scope.cancel()
# XXX: THIS IS THE KEY THING that
# happens **before** exiting the
# actor nursery block
# XXX: THIS IS THE KEY THING that
# happens **before** exiting the
# actor nursery block
# also kill off channels cuz why not
await agen1.aclose()
await agen2.aclose()
# also kill off channels cuz why not
await agen1.aclose()
await agen2.aclose()
finally:
with trio.CancelScope(shield=True):
await trio.sleep(1)
@ -339,12 +331,10 @@ def test_close_channel_explicit(
use_signal,
arb_addr,
):
'''
Verify that closing a stream explicitly and killing the actor's
"""Verify that closing a stream explicitly and killing the actor's
"root nursery" **before** the containing nursery tears down also
results in subactor(s) deregistering from the arbiter.
'''
"""
with pytest.raises(KeyboardInterrupt):
trio.run(
partial(
@ -357,18 +347,16 @@ def test_close_channel_explicit(
@pytest.mark.parametrize('use_signal', [False, True])
def test_close_channel_explicit_remote_registrar(
def test_close_channel_explicit_remote_arbiter(
daemon,
start_method,
use_signal,
arb_addr,
):
'''
Verify that closing a stream explicitly and killing the actor's
"""Verify that closing a stream explicitly and killing the actor's
"root nursery" **before** the containing nursery tears down also
results in subactor(s) deregistering from the arbiter.
'''
"""
with pytest.raises(KeyboardInterrupt):
trio.run(
partial(
@ -378,51 +366,3 @@ def test_close_channel_explicit_remote_registrar(
remote_arbiter=True,
),
)
@tractor.context
async def kill_transport(
ctx: tractor.Context,
) -> None:
await ctx.started()
actor: tractor.Actor = tractor.current_actor()
actor.cancel_server()
await trio.sleep_forever()
# @pytest.mark.parametrize('use_signal', [False, True])
def test_stale_entry_is_deleted(
daemon,
start_method,
arb_addr,
):
'''
Ensure that when a stale entry is detected in the registrar's table
that the `find_actor()` API takes care of deleting the stale entry
and not delivering a bad portal.
'''
async def main():
name: str = 'transport_fails_actor'
regport: tractor.Portal
tn: tractor.ActorNursery
async with (
tractor.open_nursery() as tn,
tractor.get_registrar(*arb_addr) as regport,
):
ptl: tractor.Portal = await tn.start_actor(
name,
enable_modules=[__name__],
)
async with ptl.open_context(
kill_transport,
) as (first, ctx):
async with tractor.find_actor(name) as maybe_portal:
assert maybe_portal is None
await ptl.cancel_actor()
trio.run(main)

View File

@ -31,7 +31,6 @@ from ._streaming import (
)
from ._discovery import (
get_arbiter,
get_registrar,
find_actor,
wait_for_actor,
query_actor,
@ -78,7 +77,6 @@ __all__ = [
'find_actor',
'query_actor',
'get_arbiter',
'get_registrar',
'is_root_process',
'msg',
'open_actor_cluster',

View File

@ -35,7 +35,7 @@ from ._state import current_actor, _runtime_vars
@acm
async def get_registrar(
async def get_arbiter(
host: str,
port: int,
@ -56,14 +56,11 @@ async def get_registrar(
# (likely a re-entrant call from the arbiter actor)
yield LocalPortal(actor, Channel((host, port)))
else:
async with (
_connect_chan(host, port) as chan,
open_portal(chan) as arb_portal,
):
yield arb_portal
async with _connect_chan(host, port) as chan:
async with open_portal(chan) as arb_portal:
get_arbiter = get_registrar
yield arb_portal
@acm
@ -104,10 +101,7 @@ async def query_actor(
# TODO: return portals to all available actors - for now just
# the last one that registered
if (
name == 'arbiter'
and actor.is_arbiter
):
if name == 'arbiter' and actor.is_arbiter:
raise RuntimeError("The current actor is the arbiter")
yield sockaddr if sockaddr else None
@ -118,7 +112,7 @@ async def find_actor(
name: str,
arbiter_sockaddr: tuple[str, int] | None = None
) -> AsyncGenerator[Portal | None, None]:
) -> AsyncGenerator[Optional[Portal], None]:
'''
Ask the arbiter to find actor(s) by name.
@ -126,49 +120,17 @@ async def find_actor(
known to the arbiter.
'''
actor = current_actor()
async with get_arbiter(
*arbiter_sockaddr or actor._arb_addr
) as arb_portal:
sockaddr = await arb_portal.run_from_ns(
'self',
'find_actor',
name=name,
)
# TODO: return portals to all available actors - for now just
# the last one that registered
if (
name == 'arbiter'
and actor.is_arbiter
):
raise RuntimeError("The current actor is the arbiter")
async with query_actor(
name=name,
arbiter_sockaddr=arbiter_sockaddr,
) as sockaddr:
if sockaddr:
try:
async with _connect_chan(*sockaddr) as chan:
async with open_portal(chan) as portal:
yield portal
return
# most likely we were unable to connect the
# transport and there is likely a stale entry in
# the registry actor's table, thus we need to
# instruct it to clear that stale entry and then
# more silently (pretend there was no reason but
# to) indicate that the target actor can't be
# contacted at that addr.
except OSError:
# NOTE: ensure we delete the stale entry from the
# registar actor.
uid: tuple[str, str] = await arb_portal.run_from_ns(
'self',
'delete_sockaddr',
sockaddr=sockaddr,
)
yield None
async with _connect_chan(*sockaddr) as chan:
async with open_portal(chan) as portal:
yield portal
else:
yield None
@acm

View File

@ -40,7 +40,6 @@ from contextlib import ExitStack
import warnings
from async_generator import aclosing
from bidict import bidict
from exceptiongroup import BaseExceptionGroup
import trio # type: ignore
from trio_typing import TaskStatus
@ -1775,10 +1774,10 @@ class Arbiter(Actor):
def __init__(self, *args, **kwargs) -> None:
self._registry: bidict[
self._registry: dict[
tuple[str, str],
tuple[str, int],
] = bidict({})
] = {}
self._waiters: dict[
str,
# either an event to sync to receiving an actor uid (which
@ -1872,23 +1871,3 @@ class Arbiter(Actor):
entry: tuple = self._registry.pop(uid, None)
if entry is None:
log.warning(f'Request to de-register {uid} failed?')
async def delete_sockaddr(
self,
sockaddr: tuple[str, int],
) -> tuple[str, str]:
uid: tuple | None = self._registry.inverse.pop(
sockaddr,
None,
)
if uid:
log.warning(
f'Deleting registry entry for {sockaddr}@{uid}!'
)
else:
log.warning(
f'No registry entry for {sockaddr}@{uid}!'
)
return uid

View File

@ -216,7 +216,7 @@ def _run_asyncio_task(
try:
result = await coro
except BaseException as aio_err:
# log.exception('asyncio task errored:')
log.exception('asyncio task errored')
chan._aio_err = aio_err
raise
@ -300,7 +300,7 @@ def _run_asyncio_task(
elif task_err is None:
assert aio_err
aio_err.with_traceback(aio_err.__traceback__)
# log.error('infected task errorred')
log.error('infected task errorred')
# XXX: alway cancel the scope on error
# in case the trio task is blocking
@ -356,7 +356,7 @@ async def translate_aio_errors(
# relay cancel through to called ``asyncio`` task
assert chan._aio_task
chan._aio_task.cancel(
msg=f'`trio`-side caller task cancelled: {trio_task.name}'
msg=f'the `trio` caller task was cancelled: {trio_task.name}'
)
raise
@ -366,7 +366,7 @@ async def translate_aio_errors(
trio.ClosedResourceError,
# trio.BrokenResourceError,
):
aio_err: BaseException = chan._aio_err
aio_err = chan._aio_err
if (
task.cancelled() and
type(aio_err) is CancelledError