Compare commits
	
		
			118 Commits 
		
	
	
		
			d1f1cd3474
			...
			785eba9a22
		
	
	| Author | SHA1 | Date | 
|---|---|---|
|  | 785eba9a22 | |
|  | 90876833a8 | |
|  | 5f5e9bad27 | |
|  | ec724e11d1 | |
|  | 8d12ece8d6 | |
|  | 8908e1283b | |
|  | 51f72801d2 | |
|  | 01b5955cce | |
|  | b0ab77a99d | |
|  | f713a7a859 | |
|  | 85f1a66b41 | |
|  | fde6d84b9d | |
|  | 37b8d77d98 | |
|  | ef7a585570 | |
|  | 4d9f6e733a | |
|  | e57ac63f56 | |
|  | 32ef2764b0 | |
|  | b6e490295b | |
|  | 08e15f441f | |
|  | 8cdf3d819f | |
|  | 9b89f87ea1 | |
|  | 29c3895527 | |
|  | c302b74008 | |
|  | e42bc33bd6 | |
|  | bb916fd815 | |
|  | 27e4fc6660 | |
|  | 5feac62d3f | |
|  | 631fcc0471 | |
|  | 6e43fe1dd0 | |
|  | 187af24bcc | |
|  | 64fbad708e | |
|  | 6bd4903d01 | |
|  | f6c098d608 | |
|  | 7920e2980b | |
|  | e92c3e63ae | |
|  | 364ae6f6c8 | |
|  | 052b36f1e7 | |
|  | 261c48e126 | |
|  | c793f177f6 | |
|  | deb84e0b2c | |
|  | 3fa86c82fd | |
|  | 1407ea26d3 | |
|  | 2903431540 | |
|  | 1da0cba380 | |
|  | 36eb30daa3 | |
|  | 5bbd9b4e54 | |
|  | 89db16c693 | |
|  | 8056a9cf9f | |
|  | c8a3a6fb2b | |
|  | 897fa3d9f2 | |
|  | 639d6a981c | |
|  | 08941c22a6 | |
|  | 010d75248e | |
|  | 47ec7e7a49 | |
|  | a66caa2397 | |
|  | b1018a13fe | |
|  | 90287b9875 | |
|  | 15f99c313e | |
|  | 39027cd330 | |
|  | d7dc51a429 | |
|  | 7720564afb | |
|  | 24c309671d | |
|  | 680501aa10 | |
|  | cc7ad719d4 | |
|  | 5149b75f25 | |
|  | eb3337a593 | |
|  | b23b55f219 | |
|  | ccb717ecc7 | |
|  | e0be3397d1 | |
|  | af60417177 | |
|  | 526e5b91d9 | |
|  | 6012628223 | |
|  | 391d3faafd | |
|  | 34c1c1713d | |
|  | 5a9a3a457c | |
|  | 3e1258f840 | |
|  | 669c09c977 | |
|  | 1a126effec | |
|  | 6982c53386 | |
|  | 7d3d1e1afb | |
|  | 1222ef1e74 | |
|  | 1cc87bda50 | |
|  | aa2a7f050f | |
|  | 636b29e440 | |
|  | ef6fdbd09b | |
|  | 5dd38643a7 | |
|  | 6e06a04e14 | |
|  | c3e68e4133 | |
|  | 8722c6a1f7 | |
|  | 9f6b9e133d | |
|  | 52de75f1d4 | |
|  | 0546b7c684 | |
|  | 6a57f28619 | |
|  | 73a82fe422 | |
|  | 61ca5b1f61 | |
|  | 70416347c1 | |
|  | a7c06271a0 | |
|  | 9ec7913562 | |
|  | 9b07e7bdeb | |
|  | 3a53921535 | |
|  | 8114f0d327 | |
|  | 0bc4a18ce6 | |
|  | db843b361d | |
|  | f45d672b54 | |
|  | cd4df52608 | |
|  | aa06452c48 | |
|  | fae0ec9edf | |
|  | 506ddb72e1 | |
|  | d007e965f0 | |
|  | 9817fa5201 | |
|  | eef6e6779c | |
|  | 82a6e5bec0 | |
|  | b0eb1b7dd6 | |
|  | b1f8741575 | |
|  | 6e66020121 | |
|  | f090bf32f2 | |
|  | 39155f9633 | |
|  | 60036cfb72 | 
|  | @ -62,7 +62,9 @@ async def recv_and_spawn_net_killers( | |||
|     await ctx.started() | ||||
|     async with ( | ||||
|         ctx.open_stream() as stream, | ||||
|         trio.open_nursery() as n, | ||||
|         trio.open_nursery( | ||||
|             strict_exception_groups=False, | ||||
|         ) as tn, | ||||
|     ): | ||||
|         async for i in stream: | ||||
|             print(f'child echoing {i}') | ||||
|  | @ -77,11 +79,11 @@ async def recv_and_spawn_net_killers( | |||
|                 i >= break_ipc_after | ||||
|             ): | ||||
|                 broke_ipc = True | ||||
|                 n.start_soon( | ||||
|                 tn.start_soon( | ||||
|                     iter_ipc_stream, | ||||
|                     stream, | ||||
|                 ) | ||||
|                 n.start_soon( | ||||
|                 tn.start_soon( | ||||
|                     partial( | ||||
|                         break_ipc_then_error, | ||||
|                         stream=stream, | ||||
|  |  | |||
|  | @ -1,8 +1,16 @@ | |||
| ''' | ||||
| Examples of using the builtin `breakpoint()` from an `asyncio.Task` | ||||
| running in a subactor spawned with `infect_asyncio=True`. | ||||
| 
 | ||||
| ''' | ||||
| import asyncio | ||||
| 
 | ||||
| import trio | ||||
| import tractor | ||||
| from tractor import to_asyncio | ||||
| from tractor import ( | ||||
|     to_asyncio, | ||||
|     Portal, | ||||
| ) | ||||
| 
 | ||||
| 
 | ||||
| async def aio_sleep_forever(): | ||||
|  | @ -17,21 +25,21 @@ async def bp_then_error( | |||
| 
 | ||||
| ) -> None: | ||||
| 
 | ||||
|     # sync with ``trio``-side (caller) task | ||||
|     # sync with `trio`-side (caller) task | ||||
|     to_trio.send_nowait('start') | ||||
| 
 | ||||
|     # NOTE: what happens here inside the hook needs some refinement.. | ||||
|     # => seems like it's still `._debug._set_trace()` but | ||||
|     #    we set `Lock.local_task_in_debug = 'sync'`, we probably want | ||||
|     #    some further, at least, meta-data about the task/actoq in debug | ||||
|     #    in terms of making it clear it's asyncio mucking about. | ||||
|     breakpoint() | ||||
|     #    some further, at least, meta-data about the task/actor in debug | ||||
|     #    in terms of making it clear it's `asyncio` mucking about. | ||||
|     breakpoint()  # asyncio-side | ||||
| 
 | ||||
|     # short checkpoint / delay | ||||
|     await asyncio.sleep(0.5) | ||||
|     await asyncio.sleep(0.5)  # asyncio-side | ||||
| 
 | ||||
|     if raise_after_bp: | ||||
|         raise ValueError('blah') | ||||
|         raise ValueError('asyncio side error!') | ||||
| 
 | ||||
|     # TODO: test case with this so that it gets cancelled? | ||||
|     else: | ||||
|  | @ -49,23 +57,21 @@ async def trio_ctx( | |||
|     # this will block until the ``asyncio`` task sends a "first" | ||||
|     # message, see first line in above func. | ||||
|     async with ( | ||||
| 
 | ||||
|         to_asyncio.open_channel_from( | ||||
|             bp_then_error, | ||||
|             raise_after_bp=not bp_before_started, | ||||
|             # raise_after_bp=not bp_before_started, | ||||
|         ) as (first, chan), | ||||
| 
 | ||||
|         trio.open_nursery() as n, | ||||
|         trio.open_nursery() as tn, | ||||
|     ): | ||||
| 
 | ||||
|         assert first == 'start' | ||||
| 
 | ||||
|         if bp_before_started: | ||||
|             await tractor.breakpoint() | ||||
|             await tractor.pause()  # trio-side | ||||
| 
 | ||||
|         await ctx.started(first) | ||||
|         await ctx.started(first)  # trio-side | ||||
| 
 | ||||
|         n.start_soon( | ||||
|         tn.start_soon( | ||||
|             to_asyncio.run_task, | ||||
|             aio_sleep_forever, | ||||
|         ) | ||||
|  | @ -73,39 +79,50 @@ async def trio_ctx( | |||
| 
 | ||||
| 
 | ||||
| async def main( | ||||
|     bps_all_over: bool = False, | ||||
|     bps_all_over: bool = True, | ||||
| 
 | ||||
|     # TODO, WHICH OF THESE HAZ BUGZ? | ||||
|     cancel_from_root: bool = False, | ||||
|     err_from_root: bool = False, | ||||
| 
 | ||||
| ) -> None: | ||||
| 
 | ||||
|     async with tractor.open_nursery( | ||||
|         # debug_mode=True, | ||||
|     ) as n: | ||||
| 
 | ||||
|         p = await n.start_actor( | ||||
|         debug_mode=True, | ||||
|         maybe_enable_greenback=True, | ||||
|         # loglevel='devx', | ||||
|     ) as an: | ||||
|         ptl: Portal = await an.start_actor( | ||||
|             'aio_daemon', | ||||
|             enable_modules=[__name__], | ||||
|             infect_asyncio=True, | ||||
|             debug_mode=True, | ||||
|             loglevel='cancel', | ||||
|             # loglevel='cancel', | ||||
|         ) | ||||
| 
 | ||||
|         async with p.open_context( | ||||
|         async with ptl.open_context( | ||||
|             trio_ctx, | ||||
|             bp_before_started=bps_all_over, | ||||
|         ) as (ctx, first): | ||||
| 
 | ||||
|             assert first == 'start' | ||||
| 
 | ||||
|             if bps_all_over: | ||||
|                 await tractor.breakpoint() | ||||
|             # pause in parent to ensure no cross-actor | ||||
|             # locking problems exist! | ||||
|             await tractor.pause()  # trio-root | ||||
| 
 | ||||
|             if cancel_from_root: | ||||
|                 await ctx.cancel() | ||||
| 
 | ||||
|             if err_from_root: | ||||
|                 assert 0 | ||||
|             else: | ||||
|                 await trio.sleep_forever() | ||||
| 
 | ||||
|             # await trio.sleep_forever() | ||||
|             await ctx.cancel() | ||||
|             assert 0 | ||||
| 
 | ||||
|         # TODO: case where we cancel from trio-side while asyncio task | ||||
|         # has debugger lock? | ||||
|         # await p.cancel_actor() | ||||
|         # await ptl.cancel_actor() | ||||
| 
 | ||||
| 
 | ||||
| if __name__ == '__main__': | ||||
|  |  | |||
|  | @ -1,5 +1,5 @@ | |||
| ''' | ||||
| Fast fail test with a context. | ||||
| Fast fail test with a `Context`. | ||||
| 
 | ||||
| Ensure the partially initialized sub-actor process | ||||
| doesn't cause a hang on error/cancel of the parent | ||||
|  |  | |||
|  | @ -7,7 +7,7 @@ async def breakpoint_forever(): | |||
|     try: | ||||
|         while True: | ||||
|             yield 'yo' | ||||
|             await tractor.breakpoint() | ||||
|             await tractor.pause() | ||||
|     except BaseException: | ||||
|         tractor.log.get_console_log().exception( | ||||
|             'Cancelled while trying to enter pause point!' | ||||
|  | @ -21,11 +21,14 @@ async def name_error(): | |||
| 
 | ||||
| 
 | ||||
| async def main(): | ||||
|     """Test breakpoint in a streaming actor. | ||||
|     """ | ||||
|     ''' | ||||
|     Test breakpoint in a streaming actor. | ||||
| 
 | ||||
|     ''' | ||||
|     async with tractor.open_nursery( | ||||
|         debug_mode=True, | ||||
|         loglevel='cancel', | ||||
|         # loglevel='devx', | ||||
|     ) as n: | ||||
| 
 | ||||
|         p0 = await n.start_actor('bp_forever', enable_modules=[__name__]) | ||||
|  |  | |||
|  | @ -10,7 +10,7 @@ async def name_error(): | |||
| async def breakpoint_forever(): | ||||
|     "Indefinitely re-enter debugger in child actor." | ||||
|     while True: | ||||
|         await tractor.breakpoint() | ||||
|         await tractor.pause() | ||||
| 
 | ||||
|         # NOTE: if the test never sent 'q'/'quit' commands | ||||
|         # on the pdb repl, without this checkpoint line the | ||||
|  |  | |||
|  | @ -40,7 +40,7 @@ async def main(): | |||
|     """ | ||||
|     async with tractor.open_nursery( | ||||
|         debug_mode=True, | ||||
|         # loglevel='cancel', | ||||
|         loglevel='devx', | ||||
|     ) as n: | ||||
| 
 | ||||
|         # spawn both actors | ||||
|  |  | |||
|  | @ -6,7 +6,7 @@ async def breakpoint_forever(): | |||
|     "Indefinitely re-enter debugger in child actor." | ||||
|     while True: | ||||
|         await trio.sleep(0.1) | ||||
|         await tractor.breakpoint() | ||||
|         await tractor.pause() | ||||
| 
 | ||||
| 
 | ||||
| async def name_error(): | ||||
|  |  | |||
|  | @ -6,19 +6,44 @@ import tractor | |||
| 
 | ||||
| 
 | ||||
| async def main() -> None: | ||||
|     async with tractor.open_nursery(debug_mode=True) as an: | ||||
| 
 | ||||
|         assert os.environ['PYTHONBREAKPOINT'] == 'tractor._debug._set_trace' | ||||
|     # intially unset, no entry. | ||||
|     orig_pybp_var: int = os.environ.get('PYTHONBREAKPOINT') | ||||
|     assert orig_pybp_var in {None, "0"} | ||||
| 
 | ||||
|     async with tractor.open_nursery( | ||||
|         debug_mode=True, | ||||
|     ) as an: | ||||
|         assert an | ||||
|         assert ( | ||||
|             (pybp_var := os.environ['PYTHONBREAKPOINT']) | ||||
|             == | ||||
|             'tractor.devx._debug._sync_pause_from_builtin' | ||||
|         ) | ||||
| 
 | ||||
|         # TODO: an assert that verifies the hook has indeed been, hooked | ||||
|         # XD | ||||
|         assert sys.breakpointhook is not tractor._debug._set_trace | ||||
|         assert ( | ||||
|             (pybp_hook := sys.breakpointhook) | ||||
|             is not tractor.devx._debug._set_trace | ||||
|         ) | ||||
| 
 | ||||
|         breakpoint() | ||||
|         print( | ||||
|             f'$PYTHONOBREAKPOINT: {pybp_var!r}\n' | ||||
|             f'`sys.breakpointhook`: {pybp_hook!r}\n' | ||||
|         ) | ||||
|         breakpoint()  # first bp, tractor hook set. | ||||
| 
 | ||||
|     # TODO: an assert that verifies the hook is unhooked.. | ||||
|     # XXX AFTER EXIT (of actor-runtime) verify the hook is unset.. | ||||
|     # | ||||
|     # YES, this is weird but it's how stdlib docs say to do it.. | ||||
|     # https://docs.python.org/3/library/sys.html#sys.breakpointhook | ||||
|     assert os.environ.get('PYTHONBREAKPOINT') is orig_pybp_var | ||||
|     assert sys.breakpointhook | ||||
|     breakpoint() | ||||
| 
 | ||||
|     # now ensure a regular builtin pause still works | ||||
|     breakpoint()  # last bp, stdlib hook restored | ||||
| 
 | ||||
| 
 | ||||
| if __name__ == '__main__': | ||||
|     trio.run(main) | ||||
|  |  | |||
|  | @ -10,7 +10,7 @@ async def main(): | |||
| 
 | ||||
|         await trio.sleep(0.1) | ||||
| 
 | ||||
|         await tractor.breakpoint() | ||||
|         await tractor.pause() | ||||
| 
 | ||||
|         await trio.sleep(0.1) | ||||
| 
 | ||||
|  |  | |||
|  | @ -11,7 +11,7 @@ async def main( | |||
|         # loglevel='runtime', | ||||
|     ): | ||||
|         while True: | ||||
|             await tractor.breakpoint() | ||||
|             await tractor.pause() | ||||
| 
 | ||||
| 
 | ||||
| if __name__ == '__main__': | ||||
|  |  | |||
|  | @ -0,0 +1,83 @@ | |||
| ''' | ||||
| Verify we can dump a `stackscope` tree on a hang. | ||||
| 
 | ||||
| ''' | ||||
| import os | ||||
| import signal | ||||
| 
 | ||||
| import trio | ||||
| import tractor | ||||
| 
 | ||||
| @tractor.context | ||||
| async def start_n_shield_hang( | ||||
|     ctx: tractor.Context, | ||||
| ): | ||||
|     # actor: tractor.Actor = tractor.current_actor() | ||||
| 
 | ||||
|     # sync to parent-side task | ||||
|     await ctx.started(os.getpid()) | ||||
| 
 | ||||
|     print('Entering shield sleep..') | ||||
|     with trio.CancelScope(shield=True): | ||||
|         await trio.sleep_forever()  # in subactor | ||||
| 
 | ||||
|     # XXX NOTE ^^^ since this shields, we expect | ||||
|     # the zombie reaper (aka T800) to engage on | ||||
|     # SIGINT from the user and eventually hard-kill | ||||
|     # this subprocess! | ||||
| 
 | ||||
| 
 | ||||
| async def main( | ||||
|     from_test: bool = False, | ||||
| ) -> None: | ||||
| 
 | ||||
|     async with ( | ||||
|         tractor.open_nursery( | ||||
|             debug_mode=True, | ||||
|             enable_stack_on_sig=True, | ||||
|             # maybe_enable_greenback=False, | ||||
|             loglevel='devx', | ||||
|         ) as an, | ||||
|     ): | ||||
|         ptl: tractor.Portal  = await an.start_actor( | ||||
|             'hanger', | ||||
|             enable_modules=[__name__], | ||||
|             debug_mode=True, | ||||
|         ) | ||||
|         async with ptl.open_context( | ||||
|             start_n_shield_hang, | ||||
|         ) as (ctx, cpid): | ||||
| 
 | ||||
|             _, proc, _ = an._children[ptl.chan.uid] | ||||
|             assert cpid == proc.pid | ||||
| 
 | ||||
|             print( | ||||
|                 'Yo my child hanging..?\n' | ||||
|                 # "i'm a user who wants to see a `stackscope` tree!\n" | ||||
|             ) | ||||
| 
 | ||||
|             # XXX simulate the wrapping test's "user actions" | ||||
|             # (i.e. if a human didn't run this manually but wants to | ||||
|             # know what they should do to reproduce test behaviour) | ||||
|             if from_test: | ||||
|                 print( | ||||
|                     f'Sending SIGUSR1 to {cpid!r}!\n' | ||||
|                 ) | ||||
|                 os.kill( | ||||
|                     cpid, | ||||
|                     signal.SIGUSR1, | ||||
|                 ) | ||||
| 
 | ||||
|                 # simulate user cancelling program | ||||
|                 await trio.sleep(0.5) | ||||
|                 os.kill( | ||||
|                     os.getpid(), | ||||
|                     signal.SIGINT, | ||||
|                 ) | ||||
|             else: | ||||
|                 # actually let user send the ctl-c | ||||
|                 await trio.sleep_forever()  # in root | ||||
| 
 | ||||
| 
 | ||||
| if __name__ == '__main__': | ||||
|     trio.run(main) | ||||
|  | @ -4,9 +4,9 @@ import trio | |||
| 
 | ||||
| async def gen(): | ||||
|     yield 'yo' | ||||
|     await tractor.breakpoint() | ||||
|     await tractor.pause() | ||||
|     yield 'yo' | ||||
|     await tractor.breakpoint() | ||||
|     await tractor.pause() | ||||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
|  | @ -15,7 +15,7 @@ async def just_bp( | |||
| ) -> None: | ||||
| 
 | ||||
|     await ctx.started() | ||||
|     await tractor.breakpoint() | ||||
|     await tractor.pause() | ||||
| 
 | ||||
|     # TODO: bps and errors in this call.. | ||||
|     async for val in gen(): | ||||
|  |  | |||
|  | @ -4,6 +4,13 @@ import time | |||
| import trio | ||||
| import tractor | ||||
| 
 | ||||
| # TODO: only import these when not running from test harness? | ||||
| # can we detect `pexpect` usage maybe? | ||||
| # from tractor.devx._debug import ( | ||||
| #     get_lock, | ||||
| #     get_debug_req, | ||||
| # ) | ||||
| 
 | ||||
| 
 | ||||
| def sync_pause( | ||||
|     use_builtin: bool = False, | ||||
|  | @ -18,7 +25,13 @@ def sync_pause( | |||
|         breakpoint(hide_tb=hide_tb) | ||||
| 
 | ||||
|     else: | ||||
|         # TODO: maybe for testing some kind of cm style interface | ||||
|         # where the `._set_trace()` call doesn't happen until block | ||||
|         # exit? | ||||
|         # assert get_lock().ctx_in_debug is None | ||||
|         # assert get_debug_req().repl is None | ||||
|         tractor.pause_from_sync() | ||||
|         # assert get_debug_req().repl is None | ||||
| 
 | ||||
|     if error: | ||||
|         raise RuntimeError('yoyo sync code error') | ||||
|  | @ -41,10 +54,11 @@ async def start_n_sync_pause( | |||
| async def main() -> None: | ||||
|     async with ( | ||||
|         tractor.open_nursery( | ||||
|             # NOTE: required for pausing from sync funcs | ||||
|             maybe_enable_greenback=True, | ||||
|             debug_mode=True, | ||||
|             # loglevel='cancel', | ||||
|             maybe_enable_greenback=True, | ||||
|             enable_stack_on_sig=True, | ||||
|             # loglevel='warning', | ||||
|             # loglevel='devx', | ||||
|         ) as an, | ||||
|         trio.open_nursery() as tn, | ||||
|     ): | ||||
|  | @ -138,7 +152,9 @@ async def main() -> None: | |||
|                     # the case 2. from above still exists! | ||||
|                     use_builtin=True, | ||||
|                 ), | ||||
|                 abandon_on_cancel=False, | ||||
|                 # TODO: with this `False` we can hang!??! | ||||
|                 # abandon_on_cancel=False, | ||||
|                 abandon_on_cancel=True, | ||||
|                 thread_name='inline_root_bg_thread', | ||||
|             ) | ||||
| 
 | ||||
|  |  | |||
|  | @ -91,7 +91,7 @@ async def main() -> list[int]: | |||
|     an: ActorNursery | ||||
|     async with tractor.open_nursery( | ||||
|         loglevel='cancel', | ||||
|         debug_mode=True, | ||||
|         # debug_mode=True, | ||||
|     ) as an: | ||||
| 
 | ||||
|         seed = int(1e3) | ||||
|  |  | |||
|  | @ -3,20 +3,18 @@ import trio | |||
| import tractor | ||||
| 
 | ||||
| 
 | ||||
| async def sleepy_jane(): | ||||
|     uid = tractor.current_actor().uid | ||||
| async def sleepy_jane() -> None: | ||||
|     uid: tuple = tractor.current_actor().uid | ||||
|     print(f'Yo i am actor {uid}') | ||||
|     await trio.sleep_forever() | ||||
| 
 | ||||
| 
 | ||||
| async def main(): | ||||
|     ''' | ||||
|     Spawn a flat actor cluster, with one process per | ||||
|     detected core. | ||||
|     Spawn a flat actor cluster, with one process per detected core. | ||||
| 
 | ||||
|     ''' | ||||
|     portal_map: dict[str, tractor.Portal] | ||||
|     results: dict[str, str] | ||||
| 
 | ||||
|     # look at this hip new syntax! | ||||
|     async with ( | ||||
|  | @ -25,11 +23,16 @@ async def main(): | |||
|             modules=[__name__] | ||||
|         ) as portal_map, | ||||
| 
 | ||||
|         trio.open_nursery() as n, | ||||
|         trio.open_nursery( | ||||
|             strict_exception_groups=False, | ||||
|         ) as tn, | ||||
|     ): | ||||
| 
 | ||||
|         for (name, portal) in portal_map.items(): | ||||
|             n.start_soon(portal.run, sleepy_jane) | ||||
|             tn.start_soon( | ||||
|                 portal.run, | ||||
|                 sleepy_jane, | ||||
|             ) | ||||
| 
 | ||||
|         await trio.sleep(0.5) | ||||
| 
 | ||||
|  | @ -41,4 +44,4 @@ if __name__ == '__main__': | |||
|     try: | ||||
|         trio.run(main) | ||||
|     except KeyboardInterrupt: | ||||
|         pass | ||||
|         print('trio cancelled by KBI') | ||||
|  |  | |||
|  | @ -0,0 +1,18 @@ | |||
| First generate a built disti: | ||||
| 
 | ||||
| ``` | ||||
| python -m pip install --upgrade build | ||||
| python -m build --sdist --outdir dist/alpha5/ | ||||
| ``` | ||||
| 
 | ||||
| Then try a test ``pypi`` upload: | ||||
| 
 | ||||
| ``` | ||||
| python -m twine upload --repository testpypi dist/alpha5/* | ||||
| ``` | ||||
| 
 | ||||
| The push to `pypi` for realz. | ||||
| 
 | ||||
| ``` | ||||
| python -m twine upload --repository testpypi dist/alpha5/* | ||||
| ``` | ||||
|  | @ -37,16 +37,14 @@ dependencies = [ | |||
|   # https://packaging.python.org/en/latest/discussions/install-requires-vs-requirements/#id5 | ||||
|   # TODO, for 3.13 we must go go `0.27` which means we have to | ||||
|   # disable strict egs or port to handling them internally! | ||||
|   # trio='^0.27' | ||||
|   "trio>=0.24,<0.25", | ||||
|   "trio>0.27", | ||||
|   "tricycle>=0.4.1,<0.5", | ||||
|   "wrapt>=1.16.0,<2", | ||||
|   "colorlog>=6.8.2,<7", | ||||
|   # built-in multi-actor `pdb` REPL | ||||
|   "pdbp>=1.5.0,<2", | ||||
|   "pdbp>=1.6,<2", # windows only (from `pdbp`) | ||||
|   # typed IPC msging | ||||
|   # TODO, get back on release once 3.13 support is out! | ||||
|   "msgspec", | ||||
|   "msgspec>=0.19.0", | ||||
| ] | ||||
| 
 | ||||
| # ------ project ------ | ||||
|  | @ -56,18 +54,14 @@ dev = [ | |||
|   # test suite | ||||
|   # TODO: maybe some of these layout choices? | ||||
|   # https://docs.pytest.org/en/8.0.x/explanation/goodpractices.html#choosing-a-test-layout-import-rules | ||||
|   "pytest>=8.2.0,<9", | ||||
|   "pytest>=8.3.5", | ||||
|   "pexpect>=4.9.0,<5", | ||||
|   # `tractor.devx` tooling | ||||
|   "greenback>=1.2.1,<2", | ||||
|   "stackscope>=0.2.2,<0.3", | ||||
| 
 | ||||
|   # xonsh usage/integration (namely as @goodboy's sh of choice Bp) | ||||
|   "xonsh>=0.19.1", | ||||
|   "xontrib-vox>=0.0.1,<0.0.2", | ||||
|   "prompt-toolkit>=3.0.43,<4", | ||||
|   "xonsh-vox-tabcomplete>=0.5,<0.6", | ||||
|   "pyperclip>=1.9.0", | ||||
|   "prompt-toolkit>=3.0.50", | ||||
|   "xonsh>=0.19.2", | ||||
| ] | ||||
| # TODO, add these with sane versions; were originally in | ||||
| # `requirements-docs.txt`.. | ||||
|  | @ -78,21 +72,39 @@ dev = [ | |||
| 
 | ||||
| # ------ dependency-groups ------ | ||||
| 
 | ||||
| # ------ dependency-groups ------ | ||||
| 
 | ||||
| [tool.uv.sources] | ||||
| msgspec = { git = "https://github.com/jcrist/msgspec.git" } | ||||
| # XXX NOTE, only for @goodboy's hacking on `pprint(sort_dicts=False)` | ||||
| # for the `pp` alias.. | ||||
| # pdbp = { path = "../pdbp", editable = true } | ||||
| 
 | ||||
| # ------ tool.uv.sources ------ | ||||
| # TODO, distributed (multi-host) extensions | ||||
| # linux kernel networking | ||||
| # 'pyroute2 | ||||
| 
 | ||||
| # ------ tool.uv.sources ------ | ||||
| 
 | ||||
| [tool.uv] | ||||
| # XXX NOTE, prefer the sys python bc apparently the distis from | ||||
| # `astral` are built in a way that breaks `pdbp`+`tabcompleter`'s | ||||
| # likely due to linking against `libedit` over `readline`.. | ||||
| # |_https://docs.astral.sh/uv/concepts/python-versions/#managed-python-distributions | ||||
| # |_https://gregoryszorc.com/docs/python-build-standalone/main/quirks.html#use-of-libedit-on-linux | ||||
| # | ||||
| # https://docs.astral.sh/uv/reference/settings/#python-preference | ||||
| python-preference = 'system' | ||||
| 
 | ||||
| # ------ tool.uv ------ | ||||
| 
 | ||||
| [tool.hatch.build.targets.sdist] | ||||
| include = ["tractor"] | ||||
| 
 | ||||
| [tool.hatch.build.targets.wheel] | ||||
| include = ["tractor"] | ||||
| 
 | ||||
| # ------ dependency-groups ------ | ||||
| # ------ tool.hatch ------ | ||||
| 
 | ||||
| [tool.towncrier] | ||||
| package = "tractor" | ||||
|  | @ -142,3 +154,5 @@ log_cli = false | |||
| # TODO: maybe some of these layout choices? | ||||
| # https://docs.pytest.org/en/8.0.x/explanation/goodpractices.html#choosing-a-test-layout-import-rules | ||||
| # pythonpath = "src" | ||||
| 
 | ||||
| # ------ tool.pytest ------ | ||||
|  |  | |||
|  | @ -75,7 +75,10 @@ def pytest_configure(config): | |||
| 
 | ||||
| @pytest.fixture(scope='session') | ||||
| def debug_mode(request): | ||||
|     return request.config.option.tractor_debug_mode | ||||
|     debug_mode: bool = request.config.option.tractor_debug_mode | ||||
|     # if debug_mode: | ||||
|     #     breakpoint() | ||||
|     return debug_mode | ||||
| 
 | ||||
| 
 | ||||
| @pytest.fixture(scope='session', autouse=True) | ||||
|  | @ -92,6 +95,12 @@ def spawn_backend(request) -> str: | |||
|     return request.config.option.spawn_backend | ||||
| 
 | ||||
| 
 | ||||
| # @pytest.fixture(scope='function', autouse=True) | ||||
| # def debug_enabled(request) -> str: | ||||
| #     from tractor import _state | ||||
| #     if _state._runtime_vars['_debug_mode']: | ||||
| #         breakpoint() | ||||
| 
 | ||||
| _ci_env: bool = os.environ.get('CI', False) | ||||
| 
 | ||||
| 
 | ||||
|  | @ -150,6 +159,18 @@ def pytest_generate_tests(metafunc): | |||
|         metafunc.parametrize("start_method", [spawn_backend], scope='module') | ||||
| 
 | ||||
| 
 | ||||
| # TODO: a way to let test scripts (like from `examples/`) | ||||
| # guarantee they won't registry addr collide! | ||||
| # @pytest.fixture | ||||
| # def open_test_runtime( | ||||
| #     reg_addr: tuple, | ||||
| # ) -> AsyncContextManager: | ||||
| #     return partial( | ||||
| #         tractor.open_nursery, | ||||
| #         registry_addrs=[reg_addr], | ||||
| #     ) | ||||
| 
 | ||||
| 
 | ||||
| def sig_prog(proc, sig): | ||||
|     "Kill the actor-process with ``sig``." | ||||
|     proc.send_signal(sig) | ||||
|  |  | |||
|  | @ -0,0 +1,243 @@ | |||
| ''' | ||||
| `tractor.devx.*` tooling sub-pkg test space. | ||||
| 
 | ||||
| ''' | ||||
| import time | ||||
| from typing import ( | ||||
|     Callable, | ||||
| ) | ||||
| 
 | ||||
| import pytest | ||||
| from pexpect.exceptions import ( | ||||
|     TIMEOUT, | ||||
| ) | ||||
| from pexpect.spawnbase import SpawnBase | ||||
| 
 | ||||
| from tractor._testing import ( | ||||
|     mk_cmd, | ||||
| ) | ||||
| from tractor.devx._debug import ( | ||||
|     _pause_msg as _pause_msg, | ||||
|     _crash_msg as _crash_msg, | ||||
|     _repl_fail_msg as _repl_fail_msg, | ||||
|     _ctlc_ignore_header as _ctlc_ignore_header, | ||||
| ) | ||||
| from ..conftest import ( | ||||
|     _ci_env, | ||||
| ) | ||||
| 
 | ||||
| 
 | ||||
| @pytest.fixture | ||||
| def spawn( | ||||
|     start_method, | ||||
|     testdir: pytest.Pytester, | ||||
|     reg_addr: tuple[str, int], | ||||
| 
 | ||||
| ) -> Callable[[str], None]: | ||||
|     ''' | ||||
|     Use the `pexpect` module shipped via `testdir.spawn()` to | ||||
|     run an `./examples/..` script by name. | ||||
| 
 | ||||
|     ''' | ||||
|     if start_method != 'trio': | ||||
|         pytest.skip( | ||||
|             '`pexpect` based tests only supported on `trio` backend' | ||||
|         ) | ||||
| 
 | ||||
|     def unset_colors(): | ||||
|         ''' | ||||
|         Python 3.13 introduced colored tracebacks that break patt | ||||
|         matching, | ||||
| 
 | ||||
|         https://docs.python.org/3/using/cmdline.html#envvar-PYTHON_COLORS | ||||
|         https://docs.python.org/3/using/cmdline.html#using-on-controlling-color | ||||
| 
 | ||||
|         ''' | ||||
|         import os | ||||
|         os.environ['PYTHON_COLORS'] = '0' | ||||
| 
 | ||||
|     def _spawn( | ||||
|         cmd: str, | ||||
|         **mkcmd_kwargs, | ||||
|     ): | ||||
|         unset_colors() | ||||
|         return testdir.spawn( | ||||
|             cmd=mk_cmd( | ||||
|                 cmd, | ||||
|                 **mkcmd_kwargs, | ||||
|             ), | ||||
|             expect_timeout=3, | ||||
|             # preexec_fn=unset_colors, | ||||
|             # ^TODO? get `pytest` core to expose underlying | ||||
|             # `pexpect.spawn()` stuff? | ||||
|         ) | ||||
| 
 | ||||
|     # such that test-dep can pass input script name. | ||||
|     return _spawn | ||||
| 
 | ||||
| 
 | ||||
| @pytest.fixture( | ||||
|     params=[False, True], | ||||
|     ids='ctl-c={}'.format, | ||||
| ) | ||||
| def ctlc( | ||||
|     request, | ||||
|     ci_env: bool, | ||||
| 
 | ||||
| ) -> bool: | ||||
| 
 | ||||
|     use_ctlc = request.param | ||||
| 
 | ||||
|     node = request.node | ||||
|     markers = node.own_markers | ||||
|     for mark in markers: | ||||
|         if mark.name == 'has_nested_actors': | ||||
|             pytest.skip( | ||||
|                 f'Test {node} has nested actors and fails with Ctrl-C.\n' | ||||
|                 f'The test can sometimes run fine locally but until' | ||||
|                 ' we solve' 'this issue this CI test will be xfail:\n' | ||||
|                 'https://github.com/goodboy/tractor/issues/320' | ||||
|             ) | ||||
| 
 | ||||
|         if mark.name == 'ctlcs_bish': | ||||
|             pytest.skip( | ||||
|                 f'Test {node} prolly uses something from the stdlib (namely `asyncio`..)\n' | ||||
|                 f'The test and/or underlying example script can *sometimes* run fine ' | ||||
|                 f'locally but more then likely until the cpython peeps get their sh#$ together, ' | ||||
|                 f'this test will definitely not behave like `trio` under SIGINT..\n' | ||||
|             ) | ||||
| 
 | ||||
|     if use_ctlc: | ||||
|         # XXX: disable pygments highlighting for auto-tests | ||||
|         # since some envs (like actions CI) will struggle | ||||
|         # the the added color-char encoding.. | ||||
|         from tractor.devx._debug import TractorConfig | ||||
|         TractorConfig.use_pygements = False | ||||
| 
 | ||||
|     yield use_ctlc | ||||
| 
 | ||||
| 
 | ||||
| def expect( | ||||
|     child, | ||||
| 
 | ||||
|     # normally a `pdb` prompt by default | ||||
|     patt: str, | ||||
| 
 | ||||
|     **kwargs, | ||||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Expect wrapper that prints last seen console | ||||
|     data before failing. | ||||
| 
 | ||||
|     ''' | ||||
|     try: | ||||
|         child.expect( | ||||
|             patt, | ||||
|             **kwargs, | ||||
|         ) | ||||
|     except TIMEOUT: | ||||
|         before = str(child.before.decode()) | ||||
|         print(before) | ||||
|         raise | ||||
| 
 | ||||
| 
 | ||||
| PROMPT = r"\(Pdb\+\)" | ||||
| 
 | ||||
| 
 | ||||
| def in_prompt_msg( | ||||
|     child: SpawnBase, | ||||
|     parts: list[str], | ||||
| 
 | ||||
|     pause_on_false: bool = False, | ||||
|     err_on_false: bool = False, | ||||
|     print_prompt_on_false: bool = True, | ||||
| 
 | ||||
| ) -> bool: | ||||
|     ''' | ||||
|     Predicate check if (the prompt's) std-streams output has all | ||||
|     `str`-parts in it. | ||||
| 
 | ||||
|     Can be used in test asserts for bulk matching expected | ||||
|     log/REPL output for a given `pdb` interact point. | ||||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = False | ||||
| 
 | ||||
|     before: str = str(child.before.decode()) | ||||
|     for part in parts: | ||||
|         if part not in before: | ||||
|             if pause_on_false: | ||||
|                 import pdbp | ||||
|                 pdbp.set_trace() | ||||
| 
 | ||||
|             if print_prompt_on_false: | ||||
|                 print(before) | ||||
| 
 | ||||
|             if err_on_false: | ||||
|                 raise ValueError( | ||||
|                     f'Could not find pattern in `before` output?\n' | ||||
|                     f'part: {part!r}\n' | ||||
|                 ) | ||||
|             return False | ||||
| 
 | ||||
|     return True | ||||
| 
 | ||||
| 
 | ||||
| # TODO: todo support terminal color-chars stripping so we can match | ||||
| # against call stack frame output from the the 'll' command the like! | ||||
| # -[ ] SO answer for stipping ANSI codes: https://stackoverflow.com/a/14693789 | ||||
| def assert_before( | ||||
|     child: SpawnBase, | ||||
|     patts: list[str], | ||||
| 
 | ||||
|     **kwargs, | ||||
| 
 | ||||
| ) -> None: | ||||
|     __tracebackhide__: bool = False | ||||
| 
 | ||||
|     assert in_prompt_msg( | ||||
|         child=child, | ||||
|         parts=patts, | ||||
| 
 | ||||
|         # since this is an "assert" helper ;) | ||||
|         err_on_false=True, | ||||
|         **kwargs | ||||
|     ) | ||||
| 
 | ||||
| 
 | ||||
| def do_ctlc( | ||||
|     child, | ||||
|     count: int = 3, | ||||
|     delay: float = 0.1, | ||||
|     patt: str|None = None, | ||||
| 
 | ||||
|     # expect repl UX to reprint the prompt after every | ||||
|     # ctrl-c send. | ||||
|     # XXX: no idea but, in CI this never seems to work even on 3.10 so | ||||
|     # needs some further investigation potentially... | ||||
|     expect_prompt: bool = not _ci_env, | ||||
| 
 | ||||
| ) -> str|None: | ||||
| 
 | ||||
|     before: str|None = None | ||||
| 
 | ||||
|     # make sure ctl-c sends don't do anything but repeat output | ||||
|     for _ in range(count): | ||||
|         time.sleep(delay) | ||||
|         child.sendcontrol('c') | ||||
| 
 | ||||
|         # TODO: figure out why this makes CI fail.. | ||||
|         # if you run this test manually it works just fine.. | ||||
|         if expect_prompt: | ||||
|             time.sleep(delay) | ||||
|             child.expect(PROMPT) | ||||
|             before = str(child.before.decode()) | ||||
|             time.sleep(delay) | ||||
| 
 | ||||
|             if patt: | ||||
|                 # should see the last line on console | ||||
|                 assert patt in before | ||||
| 
 | ||||
|     # return the console content up to the final prompt | ||||
|     return before | ||||
|  | @ -13,26 +13,25 @@ TODO: | |||
| from functools import partial | ||||
| import itertools | ||||
| import platform | ||||
| import pathlib | ||||
| import time | ||||
| 
 | ||||
| import pytest | ||||
| import pexpect | ||||
| from pexpect.exceptions import ( | ||||
|     TIMEOUT, | ||||
|     EOF, | ||||
| ) | ||||
| 
 | ||||
| from tractor._testing import ( | ||||
|     examples_dir, | ||||
| ) | ||||
| from tractor.devx._debug import ( | ||||
| from .conftest import ( | ||||
|     do_ctlc, | ||||
|     PROMPT, | ||||
|     _pause_msg, | ||||
|     _crash_msg, | ||||
|     _repl_fail_msg, | ||||
| ) | ||||
| from .conftest import ( | ||||
|     _ci_env, | ||||
|     expect, | ||||
|     in_prompt_msg, | ||||
|     assert_before, | ||||
| ) | ||||
| 
 | ||||
| # TODO: The next great debugger audit could be done by you! | ||||
|  | @ -52,15 +51,6 @@ if platform.system() == 'Windows': | |||
|     ) | ||||
| 
 | ||||
| 
 | ||||
| def mk_cmd(ex_name: str) -> str: | ||||
|     ''' | ||||
|     Generate a command suitable to pass to ``pexpect.spawn()``. | ||||
| 
 | ||||
|     ''' | ||||
|     script_path: pathlib.Path = examples_dir() / 'debugging' / f'{ex_name}.py' | ||||
|     return ' '.join(['python', str(script_path)]) | ||||
| 
 | ||||
| 
 | ||||
| # TODO: was trying to this xfail style but some weird bug i see in CI | ||||
| # that's happening at collect time.. pretty soon gonna dump actions i'm | ||||
| # thinkin... | ||||
|  | @ -79,142 +69,6 @@ has_nested_actors = pytest.mark.has_nested_actors | |||
| # ) | ||||
| 
 | ||||
| 
 | ||||
| @pytest.fixture | ||||
| def spawn( | ||||
|     start_method, | ||||
|     testdir, | ||||
|     reg_addr, | ||||
| ) -> 'pexpect.spawn': | ||||
| 
 | ||||
|     if start_method != 'trio': | ||||
|         pytest.skip( | ||||
|             "Debugger tests are only supported on the trio backend" | ||||
|         ) | ||||
| 
 | ||||
|     def _spawn(cmd): | ||||
|         return testdir.spawn( | ||||
|             cmd=mk_cmd(cmd), | ||||
|             expect_timeout=3, | ||||
|         ) | ||||
| 
 | ||||
|     return _spawn | ||||
| 
 | ||||
| 
 | ||||
| PROMPT = r"\(Pdb\+\)" | ||||
| 
 | ||||
| 
 | ||||
| def expect( | ||||
|     child, | ||||
| 
 | ||||
|     # prompt by default | ||||
|     patt: str = PROMPT, | ||||
| 
 | ||||
|     **kwargs, | ||||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Expect wrapper that prints last seen console | ||||
|     data before failing. | ||||
| 
 | ||||
|     ''' | ||||
|     try: | ||||
|         child.expect( | ||||
|             patt, | ||||
|             **kwargs, | ||||
|         ) | ||||
|     except TIMEOUT: | ||||
|         before = str(child.before.decode()) | ||||
|         print(before) | ||||
|         raise | ||||
| 
 | ||||
| 
 | ||||
| def in_prompt_msg( | ||||
|     prompt: str, | ||||
|     parts: list[str], | ||||
| 
 | ||||
|     pause_on_false: bool = False, | ||||
|     print_prompt_on_false: bool = True, | ||||
| 
 | ||||
| ) -> bool: | ||||
|     ''' | ||||
|     Predicate check if (the prompt's) std-streams output has all | ||||
|     `str`-parts in it. | ||||
| 
 | ||||
|     Can be used in test asserts for bulk matching expected | ||||
|     log/REPL output for a given `pdb` interact point. | ||||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = False | ||||
| 
 | ||||
|     for part in parts: | ||||
|         if part not in prompt: | ||||
|             if pause_on_false: | ||||
|                 import pdbp | ||||
|                 pdbp.set_trace() | ||||
| 
 | ||||
|             if print_prompt_on_false: | ||||
|                 print(prompt) | ||||
| 
 | ||||
|             return False | ||||
| 
 | ||||
|     return True | ||||
| 
 | ||||
| 
 | ||||
| # TODO: todo support terminal color-chars stripping so we can match | ||||
| # against call stack frame output from the the 'll' command the like! | ||||
| # -[ ] SO answer for stipping ANSI codes: https://stackoverflow.com/a/14693789 | ||||
| def assert_before( | ||||
|     child, | ||||
|     patts: list[str], | ||||
| 
 | ||||
|     **kwargs, | ||||
| 
 | ||||
| ) -> None: | ||||
|     __tracebackhide__: bool = False | ||||
| 
 | ||||
|     # as in before the prompt end | ||||
|     before: str = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         prompt=before, | ||||
|         parts=patts, | ||||
| 
 | ||||
|         **kwargs | ||||
|     ) | ||||
| 
 | ||||
| 
 | ||||
| @pytest.fixture( | ||||
|     params=[False, True], | ||||
|     ids='ctl-c={}'.format, | ||||
| ) | ||||
| def ctlc( | ||||
|     request, | ||||
|     ci_env: bool, | ||||
| 
 | ||||
| ) -> bool: | ||||
| 
 | ||||
|     use_ctlc = request.param | ||||
| 
 | ||||
|     node = request.node | ||||
|     markers = node.own_markers | ||||
|     for mark in markers: | ||||
|         if mark.name == 'has_nested_actors': | ||||
|             pytest.skip( | ||||
|                 f'Test {node} has nested actors and fails with Ctrl-C.\n' | ||||
|                 f'The test can sometimes run fine locally but until' | ||||
|                 ' we solve' 'this issue this CI test will be xfail:\n' | ||||
|                 'https://github.com/goodboy/tractor/issues/320' | ||||
|             ) | ||||
| 
 | ||||
|     if use_ctlc: | ||||
|         # XXX: disable pygments highlighting for auto-tests | ||||
|         # since some envs (like actions CI) will struggle | ||||
|         # the the added color-char encoding.. | ||||
|         from tractor.devx._debug import TractorConfig | ||||
|         TractorConfig.use_pygements = False | ||||
| 
 | ||||
|     yield use_ctlc | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'user_in_out', | ||||
|     [ | ||||
|  | @ -238,14 +92,15 @@ def test_root_actor_error( | |||
|     # scan for the prompt | ||||
|     expect(child, PROMPT) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
| 
 | ||||
|     # make sure expected logging and error arrives | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         [_crash_msg, "('root'"] | ||||
|         child, | ||||
|         [ | ||||
|             _crash_msg, | ||||
|             "('root'", | ||||
|             'AssertionError', | ||||
|         ] | ||||
|     ) | ||||
|     assert 'AssertionError' in before | ||||
| 
 | ||||
|     # send user command | ||||
|     child.sendline(user_input) | ||||
|  | @ -264,8 +119,10 @@ def test_root_actor_error( | |||
|     ids=lambda item: f'{item[0]} -> {item[1]}', | ||||
| ) | ||||
| def test_root_actor_bp(spawn, user_in_out): | ||||
|     """Demonstrate breakpoint from in root actor. | ||||
|     """ | ||||
|     ''' | ||||
|     Demonstrate breakpoint from in root actor. | ||||
| 
 | ||||
|     ''' | ||||
|     user_input, expect_err_str = user_in_out | ||||
|     child = spawn('root_actor_breakpoint') | ||||
| 
 | ||||
|  | @ -279,7 +136,7 @@ def test_root_actor_bp(spawn, user_in_out): | |||
|     child.expect('\r\n') | ||||
| 
 | ||||
|     # process should exit | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
|     if expect_err_str is None: | ||||
|         assert 'Error' not in str(child.before) | ||||
|  | @ -287,38 +144,6 @@ def test_root_actor_bp(spawn, user_in_out): | |||
|         assert expect_err_str in str(child.before) | ||||
| 
 | ||||
| 
 | ||||
| def do_ctlc( | ||||
|     child, | ||||
|     count: int = 3, | ||||
|     delay: float = 0.1, | ||||
|     patt: str|None = None, | ||||
| 
 | ||||
|     # expect repl UX to reprint the prompt after every | ||||
|     # ctrl-c send. | ||||
|     # XXX: no idea but, in CI this never seems to work even on 3.10 so | ||||
|     # needs some further investigation potentially... | ||||
|     expect_prompt: bool = not _ci_env, | ||||
| 
 | ||||
| ) -> None: | ||||
| 
 | ||||
|     # make sure ctl-c sends don't do anything but repeat output | ||||
|     for _ in range(count): | ||||
|         time.sleep(delay) | ||||
|         child.sendcontrol('c') | ||||
| 
 | ||||
|         # TODO: figure out why this makes CI fail.. | ||||
|         # if you run this test manually it works just fine.. | ||||
|         if expect_prompt: | ||||
|             before = str(child.before.decode()) | ||||
|             time.sleep(delay) | ||||
|             child.expect(PROMPT) | ||||
|             time.sleep(delay) | ||||
| 
 | ||||
|             if patt: | ||||
|                 # should see the last line on console | ||||
|                 assert patt in before | ||||
| 
 | ||||
| 
 | ||||
| def test_root_actor_bp_forever( | ||||
|     spawn, | ||||
|     ctlc: bool, | ||||
|  | @ -358,7 +183,7 @@ def test_root_actor_bp_forever( | |||
| 
 | ||||
|     # quit out of the loop | ||||
|     child.sendline('q') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|  | @ -380,10 +205,12 @@ def test_subactor_error( | |||
|     # scan for the prompt | ||||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         [_crash_msg, "('name_error'"] | ||||
|         child, | ||||
|         [ | ||||
|             _crash_msg, | ||||
|             "('name_error'", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     if do_next: | ||||
|  | @ -402,17 +229,15 @@ def test_subactor_error( | |||
|         child.sendline('continue') | ||||
| 
 | ||||
|     child.expect(PROMPT) | ||||
|     before = str(child.before.decode()) | ||||
| 
 | ||||
|     # root actor gets debugger engaged | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         [_crash_msg, "('root'"] | ||||
|     ) | ||||
|     # error is a remote error propagated from the subactor | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         [_crash_msg, "('name_error'"] | ||||
|         child, | ||||
|         [ | ||||
|             _crash_msg, | ||||
|             # root actor gets debugger engaged | ||||
|             "('root'", | ||||
|             # error is a remote error propagated from the subactor | ||||
|             "('name_error'", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     # another round | ||||
|  | @ -423,7 +248,7 @@ def test_subactor_error( | |||
|     child.expect('\r\n') | ||||
| 
 | ||||
|     # process should exit | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| def test_subactor_breakpoint( | ||||
|  | @ -433,14 +258,11 @@ def test_subactor_breakpoint( | |||
|     "Single subactor with an infinite breakpoint loop" | ||||
| 
 | ||||
|     child = spawn('subactor_breakpoint') | ||||
| 
 | ||||
|     # scan for the prompt | ||||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         [_pause_msg, "('breakpoint_forever'"] | ||||
|         child, | ||||
|         [_pause_msg, | ||||
|          "('breakpoint_forever'",] | ||||
|     ) | ||||
| 
 | ||||
|     # do some "next" commands to demonstrate recurrent breakpoint | ||||
|  | @ -456,9 +278,8 @@ def test_subactor_breakpoint( | |||
|     for _ in range(5): | ||||
|         child.sendline('continue') | ||||
|         child.expect(PROMPT) | ||||
|         before = str(child.before.decode()) | ||||
|         assert in_prompt_msg( | ||||
|             before, | ||||
|             child, | ||||
|             [_pause_msg, "('breakpoint_forever'"] | ||||
|         ) | ||||
| 
 | ||||
|  | @ -471,9 +292,8 @@ def test_subactor_breakpoint( | |||
|     # child process should exit but parent will capture pdb.BdbQuit | ||||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         child, | ||||
|         ['RemoteActorError:', | ||||
|          "('breakpoint_forever'", | ||||
|          'bdb.BdbQuit',] | ||||
|  | @ -486,14 +306,16 @@ def test_subactor_breakpoint( | |||
|     child.sendline('c') | ||||
| 
 | ||||
|     # process should exit | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         ['RemoteActorError:', | ||||
|         child, [ | ||||
|         'MessagingError:', | ||||
|         'RemoteActorError:', | ||||
|          "('breakpoint_forever'", | ||||
|          'bdb.BdbQuit',] | ||||
|          'bdb.BdbQuit', | ||||
|         ], | ||||
|         pause_on_false=True, | ||||
|     ) | ||||
| 
 | ||||
| 
 | ||||
|  | @ -514,7 +336,7 @@ def test_multi_subactors( | |||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         child, | ||||
|         [_pause_msg, "('breakpoint_forever'"] | ||||
|     ) | ||||
| 
 | ||||
|  | @ -535,12 +357,14 @@ def test_multi_subactors( | |||
| 
 | ||||
|     # first name_error failure | ||||
|     child.expect(PROMPT) | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         [_crash_msg, "('name_error'"] | ||||
|         child, | ||||
|         [ | ||||
|             _crash_msg, | ||||
|             "('name_error'", | ||||
|             "NameError", | ||||
|         ] | ||||
|     ) | ||||
|     assert "NameError" in before | ||||
| 
 | ||||
|     if ctlc: | ||||
|         do_ctlc(child) | ||||
|  | @ -564,9 +388,8 @@ def test_multi_subactors( | |||
|     # breakpoint loop should re-engage | ||||
|     child.sendline('c') | ||||
|     child.expect(PROMPT) | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         child, | ||||
|         [_pause_msg, "('breakpoint_forever'"] | ||||
|     ) | ||||
| 
 | ||||
|  | @ -629,7 +452,7 @@ def test_multi_subactors( | |||
| 
 | ||||
|     # process should exit | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
|     # repeat of previous multierror for final output | ||||
|     assert_before(child, [ | ||||
|  | @ -659,25 +482,28 @@ def test_multi_daemon_subactors( | |||
|     # the root's tty lock first so anticipate either crash | ||||
|     # message on the first entry. | ||||
| 
 | ||||
|     bp_forev_parts = [_pause_msg, "('bp_forever'"] | ||||
|     bp_forev_parts = [ | ||||
|         _pause_msg, | ||||
|         "('bp_forever'", | ||||
|     ] | ||||
|     bp_forev_in_msg = partial( | ||||
|         in_prompt_msg, | ||||
|         parts=bp_forev_parts, | ||||
|     ) | ||||
| 
 | ||||
|     name_error_msg = "NameError: name 'doggypants' is not defined" | ||||
|     name_error_parts = [name_error_msg] | ||||
|     name_error_msg: str = "NameError: name 'doggypants' is not defined" | ||||
|     name_error_parts: list[str] = [name_error_msg] | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
| 
 | ||||
|     if bp_forev_in_msg(prompt=before): | ||||
|     if bp_forev_in_msg(child=child): | ||||
|         next_parts = name_error_parts | ||||
| 
 | ||||
|     elif name_error_msg in before: | ||||
|         next_parts = bp_forev_parts | ||||
| 
 | ||||
|     else: | ||||
|         raise ValueError("Neither log msg was found !?") | ||||
|         raise ValueError('Neither log msg was found !?') | ||||
| 
 | ||||
|     if ctlc: | ||||
|         do_ctlc(child) | ||||
|  | @ -746,14 +572,12 @@ def test_multi_daemon_subactors( | |||
|     # wait for final error in root | ||||
|     # where it crashs with boxed error | ||||
|     while True: | ||||
|         try: | ||||
|             child.sendline('c') | ||||
|             child.expect(PROMPT) | ||||
|             assert_before( | ||||
|                 child, | ||||
|                 bp_forev_parts | ||||
|             ) | ||||
|         except AssertionError: | ||||
|         child.sendline('c') | ||||
|         child.expect(PROMPT) | ||||
|         if not in_prompt_msg( | ||||
|             child, | ||||
|             bp_forev_parts | ||||
|         ): | ||||
|             break | ||||
| 
 | ||||
|     assert_before( | ||||
|  | @ -769,7 +593,7 @@ def test_multi_daemon_subactors( | |||
|     ) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| @has_nested_actors | ||||
|  | @ -845,7 +669,7 @@ def test_multi_subactors_root_errors( | |||
|     ]) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
|     assert_before(child, [ | ||||
|         # "Attaching to pdb in crashed actor: ('root'", | ||||
|  | @ -934,10 +758,13 @@ def test_root_nursery_cancels_before_child_releases_tty_lock( | |||
|     child = spawn('root_cancelled_but_child_is_in_tty_lock') | ||||
| 
 | ||||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     assert "NameError: name 'doggypants' is not defined" in before | ||||
|     assert "tractor._exceptions.RemoteActorError: ('name_error'" not in before | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             "NameError: name 'doggypants' is not defined", | ||||
|             "tractor._exceptions.RemoteActorError: ('name_error'", | ||||
|         ], | ||||
|     ) | ||||
|     time.sleep(0.5) | ||||
| 
 | ||||
|     if ctlc: | ||||
|  | @ -975,7 +802,7 @@ def test_root_nursery_cancels_before_child_releases_tty_lock( | |||
| 
 | ||||
|     for i in range(3): | ||||
|         try: | ||||
|             child.expect(pexpect.EOF, timeout=0.5) | ||||
|             child.expect(EOF, timeout=0.5) | ||||
|             break | ||||
|         except TIMEOUT: | ||||
|             child.sendline('c') | ||||
|  | @ -1017,7 +844,7 @@ def test_root_cancels_child_context_during_startup( | |||
|         do_ctlc(child) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| def test_different_debug_mode_per_actor( | ||||
|  | @ -1028,9 +855,8 @@ def test_different_debug_mode_per_actor( | |||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     # only one actor should enter the debugger | ||||
|     before = str(child.before.decode()) | ||||
|     assert in_prompt_msg( | ||||
|         before, | ||||
|         child, | ||||
|         [_crash_msg, "('debugged_boi'", "RuntimeError"], | ||||
|     ) | ||||
| 
 | ||||
|  | @ -1038,9 +864,7 @@ def test_different_debug_mode_per_actor( | |||
|         do_ctlc(child) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
| 
 | ||||
|     before = str(child.before.decode()) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
|     # NOTE: this debugged actor error currently WON'T show up since the | ||||
|     # root will actually cancel and terminate the nursery before the error | ||||
|  | @ -1059,103 +883,6 @@ def test_different_debug_mode_per_actor( | |||
|     ) | ||||
| 
 | ||||
| 
 | ||||
| def test_pause_from_sync( | ||||
|     spawn, | ||||
|     ctlc: bool | ||||
| ): | ||||
|     ''' | ||||
|     Verify we can use the `pdbp` REPL from sync functions AND from | ||||
|     any thread spawned with `trio.to_thread.run_sync()`. | ||||
| 
 | ||||
|     `examples/debugging/sync_bp.py` | ||||
| 
 | ||||
|     ''' | ||||
|     child = spawn('sync_bp') | ||||
| 
 | ||||
|     # first `sync_pause()` after nurseries open | ||||
|     child.expect(PROMPT) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             # pre-prompt line | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|         ] | ||||
|     ) | ||||
|     if ctlc: | ||||
|         do_ctlc(child) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
| 
 | ||||
| 
 | ||||
|     # first `await tractor.pause()` inside `p.open_context()` body | ||||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     # XXX shouldn't see gb loaded message with PDB loglevel! | ||||
|     before = str(child.before.decode()) | ||||
|     assert not in_prompt_msg( | ||||
|         before, | ||||
|         ['`greenback` portal opened!'], | ||||
|     ) | ||||
|     # should be same root task | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     if ctlc: | ||||
|         do_ctlc(child) | ||||
| 
 | ||||
|     # one of the bg thread or subactor should have | ||||
|     # `Lock.acquire()`-ed | ||||
|     # (NOT both, which will result in REPL clobbering!) | ||||
|     attach_patts: dict[str, list[str]] = { | ||||
|         'subactor': [ | ||||
|             "'start_n_sync_pause'", | ||||
|             "('subactor'", | ||||
|         ], | ||||
|         'inline_root_bg_thread': [ | ||||
|             "<Thread(inline_root_bg_thread", | ||||
|             "('root'", | ||||
|         ], | ||||
|         'start_soon_root_bg_thread': [ | ||||
|             "<Thread(start_soon_root_bg_thread", | ||||
|             "('root'", | ||||
|         ], | ||||
|     } | ||||
|     while attach_patts: | ||||
|         child.sendline('c') | ||||
|         child.expect(PROMPT) | ||||
|         before = str(child.before.decode()) | ||||
|         for key in attach_patts.copy(): | ||||
|             if key in before: | ||||
|                 expected_patts: str = attach_patts.pop(key) | ||||
|                 assert_before( | ||||
|                     child, | ||||
|                     [_pause_msg] + expected_patts | ||||
|                 ) | ||||
|                 break | ||||
| 
 | ||||
|         # ensure no other task/threads engaged a REPL | ||||
|         # at the same time as the one that was detected above. | ||||
|         for key, other_patts in attach_patts.items(): | ||||
|             assert not in_prompt_msg( | ||||
|                 before, | ||||
|                 other_patts, | ||||
|             ) | ||||
| 
 | ||||
|         if ctlc: | ||||
|             do_ctlc(child) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
| 
 | ||||
| 
 | ||||
| def test_post_mortem_api( | ||||
|     spawn, | ||||
|     ctlc: bool, | ||||
|  | @ -1258,7 +985,7 @@ def test_post_mortem_api( | |||
|     # ) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| def test_shield_pause( | ||||
|  | @ -1333,9 +1060,26 @@ def test_shield_pause( | |||
|         ] | ||||
|     ) | ||||
|     child.sendline('c') | ||||
|     child.expect(pexpect.EOF) | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| # TODO: better error for "non-ideal" usage from the root actor. | ||||
| # -[ ] if called from an async scope emit a message that suggests | ||||
| #    using `await tractor.pause()` instead since it's less overhead | ||||
| #    (in terms of `greenback` and/or extra threads) and if it's from | ||||
| #    a sync scope suggest that usage must first call | ||||
| #    `ensure_portal()` in the (eventual parent) async calling scope? | ||||
| def test_sync_pause_from_bg_task_in_root_actor_(): | ||||
|     ''' | ||||
|     When used from the root actor, normally we can only implicitly | ||||
|     support `.pause_from_sync()` from the main-parent-task (that | ||||
|     opens the runtime via `open_root_actor()`) since `greenback` | ||||
|     requires a `.ensure_portal()` call per `trio.Task` where it is | ||||
|     used. | ||||
| 
 | ||||
|     ''' | ||||
|     ... | ||||
| 
 | ||||
| # TODO: needs ANSI code stripping tho, see `assert_before()` # above! | ||||
| def test_correct_frames_below_hidden(): | ||||
|     ''' | ||||
|  | @ -0,0 +1,381 @@ | |||
| ''' | ||||
| That "foreign loop/thread" debug REPL support better ALSO WORK! | ||||
| 
 | ||||
| Same as `test_native_pause.py`. | ||||
| All these tests can be understood (somewhat) by running the | ||||
| equivalent `examples/debugging/` scripts manually. | ||||
| 
 | ||||
| ''' | ||||
| from contextlib import ( | ||||
|     contextmanager as cm, | ||||
| ) | ||||
| # from functools import partial | ||||
| # import itertools | ||||
| import time | ||||
| # from typing import ( | ||||
| #     Iterator, | ||||
| # ) | ||||
| 
 | ||||
| import pytest | ||||
| from pexpect.exceptions import ( | ||||
|     TIMEOUT, | ||||
|     EOF, | ||||
| ) | ||||
| 
 | ||||
| from .conftest import ( | ||||
|     # _ci_env, | ||||
|     do_ctlc, | ||||
|     PROMPT, | ||||
|     # expect, | ||||
|     in_prompt_msg, | ||||
|     assert_before, | ||||
|     _pause_msg, | ||||
|     _crash_msg, | ||||
|     _ctlc_ignore_header, | ||||
|     # _repl_fail_msg, | ||||
| ) | ||||
| 
 | ||||
| @cm | ||||
| def maybe_expect_timeout( | ||||
|     ctlc: bool = False, | ||||
| ) -> None: | ||||
|     try: | ||||
|         yield | ||||
|     except TIMEOUT: | ||||
|         # breakpoint() | ||||
|         if ctlc: | ||||
|             pytest.xfail( | ||||
|                 'Some kinda redic threading SIGINT bug i think?\n' | ||||
|                 'See the notes in `examples/debugging/sync_bp.py`..\n' | ||||
|             ) | ||||
|         raise | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.ctlcs_bish | ||||
| def test_pause_from_sync( | ||||
|     spawn, | ||||
|     ctlc: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Verify we can use the `pdbp` REPL from sync functions AND from | ||||
|     any thread spawned with `trio.to_thread.run_sync()`. | ||||
| 
 | ||||
|     `examples/debugging/sync_bp.py` | ||||
| 
 | ||||
|     ''' | ||||
|     child = spawn('sync_bp') | ||||
| 
 | ||||
|     # first `sync_pause()` after nurseries open | ||||
|     child.expect(PROMPT) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             # pre-prompt line | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|         ] | ||||
|     ) | ||||
|     if ctlc: | ||||
|         do_ctlc(child) | ||||
|         # ^NOTE^ subactor not spawned yet; don't need extra delay. | ||||
| 
 | ||||
|     child.sendline('c') | ||||
| 
 | ||||
|     # first `await tractor.pause()` inside `p.open_context()` body | ||||
|     child.expect(PROMPT) | ||||
| 
 | ||||
|     # XXX shouldn't see gb loaded message with PDB loglevel! | ||||
|     # assert not in_prompt_msg( | ||||
|     #     child, | ||||
|     #     ['`greenback` portal opened!'], | ||||
|     # ) | ||||
|     # should be same root task | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     if ctlc: | ||||
|         do_ctlc( | ||||
|             child, | ||||
|             # NOTE: setting this to 0 (or some other sufficient | ||||
|             # small val) can cause the test to fail since the | ||||
|             # `subactor` suffers a race where the root/parent | ||||
|             # sends an actor-cancel prior to it hitting its pause | ||||
|             # point; by def the value is 0.1 | ||||
|             delay=0.4, | ||||
|         ) | ||||
| 
 | ||||
|     # XXX, fwiw without a brief sleep here the SIGINT might actually | ||||
|     # trigger "subactor" cancellation by its parent  before the | ||||
|     # shield-handler is engaged. | ||||
|     # | ||||
|     # => similar to the `delay` input to `do_ctlc()` below, setting | ||||
|     # this too low can cause the test to fail since the `subactor` | ||||
|     # suffers a race where the root/parent sends an actor-cancel | ||||
|     # prior to the context task hitting its pause point (and thus | ||||
|     # engaging the `sigint_shield()` handler in time); this value | ||||
|     # seems be good enuf? | ||||
|     time.sleep(0.6) | ||||
| 
 | ||||
|     # one of the bg thread or subactor should have | ||||
|     # `Lock.acquire()`-ed | ||||
|     # (NOT both, which will result in REPL clobbering!) | ||||
|     attach_patts: dict[str, list[str]] = { | ||||
|         'subactor': [ | ||||
|             "'start_n_sync_pause'", | ||||
|             "('subactor'", | ||||
|         ], | ||||
|         'inline_root_bg_thread': [ | ||||
|             "<Thread(inline_root_bg_thread", | ||||
|             "('root'", | ||||
|         ], | ||||
|         'start_soon_root_bg_thread': [ | ||||
|             "<Thread(start_soon_root_bg_thread", | ||||
|             "('root'", | ||||
|         ], | ||||
|     } | ||||
|     conts: int = 0  # for debugging below matching logic on failure | ||||
|     while attach_patts: | ||||
|         child.sendline('c') | ||||
|         conts += 1 | ||||
|         child.expect(PROMPT) | ||||
|         before = str(child.before.decode()) | ||||
|         for key in attach_patts: | ||||
|             if key in before: | ||||
|                 attach_key: str = key | ||||
|                 expected_patts: str = attach_patts.pop(key) | ||||
|                 assert_before( | ||||
|                     child, | ||||
|                     [_pause_msg] | ||||
|                     + | ||||
|                     expected_patts | ||||
|                 ) | ||||
|                 break | ||||
|         else: | ||||
|             pytest.fail( | ||||
|                 f'No keys found?\n\n' | ||||
|                 f'{attach_patts.keys()}\n\n' | ||||
|                 f'{before}\n' | ||||
|             ) | ||||
| 
 | ||||
|         # ensure no other task/threads engaged a REPL | ||||
|         # at the same time as the one that was detected above. | ||||
|         for key, other_patts in attach_patts.copy().items(): | ||||
|             assert not in_prompt_msg( | ||||
|                 child, | ||||
|                 other_patts, | ||||
|             ) | ||||
| 
 | ||||
|         if ctlc: | ||||
|             do_ctlc( | ||||
|                 child, | ||||
|                 patt=attach_key, | ||||
|                 # NOTE same as comment above | ||||
|                 delay=0.4, | ||||
|             ) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
| 
 | ||||
|     # XXX TODO, weird threading bug it seems despite the | ||||
|     # `abandon_on_cancel: bool` setting to | ||||
|     # `trio.to_thread.run_sync()`.. | ||||
|     with maybe_expect_timeout( | ||||
|         ctlc=ctlc, | ||||
|     ): | ||||
|         child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| def expect_any_of( | ||||
|     attach_patts: dict[str, list[str]], | ||||
|     child,   # what type? | ||||
|     ctlc: bool = False, | ||||
|     prompt: str = _ctlc_ignore_header, | ||||
|     ctlc_delay: float = .4, | ||||
| 
 | ||||
| ) -> list[str]: | ||||
|     ''' | ||||
|     Receive any of a `list[str]` of patterns provided in | ||||
|     `attach_patts`. | ||||
| 
 | ||||
|     Used to test racing prompts from multiple actors and/or | ||||
|     tasks using a common root process' `pdbp` REPL. | ||||
| 
 | ||||
|     ''' | ||||
|     assert attach_patts | ||||
| 
 | ||||
|     child.expect(PROMPT) | ||||
|     before = str(child.before.decode()) | ||||
| 
 | ||||
|     for attach_key in attach_patts: | ||||
|         if attach_key in before: | ||||
|             expected_patts: str = attach_patts.pop(attach_key) | ||||
|             assert_before( | ||||
|                 child, | ||||
|                 expected_patts | ||||
|             ) | ||||
|             break  # from for | ||||
|     else: | ||||
|         pytest.fail( | ||||
|             f'No keys found?\n\n' | ||||
|             f'{attach_patts.keys()}\n\n' | ||||
|             f'{before}\n' | ||||
|         ) | ||||
| 
 | ||||
|     # ensure no other task/threads engaged a REPL | ||||
|     # at the same time as the one that was detected above. | ||||
|     for key, other_patts in attach_patts.copy().items(): | ||||
|         assert not in_prompt_msg( | ||||
|             child, | ||||
|             other_patts, | ||||
|         ) | ||||
| 
 | ||||
|     if ctlc: | ||||
|         do_ctlc( | ||||
|             child, | ||||
|             patt=prompt, | ||||
|             # NOTE same as comment above | ||||
|             delay=ctlc_delay, | ||||
|         ) | ||||
| 
 | ||||
|     return expected_patts | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.ctlcs_bish | ||||
| def test_sync_pause_from_aio_task( | ||||
|     spawn, | ||||
| 
 | ||||
|     ctlc: bool | ||||
|     # ^TODO, fix for `asyncio`!! | ||||
| ): | ||||
|     ''' | ||||
|     Verify we can use the `pdbp` REPL from an `asyncio.Task` spawned using | ||||
|     APIs in `.to_asyncio`. | ||||
| 
 | ||||
|     `examples/debugging/asycio_bp.py` | ||||
| 
 | ||||
|     ''' | ||||
|     child = spawn('asyncio_bp') | ||||
| 
 | ||||
|     # RACE on whether trio/asyncio task bps first | ||||
|     attach_patts: dict[str, list[str]] = { | ||||
| 
 | ||||
|         # first pause in guest-mode (aka "infecting") | ||||
|         # `trio.Task`. | ||||
|         'trio-side': [ | ||||
|             _pause_msg, | ||||
|             "<Task 'trio_ctx'", | ||||
|             "('aio_daemon'", | ||||
|         ], | ||||
| 
 | ||||
|         # `breakpoint()` from `asyncio.Task`. | ||||
|         'asyncio-side': [ | ||||
|             _pause_msg, | ||||
|             "<Task pending name='Task-2' coro=<greenback_shim()", | ||||
|             "('aio_daemon'", | ||||
|         ], | ||||
|     } | ||||
| 
 | ||||
|     while attach_patts: | ||||
|         expect_any_of( | ||||
|             attach_patts=attach_patts, | ||||
|             child=child, | ||||
|             ctlc=ctlc, | ||||
|         ) | ||||
|         child.sendline('c') | ||||
| 
 | ||||
|     # NOW in race order, | ||||
|     # - the asyncio-task will error | ||||
|     # - the root-actor parent task will pause | ||||
|     # | ||||
|     attach_patts: dict[str, list[str]] = { | ||||
| 
 | ||||
|         # error raised in `asyncio.Task` | ||||
|         "raise ValueError('asyncio side error!')": [ | ||||
|             _crash_msg, | ||||
|             "<Task 'trio_ctx'", | ||||
|             "@ ('aio_daemon'", | ||||
|             "ValueError: asyncio side error!", | ||||
| 
 | ||||
|             # XXX, we no longer show this frame by default! | ||||
|             # 'return await chan.receive()',  # `.to_asyncio` impl internals in tb | ||||
|         ], | ||||
| 
 | ||||
|         # parent-side propagation via actor-nursery/portal | ||||
|         # "tractor._exceptions.RemoteActorError: remote task raised a 'ValueError'": [ | ||||
|         "remote task raised a 'ValueError'": [ | ||||
|             _crash_msg, | ||||
|             "src_uid=('aio_daemon'", | ||||
|             "('aio_daemon'", | ||||
|         ], | ||||
| 
 | ||||
|         # a final pause in root-actor | ||||
|         "<Task '__main__.main'": [ | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|         ], | ||||
|     } | ||||
|     while attach_patts: | ||||
|         expect_any_of( | ||||
|             attach_patts=attach_patts, | ||||
|             child=child, | ||||
|             ctlc=ctlc, | ||||
|         ) | ||||
|         child.sendline('c') | ||||
| 
 | ||||
|     assert not attach_patts | ||||
| 
 | ||||
|     # final boxed error propagates to root | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             _crash_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|             "remote task raised a 'ValueError'", | ||||
|             "ValueError: asyncio side error!", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     if ctlc: | ||||
|         do_ctlc( | ||||
|             child, | ||||
|             # NOTE: setting this to 0 (or some other sufficient | ||||
|             # small val) can cause the test to fail since the | ||||
|             # `subactor` suffers a race where the root/parent | ||||
|             # sends an actor-cancel prior to it hitting its pause | ||||
|             # point; by def the value is 0.1 | ||||
|             delay=0.4, | ||||
|         ) | ||||
| 
 | ||||
|     child.sendline('c') | ||||
|     # with maybe_expect_timeout(): | ||||
|     child.expect(EOF) | ||||
| 
 | ||||
| 
 | ||||
| def test_sync_pause_from_non_greenbacked_aio_task(): | ||||
|     ''' | ||||
|     Where the `breakpoint()` caller task is NOT spawned by | ||||
|     `tractor.to_asyncio` and thus never activates | ||||
|     a `greenback.ensure_portal()` beforehand, presumably bc the task | ||||
|     was started by some lib/dep as in often seen in the field. | ||||
| 
 | ||||
|     Ensure sync pausing works when the pause is in, | ||||
| 
 | ||||
|     - the root actor running in infected-mode? | ||||
|       |_ since we don't need any IPC to acquire the debug lock? | ||||
|       |_ is there some way to handle this like the non-main-thread case? | ||||
| 
 | ||||
|     All other cases need to error out appropriately right? | ||||
| 
 | ||||
|     - for any subactor we can't avoid needing the repl lock.. | ||||
|       |_ is there a way to hook into `asyncio.ensure_future(obj)`? | ||||
| 
 | ||||
|     ''' | ||||
|     pass | ||||
|  | @ -0,0 +1,172 @@ | |||
| ''' | ||||
| That "native" runtime-hackin toolset better be dang useful! | ||||
| 
 | ||||
| Verify the funtion of a variety of "developer-experience" tools we | ||||
| offer from the `.devx` sub-pkg: | ||||
| 
 | ||||
| - use of the lovely `stackscope` for dumping actor `trio`-task trees | ||||
|   during operation and hangs. | ||||
| 
 | ||||
| TODO: | ||||
| - demonstration of `CallerInfo` call stack frame filtering such that | ||||
|   for logging and REPL purposes a user sees exactly the layers needed | ||||
|   when debugging a problem inside the stack vs. in their app. | ||||
| 
 | ||||
| ''' | ||||
| import os | ||||
| import signal | ||||
| import time | ||||
| 
 | ||||
| from .conftest import ( | ||||
|     expect, | ||||
|     assert_before, | ||||
|     in_prompt_msg, | ||||
|     PROMPT, | ||||
|     _pause_msg, | ||||
| ) | ||||
| from pexpect.exceptions import ( | ||||
|     # TIMEOUT, | ||||
|     EOF, | ||||
| ) | ||||
| 
 | ||||
| 
 | ||||
| def test_shield_pause( | ||||
|     spawn, | ||||
| ): | ||||
|     ''' | ||||
|     Verify the `tractor.pause()/.post_mortem()` API works inside an | ||||
|     already cancelled `trio.CancelScope` and that you can step to the | ||||
|     next checkpoint wherein the cancelled will get raised. | ||||
| 
 | ||||
|     ''' | ||||
|     child = spawn( | ||||
|         'shield_hang_in_sub' | ||||
|     ) | ||||
|     expect( | ||||
|         child, | ||||
|         'Yo my child hanging..?', | ||||
|     ) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             'Entering shield sleep..', | ||||
|             'Enabling trace-trees on `SIGUSR1` since `stackscope` is installed @', | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     script_pid: int = child.pid | ||||
|     print( | ||||
|         f'Sending SIGUSR1 to {script_pid}\n' | ||||
|         f'(kill -s SIGUSR1 {script_pid})\n' | ||||
|     ) | ||||
|     os.kill( | ||||
|         script_pid, | ||||
|         signal.SIGUSR1, | ||||
|     ) | ||||
|     time.sleep(0.2) | ||||
|     expect( | ||||
|         child, | ||||
|         # end-of-tree delimiter | ||||
|         "end-of-\('root'", | ||||
|     ) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             # 'Srying to dump `stackscope` tree..', | ||||
|             # 'Dumping `stackscope` tree for actor', | ||||
|             "('root'",  # uid line | ||||
| 
 | ||||
|             # TODO!? this used to show? | ||||
|             # -[ ] mk reproducable for @oremanj? | ||||
|             # | ||||
|             # parent block point (non-shielded) | ||||
|             # 'await trio.sleep_forever()  # in root', | ||||
|         ] | ||||
|     ) | ||||
|     expect( | ||||
|         child, | ||||
|         # end-of-tree delimiter | ||||
|         "end-of-\('hanger'", | ||||
|     ) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             # relay to the sub should be reported | ||||
|             'Relaying `SIGUSR1`[10] to sub-actor', | ||||
| 
 | ||||
|             "('hanger'",  # uid line | ||||
| 
 | ||||
|             # TODO!? SEE ABOVE | ||||
|             # hanger LOC where it's shield-halted | ||||
|             # 'await trio.sleep_forever()  # in subactor', | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     # simulate the user sending a ctl-c to the hanging program. | ||||
|     # this should result in the terminator kicking in since | ||||
|     # the sub is shield blocking and can't respond to SIGINT. | ||||
|     os.kill( | ||||
|         child.pid, | ||||
|         signal.SIGINT, | ||||
|     ) | ||||
|     expect( | ||||
|         child, | ||||
|         'Shutting down actor runtime', | ||||
|         timeout=6, | ||||
|     ) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             'raise KeyboardInterrupt', | ||||
|             # 'Shutting down actor runtime', | ||||
|             '#T-800 deployed to collect zombie B0', | ||||
|             "'--uid', \"('hanger',", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
| 
 | ||||
| def test_breakpoint_hook_restored( | ||||
|     spawn, | ||||
| ): | ||||
|     ''' | ||||
|     Ensures our actor runtime sets a custom `breakpoint()` hook | ||||
|     on open then restores the stdlib's default on close. | ||||
| 
 | ||||
|     The hook state validation is done via `assert`s inside the | ||||
|     invoked script with only `breakpoint()` (not `tractor.pause()`) | ||||
|     calls used. | ||||
| 
 | ||||
|     ''' | ||||
|     child = spawn('restore_builtin_breakpoint') | ||||
| 
 | ||||
|     child.expect(PROMPT) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|             "first bp, tractor hook set", | ||||
|         ] | ||||
|     ) | ||||
|     child.sendline('c') | ||||
|     child.expect(PROMPT) | ||||
|     assert_before( | ||||
|         child, | ||||
|         [ | ||||
|             "last bp, stdlib hook restored", | ||||
|         ] | ||||
|     ) | ||||
| 
 | ||||
|     # since the stdlib hook was already restored there should be NO | ||||
|     # `tractor` `log.pdb()` content from console! | ||||
|     assert not in_prompt_msg( | ||||
|         child, | ||||
|         [ | ||||
|             _pause_msg, | ||||
|             "<Task '__main__.main'", | ||||
|             "('root'", | ||||
|         ], | ||||
|     ) | ||||
|     child.sendline('c') | ||||
|     child.expect(EOF) | ||||
|  | @ -3,7 +3,6 @@ Sketchy network blackoutz, ugly byzantine gens, puedes eschuchar la | |||
| cancelacion?.. | ||||
| 
 | ||||
| ''' | ||||
| import itertools | ||||
| from functools import partial | ||||
| from types import ModuleType | ||||
| 
 | ||||
|  | @ -230,13 +229,10 @@ def test_ipc_channel_break_during_stream( | |||
|     # get raw instance from pytest wrapper | ||||
|     value = excinfo.value | ||||
|     if isinstance(value, ExceptionGroup): | ||||
|         value = next( | ||||
|             itertools.dropwhile( | ||||
|                 lambda exc: not isinstance(exc, expect_final_exc), | ||||
|                 value.exceptions, | ||||
|             ) | ||||
|         ) | ||||
|         assert value | ||||
|         excs = value.exceptions | ||||
|         assert len(excs) == 1 | ||||
|         final_exc = excs[0] | ||||
|         assert isinstance(final_exc, expect_final_exc) | ||||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
|  | @ -259,15 +255,16 @@ async def break_ipc_after_started( | |||
| 
 | ||||
| def test_stream_closed_right_after_ipc_break_and_zombie_lord_engages(): | ||||
|     ''' | ||||
|     Verify that is a subactor's IPC goes down just after bringing up a stream | ||||
|     the parent can trigger a SIGINT and the child will be reaped out-of-IPC by | ||||
|     the localhost process supervision machinery: aka "zombie lord". | ||||
|     Verify that is a subactor's IPC goes down just after bringing up | ||||
|     a stream the parent can trigger a SIGINT and the child will be | ||||
|     reaped out-of-IPC by the localhost process supervision machinery: | ||||
|     aka "zombie lord". | ||||
| 
 | ||||
|     ''' | ||||
|     async def main(): | ||||
|         with trio.fail_after(3): | ||||
|             async with tractor.open_nursery() as n: | ||||
|                 portal = await n.start_actor( | ||||
|             async with tractor.open_nursery() as an: | ||||
|                 portal = await an.start_actor( | ||||
|                     'ipc_breaker', | ||||
|                     enable_modules=[__name__], | ||||
|                 ) | ||||
|  |  | |||
|  | @ -307,7 +307,15 @@ async def inf_streamer( | |||
| 
 | ||||
|     async with ( | ||||
|         ctx.open_stream() as stream, | ||||
|         trio.open_nursery() as tn, | ||||
| 
 | ||||
|         # XXX TODO, INTERESTING CASE!! | ||||
|         # - if we don't collapse the eg then the embedded | ||||
|         # `trio.EndOfChannel` doesn't propagate directly to the above | ||||
|         # .open_stream() parent, resulting in it also raising instead | ||||
|         # of gracefully absorbing as normal.. so how to handle? | ||||
|         trio.open_nursery( | ||||
|             strict_exception_groups=False, | ||||
|         ) as tn, | ||||
|     ): | ||||
|         async def close_stream_on_sentinel(): | ||||
|             async for msg in stream: | ||||
|  |  | |||
|  | @ -130,7 +130,7 @@ def test_multierror( | |||
|             try: | ||||
|                 await portal2.result() | ||||
|             except tractor.RemoteActorError as err: | ||||
|                 assert err.boxed_type == AssertionError | ||||
|                 assert err.boxed_type is AssertionError | ||||
|                 print("Look Maa that first actor failed hard, hehh") | ||||
|                 raise | ||||
| 
 | ||||
|  | @ -182,7 +182,7 @@ def test_multierror_fast_nursery(reg_addr, start_method, num_subactors, delay): | |||
| 
 | ||||
|     for exc in exceptions: | ||||
|         assert isinstance(exc, tractor.RemoteActorError) | ||||
|         assert exc.boxed_type == AssertionError | ||||
|         assert exc.boxed_type is AssertionError | ||||
| 
 | ||||
| 
 | ||||
| async def do_nothing(): | ||||
|  | @ -504,7 +504,9 @@ def test_cancel_via_SIGINT_other_task( | |||
|     if is_win():  # smh | ||||
|         timeout += 1 | ||||
| 
 | ||||
|     async def spawn_and_sleep_forever(task_status=trio.TASK_STATUS_IGNORED): | ||||
|     async def spawn_and_sleep_forever( | ||||
|         task_status=trio.TASK_STATUS_IGNORED | ||||
|     ): | ||||
|         async with tractor.open_nursery() as tn: | ||||
|             for i in range(3): | ||||
|                 await tn.run_in_actor( | ||||
|  | @ -517,7 +519,9 @@ def test_cancel_via_SIGINT_other_task( | |||
|     async def main(): | ||||
|         # should never timeout since SIGINT should cancel the current program | ||||
|         with trio.fail_after(timeout): | ||||
|             async with trio.open_nursery() as n: | ||||
|             async with trio.open_nursery( | ||||
|                 strict_exception_groups=False, | ||||
|             ) as n: | ||||
|                 await n.start(spawn_and_sleep_forever) | ||||
|                 if 'mp' in spawn_backend: | ||||
|                     time.sleep(0.1) | ||||
|  | @ -610,6 +614,12 @@ def test_fast_graceful_cancel_when_spawn_task_in_soft_proc_wait_for_daemon( | |||
|                     nurse.start_soon(delayed_kbi) | ||||
| 
 | ||||
|                     await p.run(do_nuthin) | ||||
| 
 | ||||
|         # need to explicitly re-raise the lone kbi..now | ||||
|         except* KeyboardInterrupt as kbi_eg: | ||||
|             assert (len(excs := kbi_eg.exceptions) == 1) | ||||
|             raise excs[0] | ||||
| 
 | ||||
|         finally: | ||||
|             duration = time.time() - start | ||||
|             if duration > timeout: | ||||
|  |  | |||
|  | @ -1,917 +0,0 @@ | |||
| ''' | ||||
| Low-level functional audits for our | ||||
| "capability based messaging"-spec feats. | ||||
| 
 | ||||
| B~) | ||||
| 
 | ||||
| ''' | ||||
| import typing | ||||
| from typing import ( | ||||
|     Any, | ||||
|     Type, | ||||
|     Union, | ||||
| ) | ||||
| 
 | ||||
| from msgspec import ( | ||||
|     structs, | ||||
|     msgpack, | ||||
|     Struct, | ||||
|     ValidationError, | ||||
| ) | ||||
| import pytest | ||||
| 
 | ||||
| import tractor | ||||
| from tractor import ( | ||||
|     _state, | ||||
|     MsgTypeError, | ||||
|     Context, | ||||
| ) | ||||
| from tractor.msg import ( | ||||
|     _codec, | ||||
|     _ctxvar_MsgCodec, | ||||
| 
 | ||||
|     NamespacePath, | ||||
|     MsgCodec, | ||||
|     mk_codec, | ||||
|     apply_codec, | ||||
|     current_codec, | ||||
| ) | ||||
| from tractor.msg.types import ( | ||||
|     _payload_msgs, | ||||
|     log, | ||||
|     PayloadMsg, | ||||
|     Started, | ||||
|     mk_msg_spec, | ||||
| ) | ||||
| import trio | ||||
| 
 | ||||
| 
 | ||||
| def mk_custom_codec( | ||||
|     pld_spec: Union[Type]|Any, | ||||
|     add_hooks: bool, | ||||
| 
 | ||||
| ) -> MsgCodec: | ||||
|     ''' | ||||
|     Create custom `msgpack` enc/dec-hooks and set a `Decoder` | ||||
|     which only loads `pld_spec` (like `NamespacePath`) types. | ||||
| 
 | ||||
|     ''' | ||||
|     uid: tuple[str, str] = tractor.current_actor().uid | ||||
| 
 | ||||
|     # XXX NOTE XXX: despite defining `NamespacePath` as a type | ||||
|     # field on our `PayloadMsg.pld`, we still need a enc/dec_hook() pair | ||||
|     # to cast to/from that type on the wire. See the docs: | ||||
|     # https://jcristharif.com/msgspec/extending.html#mapping-to-from-native-types | ||||
| 
 | ||||
|     def enc_nsp(obj: Any) -> Any: | ||||
|         print(f'{uid} ENC HOOK') | ||||
|         match obj: | ||||
|             case NamespacePath(): | ||||
|                 print( | ||||
|                     f'{uid}: `NamespacePath`-Only ENCODE?\n' | ||||
|                     f'obj-> `{obj}`: {type(obj)}\n' | ||||
|                 ) | ||||
|                 # if type(obj) != NamespacePath: | ||||
|                 #     breakpoint() | ||||
|                 return str(obj) | ||||
| 
 | ||||
|         print( | ||||
|             f'{uid}\n' | ||||
|             'CUSTOM ENCODE\n' | ||||
|             f'obj-arg-> `{obj}`: {type(obj)}\n' | ||||
|         ) | ||||
|         logmsg: str = ( | ||||
|             f'{uid}\n' | ||||
|             'FAILED ENCODE\n' | ||||
|             f'obj-> `{obj}: {type(obj)}`\n' | ||||
|         ) | ||||
|         raise NotImplementedError(logmsg) | ||||
| 
 | ||||
|     def dec_nsp( | ||||
|         obj_type: Type, | ||||
|         obj: Any, | ||||
| 
 | ||||
|     ) -> Any: | ||||
|         print( | ||||
|             f'{uid}\n' | ||||
|             'CUSTOM DECODE\n' | ||||
|             f'type-arg-> {obj_type}\n' | ||||
|             f'obj-arg-> `{obj}`: {type(obj)}\n' | ||||
|         ) | ||||
|         nsp = None | ||||
| 
 | ||||
|         if ( | ||||
|             obj_type is NamespacePath | ||||
|             and isinstance(obj, str) | ||||
|             and ':' in obj | ||||
|         ): | ||||
|             nsp = NamespacePath(obj) | ||||
|             # TODO: we could built a generic handler using | ||||
|             # JUST matching the obj_type part? | ||||
|             # nsp = obj_type(obj) | ||||
| 
 | ||||
|         if nsp: | ||||
|             print(f'Returning NSP instance: {nsp}') | ||||
|             return nsp | ||||
| 
 | ||||
|         logmsg: str = ( | ||||
|             f'{uid}\n' | ||||
|             'FAILED DECODE\n' | ||||
|             f'type-> {obj_type}\n' | ||||
|             f'obj-arg-> `{obj}`: {type(obj)}\n\n' | ||||
|             f'current codec:\n' | ||||
|             f'{current_codec()}\n' | ||||
|         ) | ||||
|         # TODO: figure out the ignore subsys for this! | ||||
|         # -[ ] option whether to defense-relay backc the msg | ||||
|         #   inside an `Invalid`/`Ignore` | ||||
|         # -[ ] how to make this handling pluggable such that a | ||||
|         #   `Channel`/`MsgTransport` can intercept and process | ||||
|         #   back msgs either via exception handling or some other | ||||
|         #   signal? | ||||
|         log.warning(logmsg) | ||||
|         # NOTE: this delivers the invalid | ||||
|         # value up to `msgspec`'s decoding | ||||
|         # machinery for error raising. | ||||
|         return obj | ||||
|         # raise NotImplementedError(logmsg) | ||||
| 
 | ||||
|     nsp_codec: MsgCodec = mk_codec( | ||||
|         ipc_pld_spec=pld_spec, | ||||
| 
 | ||||
|         # NOTE XXX: the encode hook MUST be used no matter what since | ||||
|         # our `NamespacePath` is not any of a `Any` native type nor | ||||
|         # a `msgspec.Struct` subtype - so `msgspec` has no way to know | ||||
|         # how to encode it unless we provide the custom hook. | ||||
|         # | ||||
|         # AGAIN that is, regardless of whether we spec an | ||||
|         # `Any`-decoded-pld the enc has no knowledge (by default) | ||||
|         # how to enc `NamespacePath` (nsp), so we add a custom | ||||
|         # hook to do that ALWAYS. | ||||
|         enc_hook=enc_nsp if add_hooks else None, | ||||
| 
 | ||||
|         # XXX NOTE: pretty sure this is mutex with the `type=` to | ||||
|         # `Decoder`? so it won't work in tandem with the | ||||
|         # `ipc_pld_spec` passed above? | ||||
|         dec_hook=dec_nsp if add_hooks else None, | ||||
|     ) | ||||
|     return nsp_codec | ||||
| 
 | ||||
| 
 | ||||
| def chk_codec_applied( | ||||
|     expect_codec: MsgCodec, | ||||
|     enter_value: MsgCodec|None = None, | ||||
| 
 | ||||
| ) -> MsgCodec: | ||||
|     ''' | ||||
|     buncha sanity checks ensuring that the IPC channel's | ||||
|     context-vars are set to the expected codec and that are | ||||
|     ctx-var wrapper APIs match the same. | ||||
| 
 | ||||
|     ''' | ||||
|     # TODO: play with tricyle again, bc this is supposed to work | ||||
|     # the way we want? | ||||
|     # | ||||
|     # TreeVar | ||||
|     # task: trio.Task = trio.lowlevel.current_task() | ||||
|     # curr_codec = _ctxvar_MsgCodec.get_in(task) | ||||
| 
 | ||||
|     # ContextVar | ||||
|     # task_ctx: Context = task.context | ||||
|     # assert _ctxvar_MsgCodec in task_ctx | ||||
|     # curr_codec: MsgCodec = task.context[_ctxvar_MsgCodec] | ||||
| 
 | ||||
|     # NOTE: currently we use this! | ||||
|     # RunVar | ||||
|     curr_codec: MsgCodec = current_codec() | ||||
|     last_read_codec = _ctxvar_MsgCodec.get() | ||||
|     # assert curr_codec is last_read_codec | ||||
| 
 | ||||
|     assert ( | ||||
|         (same_codec := expect_codec) is | ||||
|         # returned from `mk_codec()` | ||||
| 
 | ||||
|         # yielded value from `apply_codec()` | ||||
| 
 | ||||
|         # read from current task's `contextvars.Context` | ||||
|         curr_codec is | ||||
|         last_read_codec | ||||
| 
 | ||||
|         # the default `msgspec` settings | ||||
|         is not _codec._def_msgspec_codec | ||||
|         is not _codec._def_tractor_codec | ||||
|     ) | ||||
| 
 | ||||
|     if enter_value: | ||||
|         enter_value is same_codec | ||||
| 
 | ||||
| 
 | ||||
| def iter_maybe_sends( | ||||
|     send_items: dict[Union[Type], Any] | list[tuple], | ||||
|     ipc_pld_spec: Union[Type] | Any, | ||||
|     add_codec_hooks: bool, | ||||
| 
 | ||||
|     codec: MsgCodec|None = None, | ||||
| 
 | ||||
| ) -> tuple[Any, bool]: | ||||
| 
 | ||||
|     if isinstance(send_items, dict): | ||||
|         send_items = send_items.items() | ||||
| 
 | ||||
|     for ( | ||||
|         send_type_spec, | ||||
|         send_value, | ||||
|     ) in send_items: | ||||
| 
 | ||||
|         expect_roundtrip: bool = False | ||||
| 
 | ||||
|         # values-to-typespec santiy | ||||
|         send_type = type(send_value) | ||||
|         assert send_type == send_type_spec or ( | ||||
|             (subtypes := getattr(send_type_spec, '__args__', None)) | ||||
|             and send_type in subtypes | ||||
|         ) | ||||
| 
 | ||||
|         spec_subtypes: set[Union[Type]] = ( | ||||
|              getattr( | ||||
|                  ipc_pld_spec, | ||||
|                  '__args__', | ||||
|                  {ipc_pld_spec,}, | ||||
|              ) | ||||
|         ) | ||||
|         send_in_spec: bool = ( | ||||
|             send_type == ipc_pld_spec | ||||
|             or ( | ||||
|                 ipc_pld_spec != Any | ||||
|                 and  # presume `Union` of types | ||||
|                 send_type in spec_subtypes | ||||
|             ) | ||||
|             or ( | ||||
|                 ipc_pld_spec == Any | ||||
|                 and | ||||
|                 send_type != NamespacePath | ||||
|             ) | ||||
|         ) | ||||
|         expect_roundtrip = ( | ||||
|             send_in_spec | ||||
|             # any spec should support all other | ||||
|             # builtin py values that we send | ||||
|             # except our custom nsp type which | ||||
|             # we should be able to send as long | ||||
|             # as we provide the custom codec hooks. | ||||
|             or ( | ||||
|                 ipc_pld_spec == Any | ||||
|                 and | ||||
|                 send_type == NamespacePath | ||||
|                 and | ||||
|                 add_codec_hooks | ||||
|             ) | ||||
|         ) | ||||
| 
 | ||||
|         if codec is not None: | ||||
|             # XXX FIRST XXX ensure roundtripping works | ||||
|             # before touching any IPC primitives/APIs. | ||||
|             wire_bytes: bytes = codec.encode( | ||||
|                 Started( | ||||
|                     cid='blahblah', | ||||
|                     pld=send_value, | ||||
|                 ) | ||||
|             ) | ||||
|             # NOTE: demonstrates the decoder loading | ||||
|             # to via our native SCIPP msg-spec | ||||
|             # (structurred-conc-inter-proc-protocol) | ||||
|             # implemented as per, | ||||
|             try: | ||||
|                 msg: Started = codec.decode(wire_bytes) | ||||
|                 if not expect_roundtrip: | ||||
|                     pytest.fail( | ||||
|                         f'NOT-EXPECTED able to roundtrip value given spec:\n' | ||||
|                         f'ipc_pld_spec -> {ipc_pld_spec}\n' | ||||
|                         f'value -> {send_value}: {send_type}\n' | ||||
|                     ) | ||||
| 
 | ||||
|                 pld = msg.pld | ||||
|                 assert pld == send_value | ||||
| 
 | ||||
|             except ValidationError: | ||||
|                 if expect_roundtrip: | ||||
|                     pytest.fail( | ||||
|                         f'EXPECTED to roundtrip value given spec:\n' | ||||
|                         f'ipc_pld_spec -> {ipc_pld_spec}\n' | ||||
|                         f'value -> {send_value}: {send_type}\n' | ||||
|                     ) | ||||
| 
 | ||||
|         yield ( | ||||
|             str(send_type), | ||||
|             send_value, | ||||
|             expect_roundtrip, | ||||
|         ) | ||||
| 
 | ||||
| 
 | ||||
| def dec_type_union( | ||||
|     type_names: list[str], | ||||
| ) -> Type: | ||||
|     ''' | ||||
|     Look up types by name, compile into a list and then create and | ||||
|     return a `typing.Union` from the full set. | ||||
| 
 | ||||
|     ''' | ||||
|     import importlib | ||||
|     types: list[Type] = [] | ||||
|     for type_name in type_names: | ||||
|         for mod in [ | ||||
|             typing, | ||||
|             importlib.import_module(__name__), | ||||
|         ]: | ||||
|             if type_ref := getattr( | ||||
|                 mod, | ||||
|                 type_name, | ||||
|                 False, | ||||
|             ): | ||||
|                 types.append(type_ref) | ||||
| 
 | ||||
|     # special case handling only.. | ||||
|     # ipc_pld_spec: Union[Type] = eval( | ||||
|     #     pld_spec_str, | ||||
|     #     {},  # globals | ||||
|     #     {'typing': typing},  # locals | ||||
|     # ) | ||||
| 
 | ||||
|     return Union[*types] | ||||
| 
 | ||||
| 
 | ||||
| def enc_type_union( | ||||
|     union_or_type: Union[Type]|Type, | ||||
| ) -> list[str]: | ||||
|     ''' | ||||
|     Encode a type-union or single type to a list of type-name-strings | ||||
|     ready for IPC interchange. | ||||
| 
 | ||||
|     ''' | ||||
|     type_strs: list[str] = [] | ||||
|     for typ in getattr( | ||||
|         union_or_type, | ||||
|         '__args__', | ||||
|         {union_or_type,}, | ||||
|     ): | ||||
|         type_strs.append(typ.__qualname__) | ||||
| 
 | ||||
|     return type_strs | ||||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
| async def send_back_values( | ||||
|     ctx: Context, | ||||
|     expect_debug: bool, | ||||
|     pld_spec_type_strs: list[str], | ||||
|     add_hooks: bool, | ||||
|     started_msg_bytes: bytes, | ||||
|     expect_ipc_send: dict[str, tuple[Any, bool]], | ||||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Setup up a custom codec to load instances of `NamespacePath` | ||||
|     and ensure we can round trip a func ref with our parent. | ||||
| 
 | ||||
|     ''' | ||||
|     uid: tuple = tractor.current_actor().uid | ||||
| 
 | ||||
|     # debug mode sanity check (prolly superfluous but, meh) | ||||
|     assert expect_debug == _state.debug_mode() | ||||
| 
 | ||||
|     # init state in sub-actor should be default | ||||
|     chk_codec_applied( | ||||
|         expect_codec=_codec._def_tractor_codec, | ||||
|     ) | ||||
| 
 | ||||
|     # load pld spec from input str | ||||
|     ipc_pld_spec = dec_type_union( | ||||
|         pld_spec_type_strs, | ||||
|     ) | ||||
|     pld_spec_str = str(ipc_pld_spec) | ||||
| 
 | ||||
|     # same as on parent side config. | ||||
|     nsp_codec: MsgCodec = mk_custom_codec( | ||||
|         pld_spec=ipc_pld_spec, | ||||
|         add_hooks=add_hooks, | ||||
|     ) | ||||
|     with ( | ||||
|         apply_codec(nsp_codec) as codec, | ||||
|     ): | ||||
|         chk_codec_applied( | ||||
|             expect_codec=nsp_codec, | ||||
|             enter_value=codec, | ||||
|         ) | ||||
| 
 | ||||
|         print( | ||||
|             f'{uid}: attempting `Started`-bytes DECODE..\n' | ||||
|         ) | ||||
|         try: | ||||
|             msg: Started = nsp_codec.decode(started_msg_bytes) | ||||
|             expected_pld_spec_str: str = msg.pld | ||||
|             assert pld_spec_str == expected_pld_spec_str | ||||
| 
 | ||||
|         # TODO: maybe we should add our own wrapper error so as to | ||||
|         # be interchange-lib agnostic? | ||||
|         # -[ ] the error type is wtv is raised from the hook so we | ||||
|         #   could also require a type-class of errors for | ||||
|         #   indicating whether the hook-failure can be handled by | ||||
|         #   a nasty-dialog-unprot sub-sys? | ||||
|         except ValidationError: | ||||
| 
 | ||||
|             # NOTE: only in the `Any` spec case do we expect this to | ||||
|             # work since otherwise no spec covers a plain-ol' | ||||
|             # `.pld: str` | ||||
|             if pld_spec_str == 'Any': | ||||
|                 raise | ||||
|             else: | ||||
|                 print( | ||||
|                     f'{uid}: (correctly) unable to DECODE `Started`-bytes\n' | ||||
|                     f'{started_msg_bytes}\n' | ||||
|                 ) | ||||
| 
 | ||||
|         iter_send_val_items = iter(expect_ipc_send.values()) | ||||
|         sent: list[Any] = [] | ||||
|         for send_value, expect_send in iter_send_val_items: | ||||
|             try: | ||||
|                 print( | ||||
|                     f'{uid}: attempting to `.started({send_value})`\n' | ||||
|                     f'=> expect_send: {expect_send}\n' | ||||
|                     f'SINCE, ipc_pld_spec: {ipc_pld_spec}\n' | ||||
|                     f'AND, codec: {codec}\n' | ||||
|                 ) | ||||
|                 await ctx.started(send_value) | ||||
|                 sent.append(send_value) | ||||
|                 if not expect_send: | ||||
| 
 | ||||
|                     # XXX NOTE XXX THIS WON'T WORK WITHOUT SPECIAL | ||||
|                     # `str` handling! or special debug mode IPC | ||||
|                     # msgs! | ||||
|                     await tractor.pause() | ||||
| 
 | ||||
|                     raise RuntimeError( | ||||
|                         f'NOT-EXPECTED able to roundtrip value given spec:\n' | ||||
|                         f'ipc_pld_spec -> {ipc_pld_spec}\n' | ||||
|                         f'value -> {send_value}: {type(send_value)}\n' | ||||
|                     ) | ||||
| 
 | ||||
|                 break  # move on to streaming block.. | ||||
| 
 | ||||
|             except tractor.MsgTypeError: | ||||
|                 await tractor.pause() | ||||
| 
 | ||||
|                 if expect_send: | ||||
|                     raise RuntimeError( | ||||
|                         f'EXPECTED to `.started()` value given spec:\n' | ||||
|                         f'ipc_pld_spec -> {ipc_pld_spec}\n' | ||||
|                         f'value -> {send_value}: {type(send_value)}\n' | ||||
|                     ) | ||||
| 
 | ||||
|         async with ctx.open_stream() as ipc: | ||||
|             print( | ||||
|                 f'{uid}: Entering streaming block to send remaining values..' | ||||
|             ) | ||||
| 
 | ||||
|             for send_value, expect_send in iter_send_val_items: | ||||
|                 send_type: Type = type(send_value) | ||||
|                 print( | ||||
|                     '------ - ------\n' | ||||
|                     f'{uid}: SENDING NEXT VALUE\n' | ||||
|                     f'ipc_pld_spec: {ipc_pld_spec}\n' | ||||
|                     f'expect_send: {expect_send}\n' | ||||
|                     f'val: {send_value}\n' | ||||
|                     '------ - ------\n' | ||||
|                 ) | ||||
|                 try: | ||||
|                     await ipc.send(send_value) | ||||
|                     print(f'***\n{uid}-CHILD sent {send_value!r}\n***\n') | ||||
|                     sent.append(send_value) | ||||
| 
 | ||||
|                     # NOTE: should only raise above on | ||||
|                     # `.started()` or a `Return` | ||||
|                     # if not expect_send: | ||||
|                     #     raise RuntimeError( | ||||
|                     #         f'NOT-EXPECTED able to roundtrip value given spec:\n' | ||||
|                     #         f'ipc_pld_spec -> {ipc_pld_spec}\n' | ||||
|                     #         f'value -> {send_value}: {send_type}\n' | ||||
|                     #     ) | ||||
| 
 | ||||
|                 except ValidationError: | ||||
|                     print(f'{uid} FAILED TO SEND {send_value}!') | ||||
| 
 | ||||
|                     # await tractor.pause() | ||||
|                     if expect_send: | ||||
|                         raise RuntimeError( | ||||
|                             f'EXPECTED to roundtrip value given spec:\n' | ||||
|                             f'ipc_pld_spec -> {ipc_pld_spec}\n' | ||||
|                             f'value -> {send_value}: {send_type}\n' | ||||
|                         ) | ||||
|                     # continue | ||||
| 
 | ||||
|             else: | ||||
|                 print( | ||||
|                     f'{uid}: finished sending all values\n' | ||||
|                     'Should be exiting stream block!\n' | ||||
|                 ) | ||||
| 
 | ||||
|         print(f'{uid}: exited streaming block!') | ||||
| 
 | ||||
|         # TODO: this won't be true bc in streaming phase we DO NOT | ||||
|         # msgspec check outbound msgs! | ||||
|         # -[ ] once we implement the receiver side `InvalidMsg` | ||||
|         #   then we can expect it here? | ||||
|         # assert ( | ||||
|         #     len(sent) | ||||
|         #     == | ||||
|         #     len([val | ||||
|         #          for val, expect in | ||||
|         #          expect_ipc_send.values() | ||||
|         #          if expect is True]) | ||||
|         # ) | ||||
| 
 | ||||
| 
 | ||||
| def ex_func(*args): | ||||
|     print(f'ex_func({args})') | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'ipc_pld_spec', | ||||
|     [ | ||||
|         Any, | ||||
|         NamespacePath, | ||||
|         NamespacePath|None,  # the "maybe" spec Bo | ||||
|     ], | ||||
|     ids=[ | ||||
|         'any_type', | ||||
|         'nsp_type', | ||||
|         'maybe_nsp_type', | ||||
|     ] | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'add_codec_hooks', | ||||
|     [ | ||||
|         True, | ||||
|         False, | ||||
|     ], | ||||
|     ids=['use_codec_hooks', 'no_codec_hooks'], | ||||
| ) | ||||
| def test_codec_hooks_mod( | ||||
|     debug_mode: bool, | ||||
|     ipc_pld_spec: Union[Type]|Any, | ||||
|     # send_value: None|str|NamespacePath, | ||||
|     add_codec_hooks: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Audit the `.msg.MsgCodec` override apis details given our impl | ||||
|     uses `contextvars` to accomplish per `trio` task codec | ||||
|     application around an inter-proc-task-comms context. | ||||
| 
 | ||||
|     ''' | ||||
|     async def main(): | ||||
|         nsp = NamespacePath.from_ref(ex_func) | ||||
|         send_items: dict[Union, Any] = { | ||||
|             Union[None]: None, | ||||
|             Union[NamespacePath]: nsp, | ||||
|             Union[str]: str(nsp), | ||||
|         } | ||||
| 
 | ||||
|         # init default state for actor | ||||
|         chk_codec_applied( | ||||
|             expect_codec=_codec._def_tractor_codec, | ||||
|         ) | ||||
| 
 | ||||
|         async with tractor.open_nursery( | ||||
|             debug_mode=debug_mode, | ||||
|         ) as an: | ||||
|             p: tractor.Portal = await an.start_actor( | ||||
|                 'sub', | ||||
|                 enable_modules=[__name__], | ||||
|             ) | ||||
| 
 | ||||
|             # TODO: 2 cases: | ||||
|             # - codec not modified -> decode nsp as `str` | ||||
|             # - codec modified with hooks -> decode nsp as | ||||
|             #   `NamespacePath` | ||||
|             nsp_codec: MsgCodec = mk_custom_codec( | ||||
|                 pld_spec=ipc_pld_spec, | ||||
|                 add_hooks=add_codec_hooks, | ||||
|             ) | ||||
|             with apply_codec(nsp_codec) as codec: | ||||
|                 chk_codec_applied( | ||||
|                     expect_codec=nsp_codec, | ||||
|                     enter_value=codec, | ||||
|                 ) | ||||
| 
 | ||||
|                 expect_ipc_send: dict[str, tuple[Any, bool]] = {} | ||||
| 
 | ||||
|                 report: str = ( | ||||
|                     'Parent report on send values with\n' | ||||
|                     f'ipc_pld_spec: {ipc_pld_spec}\n' | ||||
|                     '       ------ - ------\n' | ||||
|                 ) | ||||
|                 for val_type_str, val, expect_send in iter_maybe_sends( | ||||
|                     send_items, | ||||
|                     ipc_pld_spec, | ||||
|                     add_codec_hooks=add_codec_hooks, | ||||
|                 ): | ||||
|                     report += ( | ||||
|                         f'send_value: {val}: {type(val)} ' | ||||
|                         f'=> expect_send: {expect_send}\n' | ||||
|                     ) | ||||
|                     expect_ipc_send[val_type_str] = (val, expect_send) | ||||
| 
 | ||||
|                 print( | ||||
|                     report + | ||||
|                     '       ------ - ------\n' | ||||
|                 ) | ||||
|                 assert len(expect_ipc_send) == len(send_items) | ||||
|                 # now try over real IPC with a the subactor | ||||
|                 # expect_ipc_rountrip: bool = True | ||||
|                 expected_started = Started( | ||||
|                     cid='cid', | ||||
|                     pld=str(ipc_pld_spec), | ||||
|                 ) | ||||
|                 # build list of values we expect to receive from | ||||
|                 # the subactor. | ||||
|                 expect_to_send: list[Any] = [ | ||||
|                     val | ||||
|                     for val, expect_send in expect_ipc_send.values() | ||||
|                     if expect_send | ||||
|                 ] | ||||
| 
 | ||||
|                 pld_spec_type_strs: list[str] = enc_type_union(ipc_pld_spec) | ||||
| 
 | ||||
|                 # XXX should raise an mte (`MsgTypeError`) | ||||
|                 # when `add_codec_hooks == False` bc the input | ||||
|                 # `expect_ipc_send` kwarg has a nsp which can't be | ||||
|                 # serialized! | ||||
|                 # | ||||
|                 # TODO:can we ensure this happens from the | ||||
|                 # `Return`-side (aka the sub) as well? | ||||
|                 if not add_codec_hooks: | ||||
|                     try: | ||||
|                         async with p.open_context( | ||||
|                             send_back_values, | ||||
|                             expect_debug=debug_mode, | ||||
|                             pld_spec_type_strs=pld_spec_type_strs, | ||||
|                             add_hooks=add_codec_hooks, | ||||
|                             started_msg_bytes=nsp_codec.encode(expected_started), | ||||
| 
 | ||||
|                             # XXX NOTE bc we send a `NamespacePath` in this kwarg | ||||
|                             expect_ipc_send=expect_ipc_send, | ||||
| 
 | ||||
|                         ) as (ctx, first): | ||||
|                             pytest.fail('ctx should fail to open without custom enc_hook!?') | ||||
| 
 | ||||
|                     # this test passes bc we can go no further! | ||||
|                     except MsgTypeError: | ||||
|                         # teardown nursery | ||||
|                         await p.cancel_actor() | ||||
|                         return | ||||
| 
 | ||||
|                 # TODO: send the original nsp here and | ||||
|                 # test with `limit_msg_spec()` above? | ||||
|                 # await tractor.pause() | ||||
|                 print('PARENT opening IPC ctx!\n') | ||||
|                 async with ( | ||||
| 
 | ||||
|                     # XXX should raise an mte (`MsgTypeError`) | ||||
|                     # when `add_codec_hooks == False`.. | ||||
|                     p.open_context( | ||||
|                         send_back_values, | ||||
|                         expect_debug=debug_mode, | ||||
|                         pld_spec_type_strs=pld_spec_type_strs, | ||||
|                         add_hooks=add_codec_hooks, | ||||
|                         started_msg_bytes=nsp_codec.encode(expected_started), | ||||
|                         expect_ipc_send=expect_ipc_send, | ||||
|                     ) as (ctx, first), | ||||
| 
 | ||||
|                     ctx.open_stream() as ipc, | ||||
|                 ): | ||||
|                     # ensure codec is still applied across | ||||
|                     # `tractor.Context` + its embedded nursery. | ||||
|                     chk_codec_applied( | ||||
|                         expect_codec=nsp_codec, | ||||
|                         enter_value=codec, | ||||
|                     ) | ||||
|                     print( | ||||
|                         'root: ENTERING CONTEXT BLOCK\n' | ||||
|                         f'type(first): {type(first)}\n' | ||||
|                         f'first: {first}\n' | ||||
|                     ) | ||||
|                     expect_to_send.remove(first) | ||||
| 
 | ||||
|                     # TODO: explicit values we expect depending on | ||||
|                     # codec config! | ||||
|                     # assert first == first_val | ||||
|                     # assert first == f'{__name__}:ex_func' | ||||
| 
 | ||||
|                     async for next_sent in ipc: | ||||
|                         print( | ||||
|                             'Parent: child sent next value\n' | ||||
|                             f'{next_sent}: {type(next_sent)}\n' | ||||
|                         ) | ||||
|                         if expect_to_send: | ||||
|                             expect_to_send.remove(next_sent) | ||||
|                         else: | ||||
|                             print('PARENT should terminate stream loop + block!') | ||||
| 
 | ||||
|                     # all sent values should have arrived! | ||||
|                     assert not expect_to_send | ||||
| 
 | ||||
|             await p.cancel_actor() | ||||
| 
 | ||||
|     trio.run(main) | ||||
| 
 | ||||
| 
 | ||||
| def chk_pld_type( | ||||
|     payload_spec: Type[Struct]|Any, | ||||
|     pld: Any, | ||||
| 
 | ||||
|     expect_roundtrip: bool|None = None, | ||||
| 
 | ||||
| ) -> bool: | ||||
| 
 | ||||
|     pld_val_type: Type = type(pld) | ||||
| 
 | ||||
|     # TODO: verify that the overridden subtypes | ||||
|     # DO NOT have modified type-annots from original! | ||||
|     # 'Start',  .pld: FuncSpec | ||||
|     # 'StartAck',  .pld: IpcCtxSpec | ||||
|     # 'Stop',  .pld: UNSEt | ||||
|     # 'Error',  .pld: ErrorData | ||||
| 
 | ||||
|     codec: MsgCodec = mk_codec( | ||||
|         # NOTE: this ONLY accepts `PayloadMsg.pld` fields of a specified | ||||
|         # type union. | ||||
|         ipc_pld_spec=payload_spec, | ||||
|     ) | ||||
| 
 | ||||
|     # make a one-off dec to compare with our `MsgCodec` instance | ||||
|     # which does the below `mk_msg_spec()` call internally | ||||
|     ipc_msg_spec: Union[Type[Struct]] | ||||
|     msg_types: list[PayloadMsg[payload_spec]] | ||||
|     ( | ||||
|         ipc_msg_spec, | ||||
|         msg_types, | ||||
|     ) = mk_msg_spec( | ||||
|         payload_type_union=payload_spec, | ||||
|     ) | ||||
|     _enc = msgpack.Encoder() | ||||
|     _dec = msgpack.Decoder( | ||||
|         type=ipc_msg_spec or Any,  # like `PayloadMsg[Any]` | ||||
|     ) | ||||
| 
 | ||||
|     assert ( | ||||
|         payload_spec | ||||
|         == | ||||
|         codec.pld_spec | ||||
|     ) | ||||
| 
 | ||||
|     # assert codec.dec == dec | ||||
|     # | ||||
|     # ^-XXX-^ not sure why these aren't "equal" but when cast | ||||
|     # to `str` they seem to match ?? .. kk | ||||
| 
 | ||||
|     assert ( | ||||
|         str(ipc_msg_spec) | ||||
|         == | ||||
|         str(codec.msg_spec) | ||||
|         == | ||||
|         str(_dec.type) | ||||
|         == | ||||
|         str(codec.dec.type) | ||||
|     ) | ||||
| 
 | ||||
|     # verify the boxed-type for all variable payload-type msgs. | ||||
|     if not msg_types: | ||||
|         breakpoint() | ||||
| 
 | ||||
|     roundtrip: bool|None = None | ||||
|     pld_spec_msg_names: list[str] = [ | ||||
|         td.__name__ for td in _payload_msgs | ||||
|     ] | ||||
|     for typedef in msg_types: | ||||
| 
 | ||||
|         skip_runtime_msg: bool = typedef.__name__ not in pld_spec_msg_names | ||||
|         if skip_runtime_msg: | ||||
|             continue | ||||
| 
 | ||||
|         pld_field = structs.fields(typedef)[1] | ||||
|         assert pld_field.type is payload_spec # TODO-^ does this need to work to get all subtypes to adhere? | ||||
| 
 | ||||
|         kwargs: dict[str, Any] = { | ||||
|             'cid': '666', | ||||
|             'pld': pld, | ||||
|         } | ||||
|         enc_msg: PayloadMsg = typedef(**kwargs) | ||||
| 
 | ||||
|         _wire_bytes: bytes = _enc.encode(enc_msg) | ||||
|         wire_bytes: bytes = codec.enc.encode(enc_msg) | ||||
|         assert _wire_bytes == wire_bytes | ||||
| 
 | ||||
|         ve: ValidationError|None = None | ||||
|         try: | ||||
|             dec_msg = codec.dec.decode(wire_bytes) | ||||
|             _dec_msg = _dec.decode(wire_bytes) | ||||
| 
 | ||||
|             # decoded msg and thus payload should be exactly same! | ||||
|             assert (roundtrip := ( | ||||
|                 _dec_msg | ||||
|                 == | ||||
|                 dec_msg | ||||
|                 == | ||||
|                 enc_msg | ||||
|             )) | ||||
| 
 | ||||
|             if ( | ||||
|                 expect_roundtrip is not None | ||||
|                 and expect_roundtrip != roundtrip | ||||
|             ): | ||||
|                 breakpoint() | ||||
| 
 | ||||
|             assert ( | ||||
|                 pld | ||||
|                 == | ||||
|                 dec_msg.pld | ||||
|                 == | ||||
|                 enc_msg.pld | ||||
|             ) | ||||
|             # assert (roundtrip := (_dec_msg == enc_msg)) | ||||
| 
 | ||||
|         except ValidationError as _ve: | ||||
|             ve = _ve | ||||
|             roundtrip: bool = False | ||||
|             if pld_val_type is payload_spec: | ||||
|                 raise ValueError( | ||||
|                    'Got `ValidationError` despite type-var match!?\n' | ||||
|                     f'pld_val_type: {pld_val_type}\n' | ||||
|                     f'payload_type: {payload_spec}\n' | ||||
|                 ) from ve | ||||
| 
 | ||||
|             else: | ||||
|                 # ow we good cuz the pld spec mismatched. | ||||
|                 print( | ||||
|                     'Got expected `ValidationError` since,\n' | ||||
|                     f'{pld_val_type} is not {payload_spec}\n' | ||||
|                 ) | ||||
|         else: | ||||
|             if ( | ||||
|                 payload_spec is not Any | ||||
|                 and | ||||
|                 pld_val_type is not payload_spec | ||||
|             ): | ||||
|                 raise ValueError( | ||||
|                    'DID NOT `ValidationError` despite expected type match!?\n' | ||||
|                     f'pld_val_type: {pld_val_type}\n' | ||||
|                     f'payload_type: {payload_spec}\n' | ||||
|                 ) | ||||
| 
 | ||||
|     # full code decode should always be attempted! | ||||
|     if roundtrip is None: | ||||
|         breakpoint() | ||||
| 
 | ||||
|     return roundtrip | ||||
| 
 | ||||
| 
 | ||||
| def test_limit_msgspec(): | ||||
| 
 | ||||
|     async def main(): | ||||
|         async with tractor.open_root_actor( | ||||
|             debug_mode=True | ||||
|         ): | ||||
| 
 | ||||
|             # ensure we can round-trip a boxing `PayloadMsg` | ||||
|             assert chk_pld_type( | ||||
|                 payload_spec=Any, | ||||
|                 pld=None, | ||||
|                 expect_roundtrip=True, | ||||
|             ) | ||||
| 
 | ||||
|             # verify that a mis-typed payload value won't decode | ||||
|             assert not chk_pld_type( | ||||
|                 payload_spec=int, | ||||
|                 pld='doggy', | ||||
|             ) | ||||
| 
 | ||||
|             # parametrize the boxed `.pld` type as a custom-struct | ||||
|             # and ensure that parametrization propagates | ||||
|             # to all payload-msg-spec-able subtypes! | ||||
|             class CustomPayload(Struct): | ||||
|                 name: str | ||||
|                 value: Any | ||||
| 
 | ||||
|             assert not chk_pld_type( | ||||
|                 payload_spec=CustomPayload, | ||||
|                 pld='doggy', | ||||
|             ) | ||||
| 
 | ||||
|             assert chk_pld_type( | ||||
|                 payload_spec=CustomPayload, | ||||
|                 pld=CustomPayload(name='doggy', value='urmom') | ||||
|             ) | ||||
| 
 | ||||
|             # yah, we can `.pause_from_sync()` now! | ||||
|             # breakpoint() | ||||
| 
 | ||||
|     trio.run(main) | ||||
|  | @ -95,8 +95,8 @@ async def trio_main( | |||
| 
 | ||||
|     # stash a "service nursery" as "actor local" (aka a Python global) | ||||
|     global _nursery | ||||
|     n = _nursery | ||||
|     assert n | ||||
|     tn = _nursery | ||||
|     assert tn | ||||
| 
 | ||||
|     async def consume_stream(): | ||||
|         async with wrapper_mngr() as stream: | ||||
|  | @ -104,10 +104,10 @@ async def trio_main( | |||
|                 print(msg) | ||||
| 
 | ||||
|     # run 2 tasks to ensure broadcaster chan use | ||||
|     n.start_soon(consume_stream) | ||||
|     n.start_soon(consume_stream) | ||||
|     tn.start_soon(consume_stream) | ||||
|     tn.start_soon(consume_stream) | ||||
| 
 | ||||
|     n.start_soon(trio_sleep_and_err) | ||||
|     tn.start_soon(trio_sleep_and_err) | ||||
| 
 | ||||
|     await trio.sleep_forever() | ||||
| 
 | ||||
|  | @ -117,8 +117,10 @@ async def open_actor_local_nursery( | |||
|     ctx: tractor.Context, | ||||
| ): | ||||
|     global _nursery | ||||
|     async with trio.open_nursery() as n: | ||||
|         _nursery = n | ||||
|     async with trio.open_nursery( | ||||
|         strict_exception_groups=False, | ||||
|     ) as tn: | ||||
|         _nursery = tn | ||||
|         await ctx.started() | ||||
|         await trio.sleep(10) | ||||
|         # await trio.sleep(1) | ||||
|  | @ -132,7 +134,7 @@ async def open_actor_local_nursery( | |||
|         # never yields back.. aka a scenario where the | ||||
|         # ``tractor.context`` task IS NOT in the service n's cancel | ||||
|         # scope. | ||||
|         n.cancel_scope.cancel() | ||||
|         tn.cancel_scope.cancel() | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|  | @ -157,7 +159,7 @@ def test_actor_managed_trio_nursery_task_error_cancels_aio( | |||
|         async with tractor.open_nursery() as n: | ||||
|             p = await n.start_actor( | ||||
|                 'nursery_mngr', | ||||
|                 infect_asyncio=asyncio_mode, | ||||
|                 infect_asyncio=asyncio_mode,  # TODO, is this enabling debug mode? | ||||
|                 enable_modules=[__name__], | ||||
|             ) | ||||
|             async with ( | ||||
|  |  | |||
|  | @ -38,9 +38,9 @@ from tractor._testing import ( | |||
| # - standard setup/teardown: | ||||
| #   ``Portal.open_context()`` starts a new | ||||
| #   remote task context in another actor. The target actor's task must | ||||
| #   call ``Context.started()`` to unblock this entry on the caller side. | ||||
| #   the callee task executes until complete and returns a final value | ||||
| #   which is delivered to the caller side and retreived via | ||||
| #   call ``Context.started()`` to unblock this entry on the parent side. | ||||
| #   the child task executes until complete and returns a final value | ||||
| #   which is delivered to the parent side and retreived via | ||||
| #   ``Context.result()``. | ||||
| 
 | ||||
| # - cancel termination: | ||||
|  | @ -170,9 +170,9 @@ async def assert_state(value: bool): | |||
|     [False, ValueError, KeyboardInterrupt], | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'callee_blocks_forever', | ||||
|     'child_blocks_forever', | ||||
|     [False, True], | ||||
|     ids=lambda item: f'callee_blocks_forever={item}' | ||||
|     ids=lambda item: f'child_blocks_forever={item}' | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'pointlessly_open_stream', | ||||
|  | @ -181,7 +181,7 @@ async def assert_state(value: bool): | |||
| ) | ||||
| def test_simple_context( | ||||
|     error_parent, | ||||
|     callee_blocks_forever, | ||||
|     child_blocks_forever, | ||||
|     pointlessly_open_stream, | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|  | @ -204,13 +204,13 @@ def test_simple_context( | |||
|                         portal.open_context( | ||||
|                             simple_setup_teardown, | ||||
|                             data=10, | ||||
|                             block_forever=callee_blocks_forever, | ||||
|                             block_forever=child_blocks_forever, | ||||
|                         ) as (ctx, sent), | ||||
|                     ): | ||||
|                         assert current_ipc_ctx() is ctx | ||||
|                         assert sent == 11 | ||||
| 
 | ||||
|                         if callee_blocks_forever: | ||||
|                         if child_blocks_forever: | ||||
|                             await portal.run(assert_state, value=True) | ||||
|                         else: | ||||
|                             assert await ctx.result() == 'yo' | ||||
|  | @ -220,7 +220,7 @@ def test_simple_context( | |||
|                                 if error_parent: | ||||
|                                     raise error_parent | ||||
| 
 | ||||
|                                 if callee_blocks_forever: | ||||
|                                 if child_blocks_forever: | ||||
|                                     await ctx.cancel() | ||||
|                                 else: | ||||
|                                     # in this case the stream will send a | ||||
|  | @ -259,9 +259,9 @@ def test_simple_context( | |||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'callee_returns_early', | ||||
|     'child_returns_early', | ||||
|     [True, False], | ||||
|     ids=lambda item: f'callee_returns_early={item}' | ||||
|     ids=lambda item: f'child_returns_early={item}' | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'cancel_method', | ||||
|  | @ -273,14 +273,14 @@ def test_simple_context( | |||
|     [True, False], | ||||
|     ids=lambda item: f'chk_ctx_result_before_exit={item}' | ||||
| ) | ||||
| def test_caller_cancels( | ||||
| def test_parent_cancels( | ||||
|     cancel_method: str, | ||||
|     chk_ctx_result_before_exit: bool, | ||||
|     callee_returns_early: bool, | ||||
|     child_returns_early: bool, | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Verify that when the opening side of a context (aka the caller) | ||||
|     Verify that when the opening side of a context (aka the parent) | ||||
|     cancels that context, the ctx does not raise a cancelled when | ||||
|     either calling `.result()` or on context exit. | ||||
| 
 | ||||
|  | @ -294,7 +294,7 @@ def test_caller_cancels( | |||
| 
 | ||||
|         if ( | ||||
|             cancel_method == 'portal' | ||||
|             and not callee_returns_early | ||||
|             and not child_returns_early | ||||
|         ): | ||||
|             try: | ||||
|                 res = await ctx.result() | ||||
|  | @ -318,7 +318,7 @@ def test_caller_cancels( | |||
|                 pytest.fail(f'should not have raised ctxc\n{ctxc}') | ||||
| 
 | ||||
|         # we actually get a result | ||||
|         if callee_returns_early: | ||||
|         if child_returns_early: | ||||
|             assert res == 'yo' | ||||
|             assert ctx.outcome is res | ||||
|             assert ctx.maybe_error is None | ||||
|  | @ -362,14 +362,14 @@ def test_caller_cancels( | |||
|             ) | ||||
|             timeout: float = ( | ||||
|                 0.5 | ||||
|                 if not callee_returns_early | ||||
|                 if not child_returns_early | ||||
|                 else 2 | ||||
|             ) | ||||
|             with trio.fail_after(timeout): | ||||
|                 async with ( | ||||
|                     expect_ctxc( | ||||
|                         yay=( | ||||
|                             not callee_returns_early | ||||
|                             not child_returns_early | ||||
|                             and cancel_method == 'portal' | ||||
|                         ) | ||||
|                     ), | ||||
|  | @ -377,13 +377,13 @@ def test_caller_cancels( | |||
|                     portal.open_context( | ||||
|                         simple_setup_teardown, | ||||
|                         data=10, | ||||
|                         block_forever=not callee_returns_early, | ||||
|                         block_forever=not child_returns_early, | ||||
|                     ) as (ctx, sent), | ||||
|                 ): | ||||
| 
 | ||||
|                     if callee_returns_early: | ||||
|                     if child_returns_early: | ||||
|                         # ensure we block long enough before sending | ||||
|                         # a cancel such that the callee has already | ||||
|                         # a cancel such that the child has already | ||||
|                         # returned it's result. | ||||
|                         await trio.sleep(0.5) | ||||
| 
 | ||||
|  | @ -421,7 +421,7 @@ def test_caller_cancels( | |||
|             #   which should in turn cause `ctx._scope` to | ||||
|             # catch any cancellation? | ||||
|             if ( | ||||
|                 not callee_returns_early | ||||
|                 not child_returns_early | ||||
|                 and cancel_method != 'portal' | ||||
|             ): | ||||
|                 assert not ctx._scope.cancelled_caught | ||||
|  | @ -430,11 +430,11 @@ def test_caller_cancels( | |||
| 
 | ||||
| 
 | ||||
| # basic stream terminations: | ||||
| # - callee context closes without using stream | ||||
| # - caller context closes without using stream | ||||
| # - caller context calls `Context.cancel()` while streaming | ||||
| #   is ongoing resulting in callee being cancelled | ||||
| # - callee calls `Context.cancel()` while streaming and caller | ||||
| # - child context closes without using stream | ||||
| # - parent context closes without using stream | ||||
| # - parent context calls `Context.cancel()` while streaming | ||||
| #   is ongoing resulting in child being cancelled | ||||
| # - child calls `Context.cancel()` while streaming and parent | ||||
| #   sees stream terminated in `RemoteActorError` | ||||
| 
 | ||||
| # TODO: future possible features | ||||
|  | @ -443,7 +443,6 @@ def test_caller_cancels( | |||
| 
 | ||||
| @tractor.context | ||||
| async def close_ctx_immediately( | ||||
| 
 | ||||
|     ctx: Context, | ||||
| 
 | ||||
| ) -> None: | ||||
|  | @ -454,13 +453,24 @@ async def close_ctx_immediately( | |||
|     async with ctx.open_stream(): | ||||
|         pass | ||||
| 
 | ||||
|     print('child returning!') | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'parent_send_before_receive', | ||||
|     [ | ||||
|         False, | ||||
|         True, | ||||
|     ], | ||||
|     ids=lambda item: f'child_send_before_receive={item}' | ||||
| ) | ||||
| @tractor_test | ||||
| async def test_callee_closes_ctx_after_stream_open( | ||||
| async def test_child_exits_ctx_after_stream_open( | ||||
|     debug_mode: bool, | ||||
|     parent_send_before_receive: bool, | ||||
| ): | ||||
|     ''' | ||||
|     callee context closes without using stream. | ||||
|     child context closes without using stream. | ||||
| 
 | ||||
|     This should result in a msg sequence | ||||
|     |_<root>_ | ||||
|  | @ -474,6 +484,9 @@ async def test_callee_closes_ctx_after_stream_open( | |||
|     => {'stop': True, 'cid': <str>} | ||||
| 
 | ||||
|     ''' | ||||
|     timeout: float = ( | ||||
|         0.5 if not debug_mode else 999 | ||||
|     ) | ||||
|     async with tractor.open_nursery( | ||||
|         debug_mode=debug_mode, | ||||
|     ) as an: | ||||
|  | @ -482,7 +495,7 @@ async def test_callee_closes_ctx_after_stream_open( | |||
|             enable_modules=[__name__], | ||||
|         ) | ||||
| 
 | ||||
|         with trio.fail_after(0.5): | ||||
|         with trio.fail_after(timeout): | ||||
|             async with portal.open_context( | ||||
|                 close_ctx_immediately, | ||||
| 
 | ||||
|  | @ -494,41 +507,56 @@ async def test_callee_closes_ctx_after_stream_open( | |||
| 
 | ||||
|                 with trio.fail_after(0.4): | ||||
|                     async with ctx.open_stream() as stream: | ||||
|                         if parent_send_before_receive: | ||||
|                             print('sending first msg from parent!') | ||||
|                             await stream.send('yo') | ||||
| 
 | ||||
|                         # should fall through since ``StopAsyncIteration`` | ||||
|                         # should be raised through translation of | ||||
|                         # a ``trio.EndOfChannel`` by | ||||
|                         # ``trio.abc.ReceiveChannel.__anext__()`` | ||||
|                         async for _ in stream: | ||||
|                         msg = 10 | ||||
|                         async for msg in stream: | ||||
|                             # trigger failure if we DO NOT | ||||
|                             # get an EOC! | ||||
|                             assert 0 | ||||
|                         else: | ||||
|                             # never should get anythinig new from | ||||
|                             # the underlying stream | ||||
|                             assert msg == 10 | ||||
| 
 | ||||
|                             # verify stream is now closed | ||||
|                             try: | ||||
|                                 with trio.fail_after(0.3): | ||||
|                                     print('parent trying to `.receive()` on EoC stream!') | ||||
|                                     await stream.receive() | ||||
|                                     assert 0, 'should have raised eoc!?' | ||||
|                             except trio.EndOfChannel: | ||||
|                                 print('parent got EoC as expected!') | ||||
|                                 pass | ||||
|                                 # raise | ||||
| 
 | ||||
|                 # TODO: should be just raise the closed resource err | ||||
|                 # directly here to enforce not allowing a re-open | ||||
|                 # of a stream to the context (at least until a time of | ||||
|                 # if/when we decide that's a good idea?) | ||||
|                 try: | ||||
|                     with trio.fail_after(0.5): | ||||
|                     with trio.fail_after(timeout): | ||||
|                         async with ctx.open_stream() as stream: | ||||
|                             pass | ||||
|                 except trio.ClosedResourceError: | ||||
|                     pass | ||||
| 
 | ||||
|                 # if ctx._rx_chan._state.data: | ||||
|                 #     await tractor.pause() | ||||
| 
 | ||||
|         await portal.cancel_actor() | ||||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
| async def expect_cancelled( | ||||
|     ctx: Context, | ||||
|     send_before_receive: bool = False, | ||||
| 
 | ||||
| ) -> None: | ||||
|     global _state | ||||
|  | @ -538,6 +566,10 @@ async def expect_cancelled( | |||
| 
 | ||||
|     try: | ||||
|         async with ctx.open_stream() as stream: | ||||
| 
 | ||||
|             if send_before_receive: | ||||
|                 await stream.send('yo') | ||||
| 
 | ||||
|             async for msg in stream: | ||||
|                 await stream.send(msg)  # echo server | ||||
| 
 | ||||
|  | @ -564,26 +596,49 @@ async def expect_cancelled( | |||
|         raise | ||||
| 
 | ||||
|     else: | ||||
|         assert 0, "callee wasn't cancelled !?" | ||||
|         assert 0, "child wasn't cancelled !?" | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'child_send_before_receive', | ||||
|     [ | ||||
|         False, | ||||
|         True, | ||||
|     ], | ||||
|     ids=lambda item: f'child_send_before_receive={item}' | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'rent_wait_for_msg', | ||||
|     [ | ||||
|         False, | ||||
|         True, | ||||
|     ], | ||||
|     ids=lambda item: f'rent_wait_for_msg={item}' | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'use_ctx_cancel_method', | ||||
|     [False, True], | ||||
|     [ | ||||
|         False, | ||||
|         'pre_stream', | ||||
|         'post_stream_open', | ||||
|         'post_stream_close', | ||||
|     ], | ||||
|     ids=lambda item: f'use_ctx_cancel_method={item}' | ||||
| ) | ||||
| @tractor_test | ||||
| async def test_caller_closes_ctx_after_callee_opens_stream( | ||||
|     use_ctx_cancel_method: bool, | ||||
| async def test_parent_exits_ctx_after_child_enters_stream( | ||||
|     use_ctx_cancel_method: bool|str, | ||||
|     debug_mode: bool, | ||||
|     rent_wait_for_msg: bool, | ||||
|     child_send_before_receive: bool, | ||||
| ): | ||||
|     ''' | ||||
|     caller context closes without using/opening stream | ||||
|     Parent-side of IPC context closes without sending on `MsgStream`. | ||||
| 
 | ||||
|     ''' | ||||
|     async with tractor.open_nursery( | ||||
|         debug_mode=debug_mode, | ||||
|     ) as an: | ||||
| 
 | ||||
|         root: Actor = current_actor() | ||||
|         portal = await an.start_actor( | ||||
|             'ctx_cancelled', | ||||
|  | @ -592,41 +647,52 @@ async def test_caller_closes_ctx_after_callee_opens_stream( | |||
| 
 | ||||
|         async with portal.open_context( | ||||
|             expect_cancelled, | ||||
|             send_before_receive=child_send_before_receive, | ||||
|         ) as (ctx, sent): | ||||
|             assert sent is None | ||||
| 
 | ||||
|             await portal.run(assert_state, value=True) | ||||
| 
 | ||||
|             # call `ctx.cancel()` explicitly | ||||
|             if use_ctx_cancel_method: | ||||
|             if use_ctx_cancel_method == 'pre_stream': | ||||
|                 await ctx.cancel() | ||||
| 
 | ||||
|                 # NOTE: means the local side `ctx._scope` will | ||||
|                 # have been cancelled by an ctxc ack and thus | ||||
|                 # `._scope.cancelled_caught` should be set. | ||||
|                 try: | ||||
|                 async with ( | ||||
|                     expect_ctxc( | ||||
|                         # XXX: the cause is US since we call | ||||
|                         # `Context.cancel()` just above! | ||||
|                         yay=True, | ||||
| 
 | ||||
|                         # XXX: must be propagated to __aexit__ | ||||
|                         # and should be silently absorbed there | ||||
|                         # since we called `.cancel()` just above ;) | ||||
|                         reraise=True, | ||||
|                     ) as maybe_ctxc, | ||||
|                 ): | ||||
|                     async with ctx.open_stream() as stream: | ||||
|                         async for msg in stream: | ||||
|                             pass | ||||
| 
 | ||||
|                 except tractor.ContextCancelled as ctxc: | ||||
|                     # XXX: the cause is US since we call | ||||
|                     # `Context.cancel()` just above! | ||||
|                     assert ( | ||||
|                         ctxc.canceller | ||||
|                         == | ||||
|                         current_actor().uid | ||||
|                         == | ||||
|                         root.uid | ||||
|                     ) | ||||
|                         if rent_wait_for_msg: | ||||
|                             async for msg in stream: | ||||
|                                 print(f'PARENT rx: {msg!r}\n') | ||||
|                                 break | ||||
| 
 | ||||
|                     # XXX: must be propagated to __aexit__ | ||||
|                     # and should be silently absorbed there | ||||
|                     # since we called `.cancel()` just above ;) | ||||
|                     raise | ||||
|                         if use_ctx_cancel_method == 'post_stream_open': | ||||
|                             await ctx.cancel() | ||||
| 
 | ||||
|                 else: | ||||
|                     assert 0, "Should have context cancelled?" | ||||
|                     if use_ctx_cancel_method == 'post_stream_close': | ||||
|                         await ctx.cancel() | ||||
| 
 | ||||
|                 ctxc: tractor.ContextCancelled = maybe_ctxc.value | ||||
|                 assert ( | ||||
|                     ctxc.canceller | ||||
|                     == | ||||
|                     current_actor().uid | ||||
|                     == | ||||
|                     root.uid | ||||
|                 ) | ||||
| 
 | ||||
|                 # channel should still be up | ||||
|                 assert portal.channel.connected() | ||||
|  | @ -637,13 +703,20 @@ async def test_caller_closes_ctx_after_callee_opens_stream( | |||
|                     value=False, | ||||
|                 ) | ||||
| 
 | ||||
|             # XXX CHILD-BLOCKS case, we SHOULD NOT exit from the | ||||
|             # `.open_context()` before the child has returned, | ||||
|             # errored or been cancelled! | ||||
|             else: | ||||
|                 try: | ||||
|                     with trio.fail_after(0.2): | ||||
|                         await ctx.result() | ||||
|                     with trio.fail_after( | ||||
|                         0.5  # if not debug_mode else 999 | ||||
|                     ): | ||||
|                         res = await ctx.wait_for_result() | ||||
|                         assert res is not tractor._context.Unresolved | ||||
|                         assert 0, "Callee should have blocked!?" | ||||
|                 except trio.TooSlowError: | ||||
|                     # NO-OP -> since already called above | ||||
|                     # NO-OP -> since already triggered by | ||||
|                     # `trio.fail_after()` above! | ||||
|                     await ctx.cancel() | ||||
| 
 | ||||
|         # NOTE: local scope should have absorbed the cancellation since | ||||
|  | @ -683,7 +756,7 @@ async def test_caller_closes_ctx_after_callee_opens_stream( | |||
| 
 | ||||
| 
 | ||||
| @tractor_test | ||||
| async def test_multitask_caller_cancels_from_nonroot_task( | ||||
| async def test_multitask_parent_cancels_from_nonroot_task( | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|     async with tractor.open_nursery( | ||||
|  | @ -735,7 +808,6 @@ async def test_multitask_caller_cancels_from_nonroot_task( | |||
| 
 | ||||
| @tractor.context | ||||
| async def cancel_self( | ||||
| 
 | ||||
|     ctx: Context, | ||||
| 
 | ||||
| ) -> None: | ||||
|  | @ -775,11 +847,11 @@ async def cancel_self( | |||
| 
 | ||||
| 
 | ||||
| @tractor_test | ||||
| async def test_callee_cancels_before_started( | ||||
| async def test_child_cancels_before_started( | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Callee calls `Context.cancel()` while streaming and caller | ||||
|     Callee calls `Context.cancel()` while streaming and parent | ||||
|     sees stream terminated in `ContextCancelled`. | ||||
| 
 | ||||
|     ''' | ||||
|  | @ -826,14 +898,13 @@ async def never_open_stream( | |||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
| async def keep_sending_from_callee( | ||||
| 
 | ||||
| async def keep_sending_from_child( | ||||
|     ctx:  Context, | ||||
|     msg_buffer_size: int|None = None, | ||||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Send endlessly on the calleee stream. | ||||
|     Send endlessly on the child stream. | ||||
| 
 | ||||
|     ''' | ||||
|     await ctx.started() | ||||
|  | @ -841,7 +912,7 @@ async def keep_sending_from_callee( | |||
|         msg_buffer_size=msg_buffer_size, | ||||
|     ) as stream: | ||||
|         for msg in count(): | ||||
|             print(f'callee sending {msg}') | ||||
|             print(f'child sending {msg}') | ||||
|             await stream.send(msg) | ||||
|             await trio.sleep(0.01) | ||||
| 
 | ||||
|  | @ -849,12 +920,12 @@ async def keep_sending_from_callee( | |||
| @pytest.mark.parametrize( | ||||
|     'overrun_by', | ||||
|     [ | ||||
|         ('caller', 1, never_open_stream), | ||||
|         ('callee', 0, keep_sending_from_callee), | ||||
|         ('parent', 1, never_open_stream), | ||||
|         ('child', 0, keep_sending_from_child), | ||||
|     ], | ||||
|     ids=[ | ||||
|          ('caller_1buf_never_open_stream'), | ||||
|          ('callee_0buf_keep_sending_from_callee'), | ||||
|          ('parent_1buf_never_open_stream'), | ||||
|          ('child_0buf_keep_sending_from_child'), | ||||
|     ] | ||||
| ) | ||||
| def test_one_end_stream_not_opened( | ||||
|  | @ -885,8 +956,7 @@ def test_one_end_stream_not_opened( | |||
|                 ) as (ctx, sent): | ||||
|                     assert sent is None | ||||
| 
 | ||||
|                     if 'caller' in overrunner: | ||||
| 
 | ||||
|                     if 'parent' in overrunner: | ||||
|                         async with ctx.open_stream() as stream: | ||||
| 
 | ||||
|                             # itersend +1 msg more then the buffer size | ||||
|  | @ -901,7 +971,7 @@ def test_one_end_stream_not_opened( | |||
|                                 await trio.sleep_forever() | ||||
| 
 | ||||
|                     else: | ||||
|                         # callee overruns caller case so we do nothing here | ||||
|                         # child overruns parent case so we do nothing here | ||||
|                         await trio.sleep_forever() | ||||
| 
 | ||||
|             await portal.cancel_actor() | ||||
|  | @ -909,19 +979,19 @@ def test_one_end_stream_not_opened( | |||
|     # 2 overrun cases and the no overrun case (which pushes right up to | ||||
|     # the msg limit) | ||||
|     if ( | ||||
|         overrunner == 'caller' | ||||
|         overrunner == 'parent' | ||||
|     ): | ||||
|         with pytest.raises(tractor.RemoteActorError) as excinfo: | ||||
|             trio.run(main) | ||||
| 
 | ||||
|         assert excinfo.value.boxed_type == StreamOverrun | ||||
| 
 | ||||
|     elif overrunner == 'callee': | ||||
|     elif overrunner == 'child': | ||||
|         with pytest.raises(tractor.RemoteActorError) as excinfo: | ||||
|             trio.run(main) | ||||
| 
 | ||||
|         # TODO: embedded remote errors so that we can verify the source | ||||
|         # error? the callee delivers an error which is an overrun | ||||
|         # error? the child delivers an error which is an overrun | ||||
|         # wrapped in a remote actor error. | ||||
|         assert excinfo.value.boxed_type == tractor.RemoteActorError | ||||
| 
 | ||||
|  | @ -931,8 +1001,7 @@ def test_one_end_stream_not_opened( | |||
| 
 | ||||
| @tractor.context | ||||
| async def echo_back_sequence( | ||||
| 
 | ||||
|     ctx:  Context, | ||||
|     ctx: Context, | ||||
|     seq: list[int], | ||||
|     wait_for_cancel: bool, | ||||
|     allow_overruns_side: str, | ||||
|  | @ -941,12 +1010,12 @@ async def echo_back_sequence( | |||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Send endlessly on the calleee stream using a small buffer size | ||||
|     Send endlessly on the child stream using a small buffer size | ||||
|     setting on the contex to simulate backlogging that would normally | ||||
|     cause overruns. | ||||
| 
 | ||||
|     ''' | ||||
|     # NOTE: ensure that if the caller is expecting to cancel this task | ||||
|     # NOTE: ensure that if the parent is expecting to cancel this task | ||||
|     # that we stay echoing much longer then they are so we don't | ||||
|     # return early instead of receive the cancel msg. | ||||
|     total_batches: int = ( | ||||
|  | @ -955,7 +1024,7 @@ async def echo_back_sequence( | |||
|     ) | ||||
| 
 | ||||
|     await ctx.started() | ||||
|     # await tractor.breakpoint() | ||||
|     # await tractor.pause() | ||||
|     async with ctx.open_stream( | ||||
|         msg_buffer_size=msg_buffer_size, | ||||
| 
 | ||||
|  | @ -996,18 +1065,18 @@ async def echo_back_sequence( | |||
|                 if be_slow: | ||||
|                     await trio.sleep(0.05) | ||||
| 
 | ||||
|                 print('callee waiting on next') | ||||
|                 print('child waiting on next') | ||||
| 
 | ||||
|             print(f'callee echoing back latest batch\n{batch}') | ||||
|             print(f'child echoing back latest batch\n{batch}') | ||||
|             for msg in batch: | ||||
|                 print(f'callee sending msg\n{msg}') | ||||
|                 print(f'child sending msg\n{msg}') | ||||
|                 await stream.send(msg) | ||||
| 
 | ||||
|     try: | ||||
|         return 'yo' | ||||
|     finally: | ||||
|         print( | ||||
|             'exiting callee with context:\n' | ||||
|             'exiting child with context:\n' | ||||
|             f'{pformat(ctx)}\n' | ||||
|         ) | ||||
| 
 | ||||
|  | @ -1061,7 +1130,7 @@ def test_maybe_allow_overruns_stream( | |||
|             debug_mode=debug_mode, | ||||
|         ) as an: | ||||
|             portal = await an.start_actor( | ||||
|                 'callee_sends_forever', | ||||
|                 'child_sends_forever', | ||||
|                 enable_modules=[__name__], | ||||
|                 loglevel=loglevel, | ||||
|                 debug_mode=debug_mode, | ||||
|  |  | |||
|  | @ -181,7 +181,9 @@ async def spawn_and_check_registry( | |||
| 
 | ||||
|             try: | ||||
|                 async with tractor.open_nursery() as n: | ||||
|                     async with trio.open_nursery() as trion: | ||||
|                     async with trio.open_nursery( | ||||
|                         strict_exception_groups=False, | ||||
|                     ) as trion: | ||||
| 
 | ||||
|                         portals = {} | ||||
|                         for i in range(3): | ||||
|  | @ -316,7 +318,9 @@ async def close_chans_before_nursery( | |||
|                         async with portal2.open_stream_from( | ||||
|                             stream_forever | ||||
|                         ) as agen2: | ||||
|                             async with trio.open_nursery() as n: | ||||
|                             async with trio.open_nursery( | ||||
|                                 strict_exception_groups=False, | ||||
|                             ) as n: | ||||
|                                 n.start_soon(streamer, agen1) | ||||
|                                 n.start_soon(cancel, use_signal, .5) | ||||
|                                 try: | ||||
|  |  | |||
|  | @ -19,7 +19,7 @@ from tractor._testing import ( | |||
| @pytest.fixture | ||||
| def run_example_in_subproc( | ||||
|     loglevel: str, | ||||
|     testdir: pytest.Testdir, | ||||
|     testdir: pytest.Pytester, | ||||
|     reg_addr: tuple[str, int], | ||||
| ): | ||||
| 
 | ||||
|  | @ -81,28 +81,36 @@ def run_example_in_subproc( | |||
| 
 | ||||
|     # walk yields: (dirpath, dirnames, filenames) | ||||
|     [ | ||||
|         (p[0], f) for p in os.walk(examples_dir()) for f in p[2] | ||||
|         (p[0], f) | ||||
|         for p in os.walk(examples_dir()) | ||||
|         for f in p[2] | ||||
| 
 | ||||
|         if '__' not in f | ||||
|         and f[0] != '_' | ||||
|         and 'debugging' not in p[0] | ||||
|         and 'integration' not in p[0] | ||||
|         and 'advanced_faults' not in p[0] | ||||
|         and 'multihost' not in p[0] | ||||
|         if ( | ||||
|             '__' not in f | ||||
|             and f[0] != '_' | ||||
|             and 'debugging' not in p[0] | ||||
|             and 'integration' not in p[0] | ||||
|             and 'advanced_faults' not in p[0] | ||||
|             and 'multihost' not in p[0] | ||||
|         ) | ||||
|     ], | ||||
| 
 | ||||
|     ids=lambda t: t[1], | ||||
| ) | ||||
| def test_example(run_example_in_subproc, example_script): | ||||
|     """Load and run scripts from this repo's ``examples/`` dir as a user | ||||
| def test_example( | ||||
|     run_example_in_subproc, | ||||
|     example_script, | ||||
| ): | ||||
|     ''' | ||||
|     Load and run scripts from this repo's ``examples/`` dir as a user | ||||
|     would copy and pasing them into their editor. | ||||
| 
 | ||||
|     On windows a little more "finessing" is done to make | ||||
|     ``multiprocessing`` play nice: we copy the ``__main__.py`` into the | ||||
|     test directory and invoke the script as a module with ``python -m | ||||
|     test_example``. | ||||
|     """ | ||||
|     ex_file = os.path.join(*example_script) | ||||
| 
 | ||||
|     ''' | ||||
|     ex_file: str = os.path.join(*example_script) | ||||
| 
 | ||||
|     if 'rpc_bidir_streaming' in ex_file and sys.version_info < (3, 9): | ||||
|         pytest.skip("2-way streaming example requires py3.9 async with syntax") | ||||
|  | @ -128,7 +136,8 @@ def test_example(run_example_in_subproc, example_script): | |||
|                     # shouldn't eventually once we figure out what's | ||||
|                     # a better way to be explicit about aio side | ||||
|                     # cancels? | ||||
|                     and 'asyncio.exceptions.CancelledError' not in last_error | ||||
|                     and | ||||
|                     'asyncio.exceptions.CancelledError' not in last_error | ||||
|                 ): | ||||
|                     raise Exception(errmsg) | ||||
| 
 | ||||
|  |  | |||
|  | @ -0,0 +1,946 @@ | |||
| ''' | ||||
| Low-level functional audits for our | ||||
| "capability based messaging"-spec feats. | ||||
| 
 | ||||
| B~) | ||||
| 
 | ||||
| ''' | ||||
| from contextlib import ( | ||||
|     contextmanager as cm, | ||||
|     # nullcontext, | ||||
| ) | ||||
| import importlib | ||||
| from typing import ( | ||||
|     Any, | ||||
|     Type, | ||||
|     Union, | ||||
| ) | ||||
| 
 | ||||
| from msgspec import ( | ||||
|     # structs, | ||||
|     # msgpack, | ||||
|     Raw, | ||||
|     # Struct, | ||||
|     ValidationError, | ||||
| ) | ||||
| import pytest | ||||
| import trio | ||||
| 
 | ||||
| import tractor | ||||
| from tractor import ( | ||||
|     Actor, | ||||
|     # _state, | ||||
|     MsgTypeError, | ||||
|     Context, | ||||
| ) | ||||
| from tractor.msg import ( | ||||
|     _codec, | ||||
|     _ctxvar_MsgCodec, | ||||
|     _exts, | ||||
| 
 | ||||
|     NamespacePath, | ||||
|     MsgCodec, | ||||
|     MsgDec, | ||||
|     mk_codec, | ||||
|     mk_dec, | ||||
|     apply_codec, | ||||
|     current_codec, | ||||
| ) | ||||
| from tractor.msg.types import ( | ||||
|     log, | ||||
|     Started, | ||||
|     # _payload_msgs, | ||||
|     # PayloadMsg, | ||||
|     # mk_msg_spec, | ||||
| ) | ||||
| from tractor.msg._ops import ( | ||||
|     limit_plds, | ||||
| ) | ||||
| 
 | ||||
| def enc_nsp(obj: Any) -> Any: | ||||
|     actor: Actor = tractor.current_actor( | ||||
|         err_on_no_runtime=False, | ||||
|     ) | ||||
|     uid: tuple[str, str]|None = None if not actor else actor.uid | ||||
|     print(f'{uid} ENC HOOK') | ||||
| 
 | ||||
|     match obj: | ||||
|         # case NamespacePath()|str(): | ||||
|         case NamespacePath(): | ||||
|             encoded: str = str(obj) | ||||
|             print( | ||||
|                 f'----- ENCODING `NamespacePath` as `str` ------\n' | ||||
|                 f'|_obj:{type(obj)!r} = {obj!r}\n' | ||||
|                 f'|_encoded: str = {encoded!r}\n' | ||||
|             ) | ||||
|             # if type(obj) != NamespacePath: | ||||
|             #     breakpoint() | ||||
|             return encoded | ||||
|         case _: | ||||
|             logmsg: str = ( | ||||
|                 f'{uid}\n' | ||||
|                 'FAILED ENCODE\n' | ||||
|                 f'obj-> `{obj}: {type(obj)}`\n' | ||||
|             ) | ||||
|             raise NotImplementedError(logmsg) | ||||
| 
 | ||||
| 
 | ||||
| def dec_nsp( | ||||
|     obj_type: Type, | ||||
|     obj: Any, | ||||
| 
 | ||||
| ) -> Any: | ||||
|     # breakpoint() | ||||
|     actor: Actor = tractor.current_actor( | ||||
|         err_on_no_runtime=False, | ||||
|     ) | ||||
|     uid: tuple[str, str]|None = None if not actor else actor.uid | ||||
|     print( | ||||
|         f'{uid}\n' | ||||
|         'CUSTOM DECODE\n' | ||||
|         f'type-arg-> {obj_type}\n' | ||||
|         f'obj-arg-> `{obj}`: {type(obj)}\n' | ||||
|     ) | ||||
|     nsp = None | ||||
|     # XXX, never happens right? | ||||
|     if obj_type is Raw: | ||||
|         breakpoint() | ||||
| 
 | ||||
|     if ( | ||||
|         obj_type is NamespacePath | ||||
|         and isinstance(obj, str) | ||||
|         and ':' in obj | ||||
|     ): | ||||
|         nsp = NamespacePath(obj) | ||||
|         # TODO: we could built a generic handler using | ||||
|         # JUST matching the obj_type part? | ||||
|         # nsp = obj_type(obj) | ||||
| 
 | ||||
|     if nsp: | ||||
|         print(f'Returning NSP instance: {nsp}') | ||||
|         return nsp | ||||
| 
 | ||||
|     logmsg: str = ( | ||||
|         f'{uid}\n' | ||||
|         'FAILED DECODE\n' | ||||
|         f'type-> {obj_type}\n' | ||||
|         f'obj-arg-> `{obj}`: {type(obj)}\n\n' | ||||
|         f'current codec:\n' | ||||
|         f'{current_codec()}\n' | ||||
|     ) | ||||
|     # TODO: figure out the ignore subsys for this! | ||||
|     # -[ ] option whether to defense-relay backc the msg | ||||
|     #   inside an `Invalid`/`Ignore` | ||||
|     # -[ ] how to make this handling pluggable such that a | ||||
|     #   `Channel`/`MsgTransport` can intercept and process | ||||
|     #   back msgs either via exception handling or some other | ||||
|     #   signal? | ||||
|     log.warning(logmsg) | ||||
|     # NOTE: this delivers the invalid | ||||
|     # value up to `msgspec`'s decoding | ||||
|     # machinery for error raising. | ||||
|     return obj | ||||
|     # raise NotImplementedError(logmsg) | ||||
| 
 | ||||
| 
 | ||||
| def ex_func(*args): | ||||
|     ''' | ||||
|     A mod level func we can ref and load via our `NamespacePath` | ||||
|     python-object pointer `str` subtype. | ||||
| 
 | ||||
|     ''' | ||||
|     print(f'ex_func({args})') | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'add_codec_hooks', | ||||
|     [ | ||||
|         True, | ||||
|         False, | ||||
|     ], | ||||
|     ids=['use_codec_hooks', 'no_codec_hooks'], | ||||
| ) | ||||
| def test_custom_extension_types( | ||||
|     debug_mode: bool, | ||||
|     add_codec_hooks: bool | ||||
| ): | ||||
|     ''' | ||||
|     Verify that a `MsgCodec` (used for encoding all outbound IPC msgs | ||||
|     and decoding all inbound `PayloadMsg`s) and a paired `MsgDec` | ||||
|     (used for decoding the `PayloadMsg.pld: Raw` received within a given | ||||
|     task's ipc `Context` scope) can both send and receive "extension types" | ||||
|     as supported via custom converter hooks passed to `msgspec`. | ||||
| 
 | ||||
|     ''' | ||||
|     nsp_pld_dec: MsgDec = mk_dec( | ||||
|         spec=None,  # ONLY support the ext type | ||||
|         dec_hook=dec_nsp if add_codec_hooks else None, | ||||
|         ext_types=[NamespacePath], | ||||
|     ) | ||||
|     nsp_codec: MsgCodec = mk_codec( | ||||
|         # ipc_pld_spec=Raw,  # default! | ||||
| 
 | ||||
|         # NOTE XXX: the encode hook MUST be used no matter what since | ||||
|         # our `NamespacePath` is not any of a `Any` native type nor | ||||
|         # a `msgspec.Struct` subtype - so `msgspec` has no way to know | ||||
|         # how to encode it unless we provide the custom hook. | ||||
|         # | ||||
|         # AGAIN that is, regardless of whether we spec an | ||||
|         # `Any`-decoded-pld the enc has no knowledge (by default) | ||||
|         # how to enc `NamespacePath` (nsp), so we add a custom | ||||
|         # hook to do that ALWAYS. | ||||
|         enc_hook=enc_nsp if add_codec_hooks else None, | ||||
| 
 | ||||
|         # XXX NOTE: pretty sure this is mutex with the `type=` to | ||||
|         # `Decoder`? so it won't work in tandem with the | ||||
|         # `ipc_pld_spec` passed above? | ||||
|         ext_types=[NamespacePath], | ||||
| 
 | ||||
|         # TODO? is it useful to have the `.pld` decoded *prior* to | ||||
|         # the `PldRx`?? like perf or mem related? | ||||
|         # ext_dec=nsp_pld_dec, | ||||
|     ) | ||||
|     if add_codec_hooks: | ||||
|         assert nsp_codec.dec.dec_hook is None | ||||
| 
 | ||||
|         # TODO? if we pass `ext_dec` above? | ||||
|         # assert nsp_codec.dec.dec_hook is dec_nsp | ||||
| 
 | ||||
|         assert nsp_codec.enc.enc_hook is enc_nsp | ||||
| 
 | ||||
|     nsp = NamespacePath.from_ref(ex_func) | ||||
| 
 | ||||
|     try: | ||||
|         nsp_bytes: bytes = nsp_codec.encode(nsp) | ||||
|         nsp_rt_sin_msg = nsp_pld_dec.decode(nsp_bytes) | ||||
|         nsp_rt_sin_msg.load_ref() is ex_func | ||||
|     except TypeError: | ||||
|         if not add_codec_hooks: | ||||
|             pass | ||||
| 
 | ||||
|     try: | ||||
|         msg_bytes: bytes = nsp_codec.encode( | ||||
|             Started( | ||||
|                 cid='cid', | ||||
|                 pld=nsp, | ||||
|             ) | ||||
|         ) | ||||
|         # since the ext-type obj should also be set as the msg.pld | ||||
|         assert nsp_bytes in msg_bytes | ||||
|         started_rt: Started = nsp_codec.decode(msg_bytes) | ||||
|         pld: Raw = started_rt.pld | ||||
|         assert isinstance(pld, Raw) | ||||
|         nsp_rt: NamespacePath = nsp_pld_dec.decode(pld) | ||||
|         assert isinstance(nsp_rt, NamespacePath) | ||||
|         # in obj comparison terms they should be the same | ||||
|         assert nsp_rt == nsp | ||||
|         # ensure we've decoded to ext type! | ||||
|         assert nsp_rt.load_ref() is ex_func | ||||
| 
 | ||||
|     except TypeError: | ||||
|         if not add_codec_hooks: | ||||
|             pass | ||||
| 
 | ||||
| @tractor.context | ||||
| async def sleep_forever_in_sub( | ||||
|     ctx: Context, | ||||
| ) -> None: | ||||
|     await trio.sleep_forever() | ||||
| 
 | ||||
| 
 | ||||
| def mk_custom_codec( | ||||
|     add_hooks: bool, | ||||
| 
 | ||||
| ) -> tuple[ | ||||
|     MsgCodec,  # encode to send | ||||
|     MsgDec,  # pld receive-n-decode | ||||
| ]: | ||||
|     ''' | ||||
|     Create custom `msgpack` enc/dec-hooks and set a `Decoder` | ||||
|     which only loads `pld_spec` (like `NamespacePath`) types. | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
|     # XXX NOTE XXX: despite defining `NamespacePath` as a type | ||||
|     # field on our `PayloadMsg.pld`, we still need a enc/dec_hook() pair | ||||
|     # to cast to/from that type on the wire. See the docs: | ||||
|     # https://jcristharif.com/msgspec/extending.html#mapping-to-from-native-types | ||||
| 
 | ||||
|     # if pld_spec is Any: | ||||
|     #     pld_spec = Raw | ||||
| 
 | ||||
|     nsp_codec: MsgCodec = mk_codec( | ||||
|         # ipc_pld_spec=Raw,  # default! | ||||
| 
 | ||||
|         # NOTE XXX: the encode hook MUST be used no matter what since | ||||
|         # our `NamespacePath` is not any of a `Any` native type nor | ||||
|         # a `msgspec.Struct` subtype - so `msgspec` has no way to know | ||||
|         # how to encode it unless we provide the custom hook. | ||||
|         # | ||||
|         # AGAIN that is, regardless of whether we spec an | ||||
|         # `Any`-decoded-pld the enc has no knowledge (by default) | ||||
|         # how to enc `NamespacePath` (nsp), so we add a custom | ||||
|         # hook to do that ALWAYS. | ||||
|         enc_hook=enc_nsp if add_hooks else None, | ||||
| 
 | ||||
|         # XXX NOTE: pretty sure this is mutex with the `type=` to | ||||
|         # `Decoder`? so it won't work in tandem with the | ||||
|         # `ipc_pld_spec` passed above? | ||||
|         ext_types=[NamespacePath], | ||||
|     ) | ||||
|     # dec_hook=dec_nsp if add_hooks else None, | ||||
|     return nsp_codec | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'limit_plds_args', | ||||
|     [ | ||||
|         ( | ||||
|             {'dec_hook': None, 'ext_types': None}, | ||||
|             None, | ||||
|         ), | ||||
|         ( | ||||
|             {'dec_hook': dec_nsp, 'ext_types': None}, | ||||
|             TypeError, | ||||
|         ), | ||||
|         ( | ||||
|             {'dec_hook': dec_nsp, 'ext_types': [NamespacePath]}, | ||||
|             None, | ||||
|         ), | ||||
|         ( | ||||
|             {'dec_hook': dec_nsp, 'ext_types': [NamespacePath|None]}, | ||||
|             None, | ||||
|         ), | ||||
|     ], | ||||
|     ids=[ | ||||
|         'no_hook_no_ext_types', | ||||
|         'only_hook', | ||||
|         'hook_and_ext_types', | ||||
|         'hook_and_ext_types_w_null', | ||||
|     ] | ||||
| ) | ||||
| def test_pld_limiting_usage( | ||||
|     limit_plds_args: tuple[dict, Exception|None], | ||||
| ): | ||||
|     ''' | ||||
|     Verify `dec_hook()` and `ext_types` need to either both be | ||||
|     provided or we raise a explanator type-error. | ||||
| 
 | ||||
|     ''' | ||||
|     kwargs, maybe_err = limit_plds_args | ||||
|     async def main(): | ||||
|         async with tractor.open_nursery() as an:  # just to open runtime | ||||
| 
 | ||||
|             # XXX SHOULD NEVER WORK outside an ipc ctx scope! | ||||
|             try: | ||||
|                 with limit_plds(**kwargs): | ||||
|                     pass | ||||
|             except RuntimeError: | ||||
|                 pass | ||||
| 
 | ||||
|             p: tractor.Portal = await an.start_actor( | ||||
|                 'sub', | ||||
|                 enable_modules=[__name__], | ||||
|             ) | ||||
|             async with ( | ||||
|                 p.open_context( | ||||
|                     sleep_forever_in_sub | ||||
|                 ) as (ctx, first), | ||||
|             ): | ||||
|                 try: | ||||
|                     with limit_plds(**kwargs): | ||||
|                         pass | ||||
|                 except maybe_err as exc: | ||||
|                     assert type(exc) is maybe_err | ||||
|                     pass | ||||
| 
 | ||||
| 
 | ||||
| def chk_codec_applied( | ||||
|     expect_codec: MsgCodec|None, | ||||
|     enter_value: MsgCodec|None = None, | ||||
| 
 | ||||
| ) -> MsgCodec: | ||||
|     ''' | ||||
|     buncha sanity checks ensuring that the IPC channel's | ||||
|     context-vars are set to the expected codec and that are | ||||
|     ctx-var wrapper APIs match the same. | ||||
| 
 | ||||
|     ''' | ||||
|     # TODO: play with tricyle again, bc this is supposed to work | ||||
|     # the way we want? | ||||
|     # | ||||
|     # TreeVar | ||||
|     # task: trio.Task = trio.lowlevel.current_task() | ||||
|     # curr_codec = _ctxvar_MsgCodec.get_in(task) | ||||
| 
 | ||||
|     # ContextVar | ||||
|     # task_ctx: Context = task.context | ||||
|     # assert _ctxvar_MsgCodec in task_ctx | ||||
|     # curr_codec: MsgCodec = task.context[_ctxvar_MsgCodec] | ||||
|     if expect_codec is None: | ||||
|         assert enter_value is None | ||||
|         return | ||||
| 
 | ||||
|     # NOTE: currently we use this! | ||||
|     # RunVar | ||||
|     curr_codec: MsgCodec = current_codec() | ||||
|     last_read_codec = _ctxvar_MsgCodec.get() | ||||
|     # assert curr_codec is last_read_codec | ||||
| 
 | ||||
|     assert ( | ||||
|         (same_codec := expect_codec) is | ||||
|         # returned from `mk_codec()` | ||||
| 
 | ||||
|         # yielded value from `apply_codec()` | ||||
| 
 | ||||
|         # read from current task's `contextvars.Context` | ||||
|         curr_codec is | ||||
|         last_read_codec | ||||
| 
 | ||||
|         # the default `msgspec` settings | ||||
|         is not _codec._def_msgspec_codec | ||||
|         is not _codec._def_tractor_codec | ||||
|     ) | ||||
| 
 | ||||
|     if enter_value: | ||||
|         assert enter_value is same_codec | ||||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
| async def send_back_values( | ||||
|     ctx: Context, | ||||
|     rent_pld_spec_type_strs: list[str], | ||||
|     add_hooks: bool, | ||||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Setup up a custom codec to load instances of `NamespacePath` | ||||
|     and ensure we can round trip a func ref with our parent. | ||||
| 
 | ||||
|     ''' | ||||
|     uid: tuple = tractor.current_actor().uid | ||||
| 
 | ||||
|     # init state in sub-actor should be default | ||||
|     chk_codec_applied( | ||||
|         expect_codec=_codec._def_tractor_codec, | ||||
|     ) | ||||
| 
 | ||||
|     # load pld spec from input str | ||||
|     rent_pld_spec = _exts.dec_type_union( | ||||
|         rent_pld_spec_type_strs, | ||||
|         mods=[ | ||||
|             importlib.import_module(__name__), | ||||
|         ], | ||||
|     ) | ||||
|     rent_pld_spec_types: set[Type] = _codec.unpack_spec_types( | ||||
|         rent_pld_spec, | ||||
|     ) | ||||
| 
 | ||||
|     # ONLY add ext-hooks if the rent specified a non-std type! | ||||
|     add_hooks: bool = ( | ||||
|         NamespacePath in rent_pld_spec_types | ||||
|         and | ||||
|         add_hooks | ||||
|     ) | ||||
| 
 | ||||
|     # same as on parent side config. | ||||
|     nsp_codec: MsgCodec|None = None | ||||
|     if add_hooks: | ||||
|         nsp_codec = mk_codec( | ||||
|             enc_hook=enc_nsp, | ||||
|             ext_types=[NamespacePath], | ||||
|         ) | ||||
| 
 | ||||
|     with ( | ||||
|         maybe_apply_codec(nsp_codec) as codec, | ||||
|         limit_plds( | ||||
|             rent_pld_spec, | ||||
|             dec_hook=dec_nsp if add_hooks else None, | ||||
|             ext_types=[NamespacePath]  if add_hooks else None, | ||||
|         ) as pld_dec, | ||||
|     ): | ||||
|         # ?XXX? SHOULD WE NOT be swapping the global codec since it | ||||
|         # breaks `Context.started()` roundtripping checks?? | ||||
|         chk_codec_applied( | ||||
|             expect_codec=nsp_codec, | ||||
|             enter_value=codec, | ||||
|         ) | ||||
| 
 | ||||
|         # ?TODO, mismatch case(s)? | ||||
|         # | ||||
|         # ensure pld spec matches on both sides | ||||
|         ctx_pld_dec: MsgDec = ctx._pld_rx._pld_dec | ||||
|         assert pld_dec is ctx_pld_dec | ||||
|         child_pld_spec: Type = pld_dec.spec | ||||
|         child_pld_spec_types: set[Type] = _codec.unpack_spec_types( | ||||
|             child_pld_spec, | ||||
|         ) | ||||
|         assert ( | ||||
|             child_pld_spec_types.issuperset( | ||||
|                 rent_pld_spec_types | ||||
|             ) | ||||
|         ) | ||||
| 
 | ||||
|         # ?TODO, try loop for each of the types in pld-superset? | ||||
|         # | ||||
|         # for send_value in [ | ||||
|         #     nsp, | ||||
|         #     str(nsp), | ||||
|         #     None, | ||||
|         # ]: | ||||
|         nsp = NamespacePath.from_ref(ex_func) | ||||
|         try: | ||||
|             print( | ||||
|                 f'{uid}: attempting to `.started({nsp})`\n' | ||||
|                 f'\n' | ||||
|                 f'rent_pld_spec: {rent_pld_spec}\n' | ||||
|                 f'child_pld_spec: {child_pld_spec}\n' | ||||
|                 f'codec: {codec}\n' | ||||
|             ) | ||||
|             # await tractor.pause() | ||||
|             await ctx.started(nsp) | ||||
| 
 | ||||
|         except tractor.MsgTypeError as _mte: | ||||
|             mte = _mte | ||||
| 
 | ||||
|             # false -ve case | ||||
|             if add_hooks: | ||||
|                 raise RuntimeError( | ||||
|                     f'EXPECTED to `.started()` value given spec ??\n\n' | ||||
|                     f'child_pld_spec -> {child_pld_spec}\n' | ||||
|                     f'value = {nsp}: {type(nsp)}\n' | ||||
|                 ) | ||||
| 
 | ||||
|             # true -ve case | ||||
|             raise mte | ||||
| 
 | ||||
|         # TODO: maybe we should add our own wrapper error so as to | ||||
|         # be interchange-lib agnostic? | ||||
|         # -[ ] the error type is wtv is raised from the hook so we | ||||
|         #   could also require a type-class of errors for | ||||
|         #   indicating whether the hook-failure can be handled by | ||||
|         #   a nasty-dialog-unprot sub-sys? | ||||
|         except TypeError as typerr: | ||||
|             # false -ve | ||||
|             if add_hooks: | ||||
|                 raise RuntimeError('Should have been able to send `nsp`??') | ||||
| 
 | ||||
|             # true -ve | ||||
|             print('Failed to send `nsp` due to no ext hooks set!') | ||||
|             raise typerr | ||||
| 
 | ||||
|         # now try sending a set of valid and invalid plds to ensure | ||||
|         # the pld spec is respected. | ||||
|         sent: list[Any] = [] | ||||
|         async with ctx.open_stream() as ipc: | ||||
|             print( | ||||
|                 f'{uid}: streaming all pld types to rent..' | ||||
|             ) | ||||
| 
 | ||||
|             # for send_value, expect_send in iter_send_val_items: | ||||
|             for send_value in [ | ||||
|                 nsp, | ||||
|                 str(nsp), | ||||
|                 None, | ||||
|             ]: | ||||
|                 send_type: Type = type(send_value) | ||||
|                 print( | ||||
|                     f'{uid}: SENDING NEXT pld\n' | ||||
|                     f'send_type: {send_type}\n' | ||||
|                     f'send_value: {send_value}\n' | ||||
|                 ) | ||||
|                 try: | ||||
|                     await ipc.send(send_value) | ||||
|                     sent.append(send_value) | ||||
| 
 | ||||
|                 except ValidationError as valerr: | ||||
|                     print(f'{uid} FAILED TO SEND {send_value}!') | ||||
| 
 | ||||
|                     # false -ve | ||||
|                     if add_hooks: | ||||
|                         raise RuntimeError( | ||||
|                             f'EXPECTED to roundtrip value given spec:\n' | ||||
|                             f'rent_pld_spec -> {rent_pld_spec}\n' | ||||
|                             f'child_pld_spec -> {child_pld_spec}\n' | ||||
|                             f'value = {send_value}: {send_type}\n' | ||||
|                         ) | ||||
| 
 | ||||
|                     # true -ve | ||||
|                     raise valerr | ||||
|                     # continue | ||||
| 
 | ||||
|             else: | ||||
|                 print( | ||||
|                     f'{uid}: finished sending all values\n' | ||||
|                     'Should be exiting stream block!\n' | ||||
|                 ) | ||||
| 
 | ||||
|         print(f'{uid}: exited streaming block!') | ||||
| 
 | ||||
| 
 | ||||
| 
 | ||||
| @cm | ||||
| def maybe_apply_codec(codec: MsgCodec|None) -> MsgCodec|None: | ||||
|     if codec is None: | ||||
|         yield None | ||||
|         return | ||||
| 
 | ||||
|     with apply_codec(codec) as codec: | ||||
|         yield codec | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'pld_spec', | ||||
|     [ | ||||
|         Any, | ||||
|         NamespacePath, | ||||
|         NamespacePath|None,  # the "maybe" spec Bo | ||||
|     ], | ||||
|     ids=[ | ||||
|         'any_type', | ||||
|         'only_nsp_ext', | ||||
|         'maybe_nsp_ext', | ||||
|     ] | ||||
| ) | ||||
| @pytest.mark.parametrize( | ||||
|     'add_hooks', | ||||
|     [ | ||||
|         True, | ||||
|         False, | ||||
|     ], | ||||
|     ids=[ | ||||
|         'use_codec_hooks', | ||||
|         'no_codec_hooks', | ||||
|     ], | ||||
| ) | ||||
| def test_ext_types_over_ipc( | ||||
|     debug_mode: bool, | ||||
|     pld_spec: Union[Type], | ||||
|     add_hooks: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Ensure we can support extension types coverted using | ||||
|     `enc/dec_hook()`s passed to the `.msg.limit_plds()` API | ||||
|     and that sane errors happen when we try do the same without | ||||
|     the codec hooks. | ||||
| 
 | ||||
|     ''' | ||||
|     pld_types: set[Type] = _codec.unpack_spec_types(pld_spec) | ||||
| 
 | ||||
|     async def main(): | ||||
| 
 | ||||
|         # sanity check the default pld-spec beforehand | ||||
|         chk_codec_applied( | ||||
|             expect_codec=_codec._def_tractor_codec, | ||||
|         ) | ||||
| 
 | ||||
|         # extension type we want to send as msg payload | ||||
|         nsp = NamespacePath.from_ref(ex_func) | ||||
| 
 | ||||
|         # ^NOTE, 2 cases: | ||||
|         # - codec hooks noto added -> decode nsp as `str` | ||||
|         # - codec with hooks -> decode nsp as `NamespacePath` | ||||
|         nsp_codec: MsgCodec|None = None | ||||
|         if ( | ||||
|             NamespacePath in pld_types | ||||
|             and | ||||
|             add_hooks | ||||
|         ): | ||||
|             nsp_codec = mk_codec( | ||||
|                 enc_hook=enc_nsp, | ||||
|                 ext_types=[NamespacePath], | ||||
|             ) | ||||
| 
 | ||||
|         async with tractor.open_nursery( | ||||
|             debug_mode=debug_mode, | ||||
|         ) as an: | ||||
|             p: tractor.Portal = await an.start_actor( | ||||
|                 'sub', | ||||
|                 enable_modules=[__name__], | ||||
|             ) | ||||
|             with ( | ||||
|                 maybe_apply_codec(nsp_codec) as codec, | ||||
|             ): | ||||
|                 chk_codec_applied( | ||||
|                     expect_codec=nsp_codec, | ||||
|                     enter_value=codec, | ||||
|                 ) | ||||
|                 rent_pld_spec_type_strs: list[str] = _exts.enc_type_union(pld_spec) | ||||
| 
 | ||||
|                 # XXX should raise an mte (`MsgTypeError`) | ||||
|                 # when `add_hooks == False` bc the input | ||||
|                 # `expect_ipc_send` kwarg has a nsp which can't be | ||||
|                 # serialized! | ||||
|                 # | ||||
|                 # TODO:can we ensure this happens from the | ||||
|                 # `Return`-side (aka the sub) as well? | ||||
|                 try: | ||||
|                     ctx: tractor.Context | ||||
|                     ipc: tractor.MsgStream | ||||
|                     async with ( | ||||
| 
 | ||||
|                         # XXX should raise an mte (`MsgTypeError`) | ||||
|                         # when `add_hooks == False`.. | ||||
|                         p.open_context( | ||||
|                             send_back_values, | ||||
|                             # expect_debug=debug_mode, | ||||
|                             rent_pld_spec_type_strs=rent_pld_spec_type_strs, | ||||
|                             add_hooks=add_hooks, | ||||
|                             # expect_ipc_send=expect_ipc_send, | ||||
|                         ) as (ctx, first), | ||||
| 
 | ||||
|                         ctx.open_stream() as ipc, | ||||
|                     ): | ||||
|                         with ( | ||||
|                             limit_plds( | ||||
|                                 pld_spec, | ||||
|                                 dec_hook=dec_nsp if add_hooks else None, | ||||
|                                 ext_types=[NamespacePath]  if add_hooks else None, | ||||
|                             ) as pld_dec, | ||||
|                         ): | ||||
|                             ctx_pld_dec: MsgDec = ctx._pld_rx._pld_dec | ||||
|                             assert pld_dec is ctx_pld_dec | ||||
| 
 | ||||
|                             # if ( | ||||
|                             #     not add_hooks | ||||
|                             #     and | ||||
|                             #     NamespacePath in  | ||||
|                             # ): | ||||
|                             #     pytest.fail('ctx should fail to open without custom enc_hook!?') | ||||
| 
 | ||||
|                             await ipc.send(nsp) | ||||
|                             nsp_rt = await ipc.receive() | ||||
| 
 | ||||
|                             assert nsp_rt == nsp | ||||
|                             assert nsp_rt.load_ref() is ex_func | ||||
| 
 | ||||
|                 # this test passes bc we can go no further! | ||||
|                 except MsgTypeError as mte: | ||||
|                     # if not add_hooks: | ||||
|                     #     # teardown nursery | ||||
|                     #     await p.cancel_actor() | ||||
|                         # return | ||||
| 
 | ||||
|                     raise mte | ||||
| 
 | ||||
|             await p.cancel_actor() | ||||
| 
 | ||||
|     if ( | ||||
|         NamespacePath in pld_types | ||||
|         and | ||||
|         add_hooks | ||||
|     ): | ||||
|         trio.run(main) | ||||
| 
 | ||||
|     else: | ||||
|         with pytest.raises( | ||||
|             expected_exception=tractor.RemoteActorError, | ||||
|         ) as excinfo: | ||||
|             trio.run(main) | ||||
| 
 | ||||
|         exc = excinfo.value | ||||
|         # bc `.started(nsp: NamespacePath)` will raise | ||||
|         assert exc.boxed_type is TypeError | ||||
| 
 | ||||
| 
 | ||||
| # def chk_pld_type( | ||||
| #     payload_spec: Type[Struct]|Any, | ||||
| #     pld: Any, | ||||
| 
 | ||||
| #     expect_roundtrip: bool|None = None, | ||||
| 
 | ||||
| # ) -> bool: | ||||
| 
 | ||||
| #     pld_val_type: Type = type(pld) | ||||
| 
 | ||||
| #     # TODO: verify that the overridden subtypes | ||||
| #     # DO NOT have modified type-annots from original! | ||||
| #     # 'Start',  .pld: FuncSpec | ||||
| #     # 'StartAck',  .pld: IpcCtxSpec | ||||
| #     # 'Stop',  .pld: UNSEt | ||||
| #     # 'Error',  .pld: ErrorData | ||||
| 
 | ||||
| #     codec: MsgCodec = mk_codec( | ||||
| #         # NOTE: this ONLY accepts `PayloadMsg.pld` fields of a specified | ||||
| #         # type union. | ||||
| #         ipc_pld_spec=payload_spec, | ||||
| #     ) | ||||
| 
 | ||||
| #     # make a one-off dec to compare with our `MsgCodec` instance | ||||
| #     # which does the below `mk_msg_spec()` call internally | ||||
| #     ipc_msg_spec: Union[Type[Struct]] | ||||
| #     msg_types: list[PayloadMsg[payload_spec]] | ||||
| #     ( | ||||
| #         ipc_msg_spec, | ||||
| #         msg_types, | ||||
| #     ) = mk_msg_spec( | ||||
| #         payload_type_union=payload_spec, | ||||
| #     ) | ||||
| #     _enc = msgpack.Encoder() | ||||
| #     _dec = msgpack.Decoder( | ||||
| #         type=ipc_msg_spec or Any,  # like `PayloadMsg[Any]` | ||||
| #     ) | ||||
| 
 | ||||
| #     assert ( | ||||
| #         payload_spec | ||||
| #         == | ||||
| #         codec.pld_spec | ||||
| #     ) | ||||
| 
 | ||||
| #     # assert codec.dec == dec | ||||
| #     # | ||||
| #     # ^-XXX-^ not sure why these aren't "equal" but when cast | ||||
| #     # to `str` they seem to match ?? .. kk | ||||
| 
 | ||||
| #     assert ( | ||||
| #         str(ipc_msg_spec) | ||||
| #         == | ||||
| #         str(codec.msg_spec) | ||||
| #         == | ||||
| #         str(_dec.type) | ||||
| #         == | ||||
| #         str(codec.dec.type) | ||||
| #     ) | ||||
| 
 | ||||
| #     # verify the boxed-type for all variable payload-type msgs. | ||||
| #     if not msg_types: | ||||
| #         breakpoint() | ||||
| 
 | ||||
| #     roundtrip: bool|None = None | ||||
| #     pld_spec_msg_names: list[str] = [ | ||||
| #         td.__name__ for td in _payload_msgs | ||||
| #     ] | ||||
| #     for typedef in msg_types: | ||||
| 
 | ||||
| #         skip_runtime_msg: bool = typedef.__name__ not in pld_spec_msg_names | ||||
| #         if skip_runtime_msg: | ||||
| #             continue | ||||
| 
 | ||||
| #         pld_field = structs.fields(typedef)[1] | ||||
| #         assert pld_field.type is payload_spec # TODO-^ does this need to work to get all subtypes to adhere? | ||||
| 
 | ||||
| #         kwargs: dict[str, Any] = { | ||||
| #             'cid': '666', | ||||
| #             'pld': pld, | ||||
| #         } | ||||
| #         enc_msg: PayloadMsg = typedef(**kwargs) | ||||
| 
 | ||||
| #         _wire_bytes: bytes = _enc.encode(enc_msg) | ||||
| #         wire_bytes: bytes = codec.enc.encode(enc_msg) | ||||
| #         assert _wire_bytes == wire_bytes | ||||
| 
 | ||||
| #         ve: ValidationError|None = None | ||||
| #         try: | ||||
| #             dec_msg = codec.dec.decode(wire_bytes) | ||||
| #             _dec_msg = _dec.decode(wire_bytes) | ||||
| 
 | ||||
| #             # decoded msg and thus payload should be exactly same! | ||||
| #             assert (roundtrip := ( | ||||
| #                 _dec_msg | ||||
| #                 == | ||||
| #                 dec_msg | ||||
| #                 == | ||||
| #                 enc_msg | ||||
| #             )) | ||||
| 
 | ||||
| #             if ( | ||||
| #                 expect_roundtrip is not None | ||||
| #                 and expect_roundtrip != roundtrip | ||||
| #             ): | ||||
| #                 breakpoint() | ||||
| 
 | ||||
| #             assert ( | ||||
| #                 pld | ||||
| #                 == | ||||
| #                 dec_msg.pld | ||||
| #                 == | ||||
| #                 enc_msg.pld | ||||
| #             ) | ||||
| #             # assert (roundtrip := (_dec_msg == enc_msg)) | ||||
| 
 | ||||
| #         except ValidationError as _ve: | ||||
| #             ve = _ve | ||||
| #             roundtrip: bool = False | ||||
| #             if pld_val_type is payload_spec: | ||||
| #                 raise ValueError( | ||||
| #                    'Got `ValidationError` despite type-var match!?\n' | ||||
| #                     f'pld_val_type: {pld_val_type}\n' | ||||
| #                     f'payload_type: {payload_spec}\n' | ||||
| #                 ) from ve | ||||
| 
 | ||||
| #             else: | ||||
| #                 # ow we good cuz the pld spec mismatched. | ||||
| #                 print( | ||||
| #                     'Got expected `ValidationError` since,\n' | ||||
| #                     f'{pld_val_type} is not {payload_spec}\n' | ||||
| #                 ) | ||||
| #         else: | ||||
| #             if ( | ||||
| #                 payload_spec is not Any | ||||
| #                 and | ||||
| #                 pld_val_type is not payload_spec | ||||
| #             ): | ||||
| #                 raise ValueError( | ||||
| #                    'DID NOT `ValidationError` despite expected type match!?\n' | ||||
| #                     f'pld_val_type: {pld_val_type}\n' | ||||
| #                     f'payload_type: {payload_spec}\n' | ||||
| #                 ) | ||||
| 
 | ||||
| #     # full code decode should always be attempted! | ||||
| #     if roundtrip is None: | ||||
| #         breakpoint() | ||||
| 
 | ||||
| #     return roundtrip | ||||
| 
 | ||||
| 
 | ||||
| # ?TODO? maybe remove since covered in the newer `test_pldrx_limiting` | ||||
| # via end-2-end testing of all this? | ||||
| # -[ ] IOW do we really NEED this lowlevel unit testing? | ||||
| # | ||||
| # def test_limit_msgspec( | ||||
| #     debug_mode: bool, | ||||
| # ): | ||||
| #     ''' | ||||
| #     Internals unit testing to verify that type-limiting an IPC ctx's | ||||
| #     msg spec with `Pldrx.limit_plds()` results in various | ||||
| #     encapsulated `msgspec` object settings and state. | ||||
| 
 | ||||
| #     ''' | ||||
| #     async def main(): | ||||
| #         async with tractor.open_root_actor( | ||||
| #             debug_mode=debug_mode, | ||||
| #         ): | ||||
| #             # ensure we can round-trip a boxing `PayloadMsg` | ||||
| #             assert chk_pld_type( | ||||
| #                 payload_spec=Any, | ||||
| #                 pld=None, | ||||
| #                 expect_roundtrip=True, | ||||
| #             ) | ||||
| 
 | ||||
| #             # verify that a mis-typed payload value won't decode | ||||
| #             assert not chk_pld_type( | ||||
| #                 payload_spec=int, | ||||
| #                 pld='doggy', | ||||
| #             ) | ||||
| 
 | ||||
| #             # parametrize the boxed `.pld` type as a custom-struct | ||||
| #             # and ensure that parametrization propagates | ||||
| #             # to all payload-msg-spec-able subtypes! | ||||
| #             class CustomPayload(Struct): | ||||
| #                 name: str | ||||
| #                 value: Any | ||||
| 
 | ||||
| #             assert not chk_pld_type( | ||||
| #                 payload_spec=CustomPayload, | ||||
| #                 pld='doggy', | ||||
| #             ) | ||||
| 
 | ||||
| #             assert chk_pld_type( | ||||
| #                 payload_spec=CustomPayload, | ||||
| #                 pld=CustomPayload(name='doggy', value='urmom') | ||||
| #             ) | ||||
| 
 | ||||
| #             # yah, we can `.pause_from_sync()` now! | ||||
| #             # breakpoint() | ||||
| 
 | ||||
| #     trio.run(main) | ||||
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							|  | @ -170,7 +170,7 @@ def test_do_not_swallow_error_before_started_by_remote_contextcancelled( | |||
|         trio.run(main) | ||||
| 
 | ||||
|     rae = excinfo.value | ||||
|     assert rae.boxed_type == TypeError | ||||
|     assert rae.boxed_type is TypeError | ||||
| 
 | ||||
| 
 | ||||
| @tractor.context | ||||
|  |  | |||
|  | @ -0,0 +1,248 @@ | |||
| ''' | ||||
| Special attention cases for using "infect `asyncio`" mode from a root | ||||
| actor; i.e. not using a std `trio.run()` bootstrap. | ||||
| 
 | ||||
| ''' | ||||
| import asyncio | ||||
| from functools import partial | ||||
| 
 | ||||
| import pytest | ||||
| import trio | ||||
| import tractor | ||||
| from tractor import ( | ||||
|     to_asyncio, | ||||
| ) | ||||
| from tests.test_infected_asyncio import ( | ||||
|     aio_echo_server, | ||||
| ) | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'raise_error_mid_stream', | ||||
|     [ | ||||
|         False, | ||||
|         Exception, | ||||
|         KeyboardInterrupt, | ||||
|     ], | ||||
|     ids='raise_error={}'.format, | ||||
| ) | ||||
| def test_infected_root_actor( | ||||
|     raise_error_mid_stream: bool|Exception, | ||||
| 
 | ||||
|     # conftest wide | ||||
|     loglevel: str, | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Verify you can run the `tractor` runtime with `Actor.is_infected_aio() == True` | ||||
|     in the root actor. | ||||
| 
 | ||||
|     ''' | ||||
|     async def _trio_main(): | ||||
|         with trio.fail_after(2 if not debug_mode else 999): | ||||
|             first: str | ||||
|             chan: to_asyncio.LinkedTaskChannel | ||||
|             async with ( | ||||
|                 tractor.open_root_actor( | ||||
|                     debug_mode=debug_mode, | ||||
|                     loglevel=loglevel, | ||||
|                 ), | ||||
|                 to_asyncio.open_channel_from( | ||||
|                     aio_echo_server, | ||||
|                 ) as (first, chan), | ||||
|             ): | ||||
|                 assert first == 'start' | ||||
| 
 | ||||
|                 for i in range(1000): | ||||
|                     await chan.send(i) | ||||
|                     out = await chan.receive() | ||||
|                     assert out == i | ||||
|                     print(f'asyncio echoing {i}') | ||||
| 
 | ||||
|                     if ( | ||||
|                         raise_error_mid_stream | ||||
|                         and | ||||
|                         i == 500 | ||||
|                     ): | ||||
|                         raise raise_error_mid_stream | ||||
| 
 | ||||
|                     if out is None: | ||||
|                         try: | ||||
|                             out = await chan.receive() | ||||
|                         except trio.EndOfChannel: | ||||
|                             break | ||||
|                         else: | ||||
|                             raise RuntimeError( | ||||
|                                 'aio channel never stopped?' | ||||
|                             ) | ||||
| 
 | ||||
|     if raise_error_mid_stream: | ||||
|         with pytest.raises(raise_error_mid_stream): | ||||
|             tractor.to_asyncio.run_as_asyncio_guest( | ||||
|                 trio_main=_trio_main, | ||||
|             ) | ||||
|     else: | ||||
|         tractor.to_asyncio.run_as_asyncio_guest( | ||||
|             trio_main=_trio_main, | ||||
|         ) | ||||
| 
 | ||||
| 
 | ||||
| 
 | ||||
| async def sync_and_err( | ||||
|     # just signature placeholders for compat with | ||||
|     # ``to_asyncio.open_channel_from()`` | ||||
|     to_trio: trio.MemorySendChannel, | ||||
|     from_trio: asyncio.Queue, | ||||
|     ev: asyncio.Event, | ||||
| 
 | ||||
| ): | ||||
|     if to_trio: | ||||
|         to_trio.send_nowait('start') | ||||
| 
 | ||||
|     await ev.wait() | ||||
|     raise RuntimeError('asyncio-side') | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     'aio_err_trigger', | ||||
|     [ | ||||
|         'before_start_point', | ||||
|         'after_trio_task_starts', | ||||
|         'after_start_point', | ||||
|     ], | ||||
|     ids='aio_err_triggered={}'.format | ||||
| ) | ||||
| def test_trio_prestarted_task_bubbles( | ||||
|     aio_err_trigger: str, | ||||
| 
 | ||||
|     # conftest wide | ||||
|     loglevel: str, | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|     async def pre_started_err( | ||||
|         raise_err: bool = False, | ||||
|         pre_sleep: float|None = None, | ||||
|         aio_trigger: asyncio.Event|None = None, | ||||
|         task_status=trio.TASK_STATUS_IGNORED, | ||||
|     ): | ||||
|         ''' | ||||
|         Maybe pre-started error then sleep. | ||||
| 
 | ||||
|         ''' | ||||
|         if pre_sleep is not None: | ||||
|             print(f'Sleeping from trio for {pre_sleep!r}s !') | ||||
|             await trio.sleep(pre_sleep) | ||||
| 
 | ||||
|         # signal aio-task to raise JUST AFTER this task | ||||
|         # starts but has not yet `.started()` | ||||
|         if aio_trigger: | ||||
|             print('Signalling aio-task to raise from `trio`!!') | ||||
|             aio_trigger.set() | ||||
| 
 | ||||
|         if raise_err: | ||||
|             print('Raising from trio!') | ||||
|             raise TypeError('trio-side') | ||||
| 
 | ||||
|         task_status.started() | ||||
|         await trio.sleep_forever() | ||||
| 
 | ||||
|     async def _trio_main(): | ||||
|         # with trio.fail_after(2): | ||||
|         with trio.fail_after(999): | ||||
|             first: str | ||||
|             chan: to_asyncio.LinkedTaskChannel | ||||
|             aio_ev = asyncio.Event() | ||||
| 
 | ||||
|             async with ( | ||||
|                 tractor.open_root_actor( | ||||
|                     debug_mode=False, | ||||
|                     loglevel=loglevel, | ||||
|                 ), | ||||
|             ): | ||||
|                 # TODO, tests for this with 3.13 egs? | ||||
|                 # from tractor.devx import open_crash_handler | ||||
|                 # with open_crash_handler(): | ||||
|                 async with ( | ||||
|                     # where we'll start a sub-task that errors BEFORE | ||||
|                     # calling `.started()` such that the error should | ||||
|                     # bubble before the guest run terminates! | ||||
|                     trio.open_nursery() as tn, | ||||
| 
 | ||||
|                     # THEN start an infect task which should error just | ||||
|                     # after the trio-side's task does. | ||||
|                     to_asyncio.open_channel_from( | ||||
|                         partial( | ||||
|                             sync_and_err, | ||||
|                             ev=aio_ev, | ||||
|                         ) | ||||
|                     ) as (first, chan), | ||||
|                 ): | ||||
| 
 | ||||
|                     for i in range(5): | ||||
|                         pre_sleep: float|None = None | ||||
|                         last_iter: bool = (i == 4) | ||||
| 
 | ||||
|                         # TODO, missing cases? | ||||
|                         # -[ ] error as well on | ||||
|                         #    'after_start_point' case as well for | ||||
|                         #    another case? | ||||
|                         raise_err: bool = False | ||||
| 
 | ||||
|                         if last_iter: | ||||
|                             raise_err: bool = True | ||||
| 
 | ||||
|                             # trigger aio task to error on next loop | ||||
|                             # tick/checkpoint | ||||
|                             if aio_err_trigger == 'before_start_point': | ||||
|                                 aio_ev.set() | ||||
| 
 | ||||
|                             pre_sleep: float = 0 | ||||
| 
 | ||||
|                         await tn.start( | ||||
|                             pre_started_err, | ||||
|                             raise_err, | ||||
|                             pre_sleep, | ||||
|                             (aio_ev if ( | ||||
|                                     aio_err_trigger == 'after_trio_task_starts' | ||||
|                                     and | ||||
|                                     last_iter | ||||
|                                 ) else None | ||||
|                             ), | ||||
|                         ) | ||||
| 
 | ||||
|                         if ( | ||||
|                             aio_err_trigger == 'after_start_point' | ||||
|                             and | ||||
|                             last_iter | ||||
|                         ): | ||||
|                             aio_ev.set() | ||||
| 
 | ||||
|     with pytest.raises( | ||||
|         expected_exception=ExceptionGroup, | ||||
|     ) as excinfo: | ||||
|         tractor.to_asyncio.run_as_asyncio_guest( | ||||
|             trio_main=_trio_main, | ||||
|         ) | ||||
| 
 | ||||
|     eg = excinfo.value | ||||
|     rte_eg, rest_eg = eg.split(RuntimeError) | ||||
| 
 | ||||
|     # ensure the trio-task's error bubbled despite the aio-side | ||||
|     # having (maybe) errored first. | ||||
|     if aio_err_trigger in ( | ||||
|         'after_trio_task_starts', | ||||
|         'after_start_point', | ||||
|     ): | ||||
|         assert len(errs := rest_eg.exceptions) == 1 | ||||
|         typerr = errs[0] | ||||
|         assert ( | ||||
|             type(typerr) is TypeError | ||||
|             and | ||||
|             'trio-side' in typerr.args | ||||
|         ) | ||||
| 
 | ||||
|     # when aio errors BEFORE (last) trio task is scheduled, we should | ||||
|     # never see anythinb but the aio-side. | ||||
|     else: | ||||
|         assert len(rtes := rte_eg.exceptions) == 1 | ||||
|         assert 'asyncio-side' in rtes[0].args[0] | ||||
|  | @ -2,7 +2,9 @@ | |||
| Broadcast channels for fan-out to local tasks. | ||||
| 
 | ||||
| """ | ||||
| from contextlib import asynccontextmanager | ||||
| from contextlib import ( | ||||
|     asynccontextmanager as acm, | ||||
| ) | ||||
| from functools import partial | ||||
| from itertools import cycle | ||||
| import time | ||||
|  | @ -15,6 +17,7 @@ import tractor | |||
| from tractor.trionics import ( | ||||
|     broadcast_receiver, | ||||
|     Lagged, | ||||
|     collapse_eg, | ||||
| ) | ||||
| 
 | ||||
| 
 | ||||
|  | @ -62,7 +65,7 @@ async def ensure_sequence( | |||
|                 break | ||||
| 
 | ||||
| 
 | ||||
| @asynccontextmanager | ||||
| @acm | ||||
| async def open_sequence_streamer( | ||||
| 
 | ||||
|     sequence: list[int], | ||||
|  | @ -74,9 +77,9 @@ async def open_sequence_streamer( | |||
|     async with tractor.open_nursery( | ||||
|         arbiter_addr=reg_addr, | ||||
|         start_method=start_method, | ||||
|     ) as tn: | ||||
|     ) as an: | ||||
| 
 | ||||
|         portal = await tn.start_actor( | ||||
|         portal = await an.start_actor( | ||||
|             'sequence_echoer', | ||||
|             enable_modules=[__name__], | ||||
|         ) | ||||
|  | @ -155,9 +158,12 @@ def test_consumer_and_parent_maybe_lag( | |||
|         ) as stream: | ||||
| 
 | ||||
|             try: | ||||
|                 async with trio.open_nursery() as n: | ||||
|                 async with ( | ||||
|                     collapse_eg(), | ||||
|                     trio.open_nursery() as tn, | ||||
|                 ): | ||||
| 
 | ||||
|                     n.start_soon( | ||||
|                     tn.start_soon( | ||||
|                         ensure_sequence, | ||||
|                         stream, | ||||
|                         sequence.copy(), | ||||
|  | @ -230,8 +236,8 @@ def test_faster_task_to_recv_is_cancelled_by_slower( | |||
| 
 | ||||
|         ) as stream: | ||||
| 
 | ||||
|             async with trio.open_nursery() as n: | ||||
|                 n.start_soon( | ||||
|             async with trio.open_nursery() as tn: | ||||
|                 tn.start_soon( | ||||
|                     ensure_sequence, | ||||
|                     stream, | ||||
|                     sequence.copy(), | ||||
|  | @ -253,7 +259,7 @@ def test_faster_task_to_recv_is_cancelled_by_slower( | |||
|                         continue | ||||
| 
 | ||||
|                 print('cancelling faster subtask') | ||||
|                 n.cancel_scope.cancel() | ||||
|                 tn.cancel_scope.cancel() | ||||
| 
 | ||||
|             try: | ||||
|                 value = await stream.receive() | ||||
|  | @ -271,7 +277,7 @@ def test_faster_task_to_recv_is_cancelled_by_slower( | |||
|                         # the faster subtask was cancelled | ||||
|                         break | ||||
| 
 | ||||
|                 # await tractor.breakpoint() | ||||
|                 # await tractor.pause() | ||||
|                 # await stream.receive() | ||||
|                 print(f'final value: {value}') | ||||
| 
 | ||||
|  | @ -371,13 +377,13 @@ def test_ensure_slow_consumers_lag_out( | |||
|                                     f'on {lags}:{value}') | ||||
|                                 return | ||||
| 
 | ||||
|             async with trio.open_nursery() as nursery: | ||||
|             async with trio.open_nursery() as tn: | ||||
| 
 | ||||
|                 for i in range(1, num_laggers): | ||||
| 
 | ||||
|                     task_name = f'sub_{i}' | ||||
|                     laggers[task_name] = 0 | ||||
|                     nursery.start_soon( | ||||
|                     tn.start_soon( | ||||
|                         partial( | ||||
|                             sub_and_print, | ||||
|                             delay=i*0.001, | ||||
|  | @ -497,6 +503,7 @@ def test_no_raise_on_lag(): | |||
|                 # internals when the no raise flag is set. | ||||
|                 loglevel='warning', | ||||
|             ), | ||||
|             collapse_eg(), | ||||
|             trio.open_nursery() as n, | ||||
|         ): | ||||
|             n.start_soon(slow) | ||||
|  |  | |||
|  | @ -3,6 +3,10 @@ Reminders for oddities in `trio` that we need to stay aware of and/or | |||
| want to see changed. | ||||
| 
 | ||||
| ''' | ||||
| from contextlib import ( | ||||
|     asynccontextmanager as acm, | ||||
| ) | ||||
| 
 | ||||
| import pytest | ||||
| import trio | ||||
| from trio import TaskStatus | ||||
|  | @ -60,7 +64,9 @@ def test_stashed_child_nursery(use_start_soon): | |||
|     async def main(): | ||||
| 
 | ||||
|         async with ( | ||||
|             trio.open_nursery() as pn, | ||||
|             trio.open_nursery( | ||||
|                 strict_exception_groups=False, | ||||
|             ) as pn, | ||||
|         ): | ||||
|             cn = await pn.start(mk_child_nursery) | ||||
|             assert cn | ||||
|  | @ -80,3 +86,118 @@ def test_stashed_child_nursery(use_start_soon): | |||
| 
 | ||||
|     with pytest.raises(NameError): | ||||
|         trio.run(main) | ||||
| 
 | ||||
| 
 | ||||
| @pytest.mark.parametrize( | ||||
|     ('unmask_from_canc', 'canc_from_finally'), | ||||
|     [ | ||||
|         (True, False), | ||||
|         (True, True), | ||||
|         pytest.param(False, True, | ||||
|                      marks=pytest.mark.xfail(reason="never raises!") | ||||
|         ), | ||||
|     ], | ||||
|     # TODO, ask ronny how to impl this .. XD | ||||
|     # ids='unmask_from_canc={0}, canc_from_finally={1}',#.format, | ||||
| ) | ||||
| def test_acm_embedded_nursery_propagates_enter_err( | ||||
|     canc_from_finally: bool, | ||||
|     unmask_from_canc: bool, | ||||
|     debug_mode: bool, | ||||
| ): | ||||
|     ''' | ||||
|     Demo how a masking `trio.Cancelled` could be handled by unmasking from the | ||||
|     `.__context__` field when a user (by accident) re-raises from a `finally:`. | ||||
| 
 | ||||
|     ''' | ||||
|     import tractor | ||||
| 
 | ||||
|     @acm | ||||
|     async def maybe_raise_from_masking_exc( | ||||
|         tn: trio.Nursery, | ||||
|         unmask_from: BaseException|None = trio.Cancelled | ||||
| 
 | ||||
|         # TODO, maybe offer a collection? | ||||
|         # unmask_from: set[BaseException] = { | ||||
|         #     trio.Cancelled, | ||||
|         # }, | ||||
|     ): | ||||
|         if not unmask_from: | ||||
|             yield | ||||
|             return | ||||
| 
 | ||||
|         try: | ||||
|             yield | ||||
|         except* unmask_from as be_eg: | ||||
| 
 | ||||
|             # TODO, if we offer `unmask_from: set` | ||||
|             # for masker_exc_type in unmask_from: | ||||
| 
 | ||||
|             matches, rest = be_eg.split(unmask_from) | ||||
|             if not matches: | ||||
|                 raise | ||||
| 
 | ||||
|             for exc_match in be_eg.exceptions: | ||||
|                 if ( | ||||
|                     (exc_ctx := exc_match.__context__) | ||||
|                     and | ||||
|                     type(exc_ctx) not in { | ||||
|                         # trio.Cancelled,  # always by default? | ||||
|                         unmask_from, | ||||
|                     } | ||||
|                 ): | ||||
|                     exc_ctx.add_note( | ||||
|                         f'\n' | ||||
|                         f'WARNING: the above error was masked by a {unmask_from!r} !?!\n' | ||||
|                         f'Are you always cancelling? Say from a `finally:` ?\n\n' | ||||
| 
 | ||||
|                         f'{tn!r}' | ||||
|                     ) | ||||
|                     raise exc_ctx from exc_match | ||||
| 
 | ||||
| 
 | ||||
|     @acm | ||||
|     async def wraps_tn_that_always_cancels(): | ||||
|         async with ( | ||||
|             trio.open_nursery() as tn, | ||||
|             maybe_raise_from_masking_exc( | ||||
|                 tn=tn, | ||||
|                 unmask_from=( | ||||
|                     trio.Cancelled | ||||
|                     if unmask_from_canc | ||||
|                     else None | ||||
|                 ), | ||||
|             ) | ||||
|         ): | ||||
|             try: | ||||
|                 yield tn | ||||
|             finally: | ||||
|                 if canc_from_finally: | ||||
|                     tn.cancel_scope.cancel() | ||||
|                     await trio.lowlevel.checkpoint() | ||||
| 
 | ||||
|     async def _main(): | ||||
|         with tractor.devx.maybe_open_crash_handler( | ||||
|             pdb=debug_mode, | ||||
|         ) as bxerr: | ||||
|             assert not bxerr.value | ||||
| 
 | ||||
|             async with ( | ||||
|                 wraps_tn_that_always_cancels() as tn, | ||||
|             ): | ||||
|                 assert not tn.cancel_scope.cancel_called | ||||
|                 assert 0 | ||||
| 
 | ||||
|         assert ( | ||||
|             (err := bxerr.value) | ||||
|             and | ||||
|             type(err) is AssertionError | ||||
|         ) | ||||
| 
 | ||||
|     with pytest.raises(ExceptionGroup) as excinfo: | ||||
|         trio.run(_main) | ||||
| 
 | ||||
|     eg: ExceptionGroup = excinfo.value | ||||
|     assert_eg, rest_eg = eg.split(AssertionError) | ||||
| 
 | ||||
|     assert len(assert_eg.exceptions) == 1 | ||||
|  |  | |||
|  | @ -44,6 +44,7 @@ from ._state import ( | |||
|     current_actor as current_actor, | ||||
|     is_root_process as is_root_process, | ||||
|     current_ipc_ctx as current_ipc_ctx, | ||||
|     debug_mode as debug_mode | ||||
| ) | ||||
| from ._exceptions import ( | ||||
|     ContextCancelled as ContextCancelled, | ||||
|  | @ -66,3 +67,4 @@ from ._root import ( | |||
| from ._ipc import Channel as Channel | ||||
| from ._portal import Portal as Portal | ||||
| from ._runtime import Actor as Actor | ||||
| # from . import hilevel as hilevel | ||||
|  |  | |||
|  | @ -19,10 +19,13 @@ Actor cluster helpers. | |||
| 
 | ||||
| ''' | ||||
| from __future__ import annotations | ||||
| 
 | ||||
| from contextlib import asynccontextmanager as acm | ||||
| from contextlib import ( | ||||
|     asynccontextmanager as acm, | ||||
| ) | ||||
| from multiprocessing import cpu_count | ||||
| from typing import AsyncGenerator, Optional | ||||
| from typing import ( | ||||
|     AsyncGenerator, | ||||
| ) | ||||
| 
 | ||||
| import trio | ||||
| import tractor | ||||
|  |  | |||
|  | @ -47,6 +47,9 @@ from functools import partial | |||
| import inspect | ||||
| from pprint import pformat | ||||
| import textwrap | ||||
| from types import ( | ||||
|     UnionType, | ||||
| ) | ||||
| from typing import ( | ||||
|     Any, | ||||
|     AsyncGenerator, | ||||
|  | @ -79,6 +82,7 @@ from .msg import ( | |||
|     MsgType, | ||||
|     NamespacePath, | ||||
|     PayloadT, | ||||
|     Return, | ||||
|     Started, | ||||
|     Stop, | ||||
|     Yield, | ||||
|  | @ -242,11 +246,13 @@ class Context: | |||
|     # a drain loop? | ||||
|     # _res_scope: trio.CancelScope|None = None | ||||
| 
 | ||||
|     _outcome_msg: Return|Error|ContextCancelled = Unresolved | ||||
| 
 | ||||
|     # on a clean exit there should be a final value | ||||
|     # delivered from the far end "callee" task, so | ||||
|     # this value is only set on one side. | ||||
|     # _result: Any | int = None | ||||
|     _result: Any|Unresolved = Unresolved | ||||
|     _result: PayloadT|Unresolved = Unresolved | ||||
| 
 | ||||
|     # if the local "caller"  task errors this value is always set | ||||
|     # to the error that was captured in the | ||||
|  | @ -950,7 +956,7 @@ class Context: | |||
|             # f'Context.cancel() => {self.chan.uid}\n' | ||||
|             f'c)=> {self.chan.uid}\n' | ||||
|             # f'{self.chan.uid}\n' | ||||
|             f' |_ @{self.dst_maddr}\n' | ||||
|             f'  |_ @{self.dst_maddr}\n' | ||||
|             f'    >> {self.repr_rpc}\n' | ||||
|             # f'    >> {self._nsf}() -> {codec}[dict]:\n\n' | ||||
|             # TODO: pull msg-type from spec re #320 | ||||
|  | @ -1003,7 +1009,8 @@ class Context: | |||
|                     ) | ||||
|                 else: | ||||
|                     log.cancel( | ||||
|                         'Timed out on cancel request of remote task?\n' | ||||
|                         f'Timed out on cancel request of remote task?\n' | ||||
|                         f'\n' | ||||
|                         f'{reminfo}' | ||||
|                     ) | ||||
| 
 | ||||
|  | @ -1195,9 +1202,11 @@ class Context: | |||
| 
 | ||||
|         ''' | ||||
|         __tracebackhide__: bool = hide_tb | ||||
|         assert self._portal, ( | ||||
|             '`Context.wait_for_result()` can not be called from callee side!' | ||||
|         ) | ||||
|         if not self._portal: | ||||
|             raise RuntimeError( | ||||
|                 'Invalid usage of `Context.wait_for_result()`!\n' | ||||
|                 'Not valid on child-side IPC ctx!\n' | ||||
|             ) | ||||
|         if self._final_result_is_set(): | ||||
|             return self._result | ||||
| 
 | ||||
|  | @ -1218,6 +1227,8 @@ class Context: | |||
|             # since every message should be delivered via the normal | ||||
|             # `._deliver_msg()` route which will appropriately set | ||||
|             # any `.maybe_error`. | ||||
|             outcome_msg: Return|Error|ContextCancelled | ||||
|             drained_msgs: list[MsgType] | ||||
|             ( | ||||
|                 outcome_msg, | ||||
|                 drained_msgs, | ||||
|  | @ -1225,11 +1236,19 @@ class Context: | |||
|                 ctx=self, | ||||
|                 hide_tb=hide_tb, | ||||
|             ) | ||||
| 
 | ||||
|             drained_status: str = ( | ||||
|                 'Ctx drained to final outcome msg\n\n' | ||||
|                 f'{outcome_msg}\n' | ||||
|             ) | ||||
| 
 | ||||
|             # ?XXX, should already be set in `._deliver_msg()` right? | ||||
|             if self._outcome_msg is not Unresolved: | ||||
|                 # from .devx import _debug | ||||
|                 # await _debug.pause() | ||||
|                 assert self._outcome_msg is outcome_msg | ||||
|             else: | ||||
|                 self._outcome_msg = outcome_msg | ||||
| 
 | ||||
|             if drained_msgs: | ||||
|                 drained_status += ( | ||||
|                     '\n' | ||||
|  | @ -1560,12 +1579,12 @@ class Context: | |||
|                     strict_pld_parity=strict_pld_parity, | ||||
|                     hide_tb=hide_tb, | ||||
|                 ) | ||||
|             except BaseException as err: | ||||
|             except BaseException as _bexc: | ||||
|                 err = _bexc | ||||
|                 if not isinstance(err, MsgTypeError): | ||||
|                     __tracebackhide__: bool = False | ||||
| 
 | ||||
|                 raise | ||||
| 
 | ||||
|                 raise err | ||||
| 
 | ||||
|         # TODO: maybe a flag to by-pass encode op if already done | ||||
|         # here in caller? | ||||
|  | @ -1703,15 +1722,28 @@ class Context: | |||
|         # TODO: expose as mod func instead! | ||||
|         structfmt = pretty_struct.Struct.pformat | ||||
|         if self._in_overrun: | ||||
|             log.warning( | ||||
|                 f'Queueing OVERRUN msg on caller task:\n\n' | ||||
| 
 | ||||
|             report: str = ( | ||||
|                 f'{flow_body}' | ||||
| 
 | ||||
|                 f'{structfmt(msg)}\n' | ||||
|             ) | ||||
|             over_q: deque = self._overflow_q | ||||
|             self._overflow_q.append(msg) | ||||
| 
 | ||||
|             if len(over_q) == over_q.maxlen: | ||||
|                 report = ( | ||||
|                     'FAILED to queue OVERRUN msg, OVERAN the OVERRUN QUEUE !!\n\n' | ||||
|                     + report | ||||
|                 ) | ||||
|                 # log.error(report) | ||||
|                 log.debug(report) | ||||
| 
 | ||||
|             else: | ||||
|                 report = ( | ||||
|                     'Queueing OVERRUN msg on caller task:\n\n' | ||||
|                     + report | ||||
|                 ) | ||||
|                 log.debug(report) | ||||
| 
 | ||||
|             # XXX NOTE XXX | ||||
|             # overrun is the ONLY case where returning early is fine! | ||||
|             return False | ||||
|  | @ -1724,7 +1756,6 @@ class Context: | |||
| 
 | ||||
|                 f'{structfmt(msg)}\n' | ||||
|             ) | ||||
| 
 | ||||
|             # NOTE: if an error is deteced we should always still | ||||
|             # send it through the feeder-mem-chan and expect | ||||
|             # it to be raised by any context (stream) consumer | ||||
|  | @ -1736,6 +1767,21 @@ class Context: | |||
|             # normally the task that should get cancelled/error | ||||
|             # from some remote fault! | ||||
|             send_chan.send_nowait(msg) | ||||
|             match msg: | ||||
|                 case Stop(): | ||||
|                     if (stream := self._stream): | ||||
|                         stream._stop_msg = msg | ||||
| 
 | ||||
|                 case Return(): | ||||
|                     if not self._outcome_msg: | ||||
|                         log.warning( | ||||
|                             f'Setting final outcome msg AFTER ' | ||||
|                             f'`._rx_chan.send()`??\n' | ||||
|                             f'\n' | ||||
|                             f'{msg}' | ||||
|                         ) | ||||
|                         self._outcome_msg = msg | ||||
| 
 | ||||
|             return True | ||||
| 
 | ||||
|         except trio.BrokenResourceError: | ||||
|  | @ -1969,7 +2015,10 @@ async def open_context_from_portal( | |||
|     ctxc_from_callee: ContextCancelled|None = None | ||||
|     try: | ||||
|         async with ( | ||||
|             trio.open_nursery() as tn, | ||||
|             trio.open_nursery( | ||||
|                 strict_exception_groups=False, | ||||
|             ) as tn, | ||||
| 
 | ||||
|             msgops.maybe_limit_plds( | ||||
|                 ctx=ctx, | ||||
|                 spec=ctx_meta.get('pld_spec'), | ||||
|  | @ -1989,7 +2038,7 @@ async def open_context_from_portal( | |||
|             # the dialog, the `Error` msg should be raised from the `msg` | ||||
|             # handling block below. | ||||
|             try: | ||||
|                 started_msg, first = await ctx._pld_rx.recv_msg_w_pld( | ||||
|                 started_msg, first = await ctx._pld_rx.recv_msg( | ||||
|                     ipc=ctx, | ||||
|                     expect_msg=Started, | ||||
|                     passthrough_non_pld_msgs=False, | ||||
|  | @ -2354,7 +2403,8 @@ async def open_context_from_portal( | |||
|             # displaying `ContextCancelled` traces where the | ||||
|             # cause of crash/exit IS due to something in | ||||
|             # user/app code on either end of the context. | ||||
|             and not rxchan._closed | ||||
|             and | ||||
|             not rxchan._closed | ||||
|         ): | ||||
|             # XXX NOTE XXX: and again as per above, we mask any | ||||
|             # `trio.Cancelled` raised here so as to NOT mask | ||||
|  | @ -2413,6 +2463,7 @@ async def open_context_from_portal( | |||
|         # FINALLY, remove the context from runtime tracking and | ||||
|         # exit! | ||||
|         log.runtime( | ||||
|         # log.cancel( | ||||
|             f'De-allocating IPC ctx opened with {ctx.side!r} peer \n' | ||||
|             f'uid: {uid}\n' | ||||
|             f'cid: {ctx.cid}\n' | ||||
|  | @ -2468,7 +2519,6 @@ def mk_context( | |||
|         _caller_info=caller_info, | ||||
|         **kwargs, | ||||
|     ) | ||||
|     pld_rx._ctx = ctx | ||||
|     ctx._result = Unresolved | ||||
|     return ctx | ||||
| 
 | ||||
|  | @ -2531,7 +2581,14 @@ def context( | |||
|     name: str | ||||
|     param: Type | ||||
|     for name, param in annots.items(): | ||||
|         if param is Context: | ||||
|         if ( | ||||
|             param is Context | ||||
|             or ( | ||||
|                 isinstance(param, UnionType) | ||||
|                 and | ||||
|                 Context in param.__args__ | ||||
|             ) | ||||
|         ): | ||||
|             ctx_var_name: str = name | ||||
|             break | ||||
|     else: | ||||
|  |  | |||
|  | @ -20,6 +20,7 @@ Sub-process entry points. | |||
| """ | ||||
| from __future__ import annotations | ||||
| from functools import partial | ||||
| import multiprocessing as mp | ||||
| import os | ||||
| import textwrap | ||||
| from typing import ( | ||||
|  | @ -64,20 +65,22 @@ def _mp_main( | |||
|     ''' | ||||
|     actor._forkserver_info = forkserver_info | ||||
|     from ._spawn import try_set_start_method | ||||
|     spawn_ctx = try_set_start_method(start_method) | ||||
|     spawn_ctx: mp.context.BaseContext = try_set_start_method(start_method) | ||||
|     assert spawn_ctx | ||||
| 
 | ||||
|     if actor.loglevel is not None: | ||||
|         log.info( | ||||
|             f"Setting loglevel for {actor.uid} to {actor.loglevel}") | ||||
|             f'Setting loglevel for {actor.uid} to {actor.loglevel}' | ||||
|         ) | ||||
|         get_console_log(actor.loglevel) | ||||
| 
 | ||||
|     assert spawn_ctx | ||||
|     # TODO: use scops headers like for `trio` below! | ||||
|     # (well after we libify it maybe..) | ||||
|     log.info( | ||||
|         f"Started new {spawn_ctx.current_process()} for {actor.uid}") | ||||
| 
 | ||||
|     _state._current_actor = actor | ||||
| 
 | ||||
|     log.debug(f"parent_addr is {parent_addr}") | ||||
|         f'Started new {spawn_ctx.current_process()} for {actor.uid}' | ||||
|     #     f"parent_addr is {parent_addr}" | ||||
|     ) | ||||
|     _state._current_actor: Actor = actor | ||||
|     trio_main = partial( | ||||
|         async_main, | ||||
|         actor=actor, | ||||
|  | @ -94,7 +97,9 @@ def _mp_main( | |||
|         pass  # handle it the same way trio does? | ||||
| 
 | ||||
|     finally: | ||||
|         log.info(f"Subactor {actor.uid} terminated") | ||||
|         log.info( | ||||
|             f'`mp`-subactor {actor.uid} exited' | ||||
|         ) | ||||
| 
 | ||||
| 
 | ||||
| # TODO: move this func to some kinda `.devx._conc_lang.py` eventually | ||||
|  | @ -233,7 +238,7 @@ def _trio_main( | |||
|             nest_from_op( | ||||
|                 input_op='>(',  # see syntax ideas above | ||||
|                 tree_str=actor_info, | ||||
|                 back_from_op=1, | ||||
|                 back_from_op=2,  # since "complete" | ||||
|             ) | ||||
|         ) | ||||
|     logmeth = log.info | ||||
|  |  | |||
|  | @ -22,6 +22,7 @@ from __future__ import annotations | |||
| import builtins | ||||
| import importlib | ||||
| from pprint import pformat | ||||
| from pdb import bdb | ||||
| import sys | ||||
| from types import ( | ||||
|     TracebackType, | ||||
|  | @ -82,6 +83,48 @@ class InternalError(RuntimeError): | |||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| class AsyncioCancelled(Exception): | ||||
|     ''' | ||||
|     Asyncio cancelled translation (non-base) error | ||||
|     for use with the ``to_asyncio`` module | ||||
|     to be raised in the ``trio`` side task | ||||
| 
 | ||||
|     NOTE: this should NOT inherit from `asyncio.CancelledError` or | ||||
|     tests should break! | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| 
 | ||||
| class AsyncioTaskExited(Exception): | ||||
|     ''' | ||||
|     asyncio.Task "exited" translation error for use with the | ||||
|     `to_asyncio` APIs to be raised in the `trio` side task indicating | ||||
|     on `.run_task()`/`.open_channel_from()` exit that the aio side | ||||
|     exited early/silently. | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| class TrioCancelled(Exception): | ||||
|     ''' | ||||
|     Trio cancelled translation (non-base) error | ||||
|     for use with the `to_asyncio` module | ||||
|     to be raised in the `asyncio.Task` to indicate | ||||
|     that the `trio` side raised `Cancelled` or an error. | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| class TrioTaskExited(Exception): | ||||
|     ''' | ||||
|     The `trio`-side task exited without explicitly cancelling the | ||||
|     `asyncio.Task` peer. | ||||
| 
 | ||||
|     This is very similar to how `trio.ClosedResource` acts as | ||||
|     a "clean shutdown" signal to the consumer side of a mem-chan, | ||||
| 
 | ||||
|     https://trio.readthedocs.io/en/stable/reference-core.html#clean-shutdown-with-channels | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| 
 | ||||
| # NOTE: more or less should be close to these: | ||||
| # 'boxed_type', | ||||
|  | @ -127,8 +170,8 @@ _body_fields: list[str] = list( | |||
| 
 | ||||
| def get_err_type(type_name: str) -> BaseException|None: | ||||
|     ''' | ||||
|     Look up an exception type by name from the set of locally | ||||
|     known namespaces: | ||||
|     Look up an exception type by name from the set of locally known | ||||
|     namespaces: | ||||
| 
 | ||||
|     - `builtins` | ||||
|     - `tractor._exceptions` | ||||
|  | @ -139,6 +182,7 @@ def get_err_type(type_name: str) -> BaseException|None: | |||
|         builtins, | ||||
|         _this_mod, | ||||
|         trio, | ||||
|         bdb, | ||||
|     ]: | ||||
|         if type_ref := getattr( | ||||
|             ns, | ||||
|  | @ -358,6 +402,13 @@ class RemoteActorError(Exception): | |||
|                 self._ipc_msg.src_type_str | ||||
|             ) | ||||
| 
 | ||||
|             if not self._src_type: | ||||
|                 raise TypeError( | ||||
|                     f'Failed to lookup src error type with ' | ||||
|                     f'`tractor._exceptions.get_err_type()` :\n' | ||||
|                     f'{self.src_type_str}' | ||||
|                 ) | ||||
| 
 | ||||
|         return self._src_type | ||||
| 
 | ||||
|     @property | ||||
|  | @ -366,6 +417,9 @@ class RemoteActorError(Exception): | |||
|         String-name of the (last hop's) boxed error type. | ||||
| 
 | ||||
|         ''' | ||||
|         # TODO, maybe support also serializing the | ||||
|         # `ExceptionGroup.exeptions: list[BaseException]` set under | ||||
|         # certain conditions? | ||||
|         bt: Type[BaseException] = self.boxed_type | ||||
|         if bt: | ||||
|             return str(bt.__name__) | ||||
|  | @ -378,9 +432,13 @@ class RemoteActorError(Exception): | |||
|         Error type boxed by last actor IPC hop. | ||||
| 
 | ||||
|         ''' | ||||
|         if self._boxed_type is None: | ||||
|         if ( | ||||
|             self._boxed_type is None | ||||
|             and | ||||
|             (ipc_msg := self._ipc_msg) | ||||
|         ): | ||||
|             self._boxed_type = get_err_type( | ||||
|                 self._ipc_msg.boxed_type_str | ||||
|                 ipc_msg.boxed_type_str | ||||
|             ) | ||||
| 
 | ||||
|         return self._boxed_type | ||||
|  | @ -609,6 +667,7 @@ class RemoteActorError(Exception): | |||
|                 # just after <Type( | ||||
|                 #             |___ .. | ||||
|                 tb_body_indent=1, | ||||
|                 boxer_header=self.relay_uid, | ||||
|             ) | ||||
| 
 | ||||
|         tail = '' | ||||
|  | @ -651,16 +710,10 @@ class RemoteActorError(Exception): | |||
|         failing actor's remote env. | ||||
| 
 | ||||
|         ''' | ||||
|         src_type_ref: Type[BaseException] = self.src_type | ||||
|         if not src_type_ref: | ||||
|             raise TypeError( | ||||
|                 'Failed to lookup src error type:\n' | ||||
|                 f'{self.src_type_str}' | ||||
|             ) | ||||
| 
 | ||||
|         # TODO: better tb insertion and all the fancier dunder | ||||
|         # metadata stuff as per `.__context__` etc. and friends: | ||||
|         # https://github.com/python-trio/trio/issues/611 | ||||
|         src_type_ref: Type[BaseException] = self.src_type | ||||
|         return src_type_ref(self.tb_str) | ||||
| 
 | ||||
|     # TODO: local recontruction of nested inception for a given | ||||
|  | @ -786,8 +839,11 @@ class MsgTypeError( | |||
|         ''' | ||||
|         if ( | ||||
|             (_bad_msg := self.msgdata.get('_bad_msg')) | ||||
|             and | ||||
|             isinstance(_bad_msg, PayloadMsg) | ||||
|             and ( | ||||
|                 isinstance(_bad_msg, PayloadMsg) | ||||
|                 or | ||||
|                 isinstance(_bad_msg, msgtypes.Start) | ||||
|             ) | ||||
|         ): | ||||
|             return _bad_msg | ||||
| 
 | ||||
|  | @ -973,15 +1029,6 @@ class NoRuntime(RuntimeError): | |||
|     "The root actor has not been initialized yet" | ||||
| 
 | ||||
| 
 | ||||
| 
 | ||||
| class AsyncioCancelled(Exception): | ||||
|     ''' | ||||
|     Asyncio cancelled translation (non-base) error | ||||
|     for use with the ``to_asyncio`` module | ||||
|     to be raised in the ``trio`` side task | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| class MessagingError(Exception): | ||||
|     ''' | ||||
|     IPC related msg (typing), transaction (ordering) or dialog | ||||
|  | @ -989,7 +1036,6 @@ class MessagingError(Exception): | |||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
| 
 | ||||
| def pack_error( | ||||
|     exc: BaseException|RemoteActorError, | ||||
| 
 | ||||
|  | @ -1101,6 +1147,8 @@ def unpack_error( | |||
|     which is the responsibilitiy of the caller. | ||||
| 
 | ||||
|     ''' | ||||
|     # XXX, apparently we pass all sorts of msgs here? | ||||
|     # kinda odd but seems like maybe they shouldn't be? | ||||
|     if not isinstance(msg, Error): | ||||
|         return None | ||||
| 
 | ||||
|  | @ -1143,19 +1191,51 @@ def unpack_error( | |||
| 
 | ||||
| 
 | ||||
| def is_multi_cancelled( | ||||
|     exc: BaseException|BaseExceptionGroup | ||||
| ) -> bool: | ||||
|     exc: BaseException|BaseExceptionGroup, | ||||
| 
 | ||||
|     ignore_nested: set[BaseException] = set(), | ||||
| 
 | ||||
| ) -> bool|BaseExceptionGroup: | ||||
|     ''' | ||||
|     Predicate to determine if a possible ``BaseExceptionGroup`` contains | ||||
|     only ``trio.Cancelled`` sub-exceptions (and is likely the result of | ||||
|     cancelling a collection of subtasks. | ||||
|     Predicate to determine if an `BaseExceptionGroup` only contains | ||||
|     some (maybe nested) set of sub-grouped exceptions (like only | ||||
|     `trio.Cancelled`s which get swallowed silently by default) and is | ||||
|     thus the result of "gracefully cancelling" a collection of | ||||
|     sub-tasks (or other conc primitives) and receiving a "cancelled | ||||
|     ACK" from each after termination. | ||||
| 
 | ||||
|     Docs: | ||||
|     ---- | ||||
|     - https://docs.python.org/3/library/exceptions.html#exception-groups | ||||
|     - https://docs.python.org/3/library/exceptions.html#BaseExceptionGroup.subgroup | ||||
| 
 | ||||
|     ''' | ||||
| 
 | ||||
|     if ( | ||||
|         not ignore_nested | ||||
|         or | ||||
|         trio.Cancelled in ignore_nested | ||||
|         # XXX always count-in `trio`'s native signal | ||||
|     ): | ||||
|         ignore_nested.update({trio.Cancelled}) | ||||
| 
 | ||||
|     if isinstance(exc, BaseExceptionGroup): | ||||
|         return exc.subgroup( | ||||
|             lambda exc: isinstance(exc, trio.Cancelled) | ||||
|         ) is not None | ||||
|         matched_exc: BaseExceptionGroup|None = exc.subgroup( | ||||
|             tuple(ignore_nested), | ||||
| 
 | ||||
|             # TODO, complain about why not allowed XD | ||||
|             # condition=tuple(ignore_nested), | ||||
|         ) | ||||
|         if matched_exc is not None: | ||||
|             return matched_exc | ||||
| 
 | ||||
|     # NOTE, IFF no excs types match (throughout the error-tree) | ||||
|     # -> return `False`, OW return the matched sub-eg. | ||||
|     # | ||||
|     # IOW, for the inverse of ^ for the purpose of | ||||
|     # maybe-enter-REPL--logic: "only debug when the err-tree contains | ||||
|     # at least one exc-type NOT in `ignore_nested`" ; i.e. the case where | ||||
|     # we fallthrough and return `False` here. | ||||
|     return False | ||||
| 
 | ||||
| 
 | ||||
|  | @ -1375,7 +1455,9 @@ def _mk_recv_mte( | |||
|         any_pld: Any = msgpack.decode(msg.pld) | ||||
|         message: str = ( | ||||
|             f'invalid `{msg_type.__qualname__}` msg payload\n\n' | ||||
|             f'value: `{any_pld!r}` does not match type-spec: ' | ||||
|             f'{any_pld!r}\n\n' | ||||
|             f'has type {type(any_pld)!r}\n\n' | ||||
|             f'and does not match type-spec ' | ||||
|             f'`{type(msg).__qualname__}.pld: {codec.pld_spec_str}`' | ||||
|         ) | ||||
|         bad_msg = msg | ||||
|  |  | |||
|  | @ -255,8 +255,8 @@ class MsgpackTCPStream(MsgTransport): | |||
|                 raise TransportClosed( | ||||
|                     message=( | ||||
|                         f'IPC transport already closed by peer\n' | ||||
|                         f'x)> {type(trans_err)}\n' | ||||
|                         f' |_{self}\n' | ||||
|                         f'x]> {type(trans_err)}\n' | ||||
|                         f'  |_{self}\n' | ||||
|                     ), | ||||
|                     loglevel=loglevel, | ||||
|                 ) from trans_err | ||||
|  | @ -273,8 +273,8 @@ class MsgpackTCPStream(MsgTransport): | |||
|                 raise TransportClosed( | ||||
|                     message=( | ||||
|                         f'IPC transport already manually closed locally?\n' | ||||
|                         f'x)> {type(closure_err)} \n' | ||||
|                         f' |_{self}\n' | ||||
|                         f'x]> {type(closure_err)} \n' | ||||
|                         f'  |_{self}\n' | ||||
|                     ), | ||||
|                     loglevel='error', | ||||
|                     raise_on_report=( | ||||
|  | @ -289,8 +289,8 @@ class MsgpackTCPStream(MsgTransport): | |||
|                 raise TransportClosed( | ||||
|                     message=( | ||||
|                         f'IPC transport already gracefully closed\n' | ||||
|                         f')>\n' | ||||
|                         f'|_{self}\n' | ||||
|                         f']>\n' | ||||
|                         f' |_{self}\n' | ||||
|                     ), | ||||
|                     loglevel='transport', | ||||
|                     # cause=???  # handy or no? | ||||
|  |  | |||
|  | @ -184,7 +184,7 @@ class Portal: | |||
|                 ( | ||||
|                     self._final_result_msg, | ||||
|                     self._final_result_pld, | ||||
|                 ) = await self._expect_result_ctx._pld_rx.recv_msg_w_pld( | ||||
|                 ) = await self._expect_result_ctx._pld_rx.recv_msg( | ||||
|                     ipc=self._expect_result_ctx, | ||||
|                     expect_msg=Return, | ||||
|                 ) | ||||
|  | @ -533,6 +533,10 @@ async def open_portal( | |||
|     async with maybe_open_nursery( | ||||
|         tn, | ||||
|         shield=shield, | ||||
|         strict_exception_groups=False, | ||||
|         # ^XXX^ TODO? soo roll our own then ?? | ||||
|         # -> since we kinda want the "if only one `.exception` then | ||||
|         # just raise that" interface? | ||||
|     ) as tn: | ||||
| 
 | ||||
|         if not channel.connected(): | ||||
|  |  | |||
|  | @ -80,7 +80,7 @@ async def open_root_actor( | |||
| 
 | ||||
|     # enables the multi-process debugger support | ||||
|     debug_mode: bool = False, | ||||
|     maybe_enable_greenback: bool = False,  # `.pause_from_sync()/breakpoint()` support | ||||
|     maybe_enable_greenback: bool = True,  # `.pause_from_sync()/breakpoint()` support | ||||
|     enable_stack_on_sig: bool = False, | ||||
| 
 | ||||
|     # internal logging | ||||
|  | @ -95,13 +95,24 @@ async def open_root_actor( | |||
| 
 | ||||
|     hide_tb: bool = True, | ||||
| 
 | ||||
|     # XXX, proxied directly to `.devx._debug._maybe_enter_pm()` | ||||
|     # for REPL-entry logic. | ||||
|     debug_filter: Callable[ | ||||
|         [BaseException|BaseExceptionGroup], | ||||
|         bool, | ||||
|     ] = lambda err: not is_multi_cancelled(err), | ||||
| 
 | ||||
|     # TODO, a way for actors to augment passing derived | ||||
|     # read-only state to sublayers? | ||||
|     # extra_rt_vars: dict|None = None, | ||||
| 
 | ||||
| ) -> Actor: | ||||
|     ''' | ||||
|     Runtime init entry point for ``tractor``. | ||||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = hide_tb | ||||
|     _debug.hide_runtime_frames() | ||||
|     __tracebackhide__: bool = hide_tb | ||||
| 
 | ||||
|     # TODO: stick this in a `@cm` defined in `devx._debug`? | ||||
|     # | ||||
|  | @ -233,14 +244,8 @@ async def open_root_actor( | |||
|         and | ||||
|         enable_stack_on_sig | ||||
|     ): | ||||
|         try: | ||||
|             logger.info('Enabling `stackscope` traces on SIGUSR1') | ||||
|             from .devx import enable_stack_on_sig | ||||
|             enable_stack_on_sig() | ||||
|         except ImportError: | ||||
|             logger.warning( | ||||
|                 '`stackscope` not installed for use in debug mode!' | ||||
|             ) | ||||
|         from .devx._stackscope import enable_stack_on_sig | ||||
|         enable_stack_on_sig() | ||||
| 
 | ||||
|     # closed into below ping task-func | ||||
|     ponged_addrs: list[tuple[str, int]] = [] | ||||
|  | @ -336,6 +341,10 @@ async def open_root_actor( | |||
|             loglevel=loglevel, | ||||
|             enable_modules=enable_modules, | ||||
|         ) | ||||
|         # XXX, in case the root actor runtime was actually run from | ||||
|         # `tractor.to_asyncio.run_as_asyncio_guest()` and NOt | ||||
|         # `.trio.run()`. | ||||
|         actor._infected_aio = _state._runtime_vars['_is_infected_aio'] | ||||
| 
 | ||||
|     # Start up main task set via core actor-runtime nurseries. | ||||
|     try: | ||||
|  | @ -353,7 +362,10 @@ async def open_root_actor( | |||
|         ) | ||||
| 
 | ||||
|         # start the actor runtime in a new task | ||||
|         async with trio.open_nursery() as nursery: | ||||
|         async with trio.open_nursery( | ||||
|             strict_exception_groups=False, | ||||
|             # ^XXX^ TODO? instead unpack any RAE as per "loose" style? | ||||
|         ) as nursery: | ||||
| 
 | ||||
|             # ``_runtime.async_main()`` creates an internal nursery | ||||
|             # and blocks here until any underlying actor(-process) | ||||
|  | @ -377,6 +389,13 @@ async def open_root_actor( | |||
|                 Exception, | ||||
|                 BaseExceptionGroup, | ||||
|             ) as err: | ||||
| 
 | ||||
|                 # TODO, in beginning to handle the subsubactor with | ||||
|                 # crashed grandparent cases.. | ||||
|                 # | ||||
|                 # was_locked: bool = await _debug.maybe_wait_for_debugger( | ||||
|                 #     child_in_debug=True, | ||||
|                 # ) | ||||
|                 # XXX NOTE XXX see equiv note inside | ||||
|                 # `._runtime.Actor._stream_handler()` where in the | ||||
|                 # non-root or root-that-opened-this-mahually case we | ||||
|  | @ -385,11 +404,15 @@ async def open_root_actor( | |||
|                 entered: bool = await _debug._maybe_enter_pm( | ||||
|                     err, | ||||
|                     api_frame=inspect.currentframe(), | ||||
|                     debug_filter=debug_filter, | ||||
|                 ) | ||||
| 
 | ||||
|                 if ( | ||||
|                     not entered | ||||
|                     and | ||||
|                     not is_multi_cancelled(err) | ||||
|                     not is_multi_cancelled( | ||||
|                         err, | ||||
|                     ) | ||||
|                 ): | ||||
|                     logger.exception('Root actor crashed\n') | ||||
| 
 | ||||
|  | @ -443,12 +466,19 @@ def run_daemon( | |||
| 
 | ||||
|     start_method: str | None = None, | ||||
|     debug_mode: bool = False, | ||||
| 
 | ||||
|     # TODO, support `infected_aio=True` mode by, | ||||
|     # - calling the appropriate entrypoint-func from `.to_asyncio` | ||||
|     # - maybe init-ing `greenback` as done above in | ||||
|     #   `open_root_actor()`. | ||||
| 
 | ||||
|     **kwargs | ||||
| 
 | ||||
| ) -> None: | ||||
|     ''' | ||||
|     Spawn daemon actor which will respond to RPC; the main task simply | ||||
|     starts the runtime and then sleeps forever. | ||||
|     Spawn a root (daemon) actor which will respond to RPC; the main | ||||
|     task simply starts the runtime and then blocks via embedded | ||||
|     `trio.sleep_forever()`. | ||||
| 
 | ||||
|     This is a very minimal convenience wrapper around starting | ||||
|     a "run-until-cancelled" root actor which can be started with a set | ||||
|  | @ -461,7 +491,6 @@ def run_daemon( | |||
|         importlib.import_module(path) | ||||
| 
 | ||||
|     async def _main(): | ||||
| 
 | ||||
|         async with open_root_actor( | ||||
|             registry_addrs=registry_addrs, | ||||
|             name=name, | ||||
|  |  | |||
|  | @ -620,7 +620,11 @@ async def _invoke( | |||
|             tn: trio.Nursery | ||||
|             rpc_ctx_cs: CancelScope | ||||
|             async with ( | ||||
|                 trio.open_nursery() as tn, | ||||
|                 trio.open_nursery( | ||||
|                     strict_exception_groups=False, | ||||
|                     # ^XXX^ TODO? instead unpack any RAE as per "loose" style? | ||||
| 
 | ||||
|                 ) as tn, | ||||
|                 msgops.maybe_limit_plds( | ||||
|                     ctx=ctx, | ||||
|                     spec=ctx_meta.get('pld_spec'), | ||||
|  | @ -645,6 +649,10 @@ async def _invoke( | |||
|                 ) | ||||
|                 # set and shuttle final result to "parent"-side task. | ||||
|                 ctx._result = res | ||||
|                 log.runtime( | ||||
|                     f'Sending result msg and exiting {ctx.side!r}\n' | ||||
|                     f'{return_msg}\n' | ||||
|                 ) | ||||
|                 await chan.send(return_msg) | ||||
| 
 | ||||
|             # NOTE: this happens IFF `ctx._scope.cancel()` is | ||||
|  | @ -733,8 +741,8 @@ async def _invoke( | |||
|         # XXX: do we ever trigger this block any more? | ||||
|         except ( | ||||
|             BaseExceptionGroup, | ||||
|             trio.Cancelled, | ||||
|             BaseException, | ||||
|             trio.Cancelled, | ||||
| 
 | ||||
|         ) as scope_error: | ||||
|             if ( | ||||
|  | @ -847,8 +855,8 @@ async def try_ship_error_to_remote( | |||
|             log.critical( | ||||
|                 'IPC transport failure -> ' | ||||
|                 f'failed to ship error to {remote_descr}!\n\n' | ||||
|                 f'X=> {channel.uid}\n\n' | ||||
| 
 | ||||
|                 f'{type(msg)!r}[{msg.boxed_type_str}] X=> {channel.uid}\n' | ||||
|                 f'\n' | ||||
|                 # TODO: use `.msg.preetty_struct` for this! | ||||
|                 f'{msg}\n' | ||||
|             ) | ||||
|  |  | |||
|  | @ -59,6 +59,7 @@ from types import ModuleType | |||
| import warnings | ||||
| 
 | ||||
| import trio | ||||
| from trio._core import _run as trio_runtime | ||||
| from trio import ( | ||||
|     CancelScope, | ||||
|     Nursery, | ||||
|  | @ -80,6 +81,7 @@ from ._context import ( | |||
| from .log import get_logger | ||||
| from ._exceptions import ( | ||||
|     ContextCancelled, | ||||
|     InternalError, | ||||
|     ModuleNotExposed, | ||||
|     MsgTypeError, | ||||
|     unpack_error, | ||||
|  | @ -98,6 +100,7 @@ from ._rpc import ( | |||
| 
 | ||||
| if TYPE_CHECKING: | ||||
|     from ._supervise import ActorNursery | ||||
|     from trio._channel import MemoryChannelState | ||||
| 
 | ||||
| 
 | ||||
| log = get_logger('tractor') | ||||
|  | @ -833,8 +836,10 @@ class Actor: | |||
|             )] | ||||
|         except KeyError: | ||||
|             report: str = ( | ||||
|                 'Ignoring invalid IPC ctx msg!\n\n' | ||||
|                 f'<=? {uid}\n\n' | ||||
|                 'Ignoring invalid IPC msg!?\n' | ||||
|                 f'Ctx seems to not/no-longer exist??\n' | ||||
|                 f'\n' | ||||
|                 f'<=? {uid}\n' | ||||
|                 f'  |_{pretty_struct.pformat(msg)}\n' | ||||
|             ) | ||||
|             match msg: | ||||
|  | @ -896,11 +901,15 @@ class Actor: | |||
|                 f'peer: {chan.uid}\n' | ||||
|                 f'cid:{cid}\n' | ||||
|             ) | ||||
|             ctx._allow_overruns = allow_overruns | ||||
|             ctx._allow_overruns: bool = allow_overruns | ||||
| 
 | ||||
|             # adjust buffer size if specified | ||||
|             state = ctx._send_chan._state  # type: ignore | ||||
|             if msg_buffer_size and state.max_buffer_size != msg_buffer_size: | ||||
|             state: MemoryChannelState  = ctx._send_chan._state  # type: ignore | ||||
|             if ( | ||||
|                 msg_buffer_size | ||||
|                 and | ||||
|                 state.max_buffer_size != msg_buffer_size | ||||
|             ): | ||||
|                 state.max_buffer_size = msg_buffer_size | ||||
| 
 | ||||
|         except KeyError: | ||||
|  | @ -1094,7 +1103,36 @@ class Actor: | |||
|                                 '`tractor.pause_from_sync()` not available!' | ||||
|                             ) | ||||
| 
 | ||||
|                 rvs['_is_root'] = False | ||||
|                 # XXX ensure the "infected `asyncio` mode" setting | ||||
|                 # passed down from our spawning parent is consistent | ||||
|                 # with `trio`-runtime initialization: | ||||
|                 # - during sub-proc boot, the entrypoint func | ||||
|                 #   (`._entry.<spawn_backend>_main()`) should set | ||||
|                 #   `._infected_aio = True` before calling | ||||
|                 #   `run_as_asyncio_guest()`, | ||||
|                 # - the value of `infect_asyncio: bool = True` as | ||||
|                 #   passed to `ActorNursery.start_actor()` must be | ||||
|                 #   the same as `_runtime_vars['_is_infected_aio']` | ||||
|                 if ( | ||||
|                     (aio_rtv := rvs['_is_infected_aio']) | ||||
|                     != | ||||
|                     (aio_attr := self._infected_aio) | ||||
|                 ): | ||||
|                     raise InternalError( | ||||
|                         'Parent sent runtime-vars that mismatch for the ' | ||||
|                         '"infected `asyncio` mode" settings ?!?\n\n' | ||||
| 
 | ||||
|                         f'rvs["_is_infected_aio"] = {aio_rtv}\n' | ||||
|                         f'self._infected_aio = {aio_attr}\n' | ||||
|                     ) | ||||
|                 if aio_rtv: | ||||
|                     assert trio_runtime.GLOBAL_RUN_CONTEXT.runner.is_guest | ||||
|                     # ^TODO^ possibly add a `sniffio` or | ||||
|                     # `trio` pub-API for `is_guest_mode()`? | ||||
| 
 | ||||
|                 rvs['_is_root'] = False  # obvi XD | ||||
| 
 | ||||
|                 # update process-wide globals | ||||
|                 _state._runtime_vars.update(rvs) | ||||
| 
 | ||||
|                 # XXX: ``msgspec`` doesn't support serializing tuples | ||||
|  | @ -1247,7 +1285,8 @@ class Actor: | |||
|         msg: str = ( | ||||
|             f'Actor-runtime cancel request from {requester_type}\n\n' | ||||
|             f'<=c) {requesting_uid}\n' | ||||
|             f' |_{self}\n' | ||||
|             f'  |_{self}\n' | ||||
|             f'\n' | ||||
|         ) | ||||
| 
 | ||||
|         # TODO: what happens here when we self-cancel tho? | ||||
|  | @ -1267,13 +1306,15 @@ class Actor: | |||
|                 lock_req_ctx.has_outcome | ||||
|             ): | ||||
|                 msg += ( | ||||
|                     '-> Cancelling active debugger request..\n' | ||||
|                     f'\n' | ||||
|                     f'-> Cancelling active debugger request..\n' | ||||
|                     f'|_{_debug.Lock.repr()}\n\n' | ||||
|                     f'|_{lock_req_ctx}\n\n' | ||||
|                 ) | ||||
|                 # lock_req_ctx._scope.cancel() | ||||
|                 # TODO: wrap this in a method-API.. | ||||
|                 debug_req.req_cs.cancel() | ||||
|                 # if lock_req_ctx: | ||||
| 
 | ||||
|             # self-cancel **all** ongoing RPC tasks | ||||
|             await self.cancel_rpc_tasks( | ||||
|  | @ -1682,11 +1723,15 @@ async def async_main( | |||
|         # parent is kept alive as a resilient service until | ||||
|         # cancellation steps have (mostly) occurred in | ||||
|         # a deterministic way. | ||||
|         async with trio.open_nursery() as root_nursery: | ||||
|         async with trio.open_nursery( | ||||
|             strict_exception_groups=False, | ||||
|         ) as root_nursery: | ||||
|             actor._root_n = root_nursery | ||||
|             assert actor._root_n | ||||
| 
 | ||||
|             async with trio.open_nursery() as service_nursery: | ||||
|             async with trio.open_nursery( | ||||
|                 strict_exception_groups=False, | ||||
|             ) as service_nursery: | ||||
|                 # This nursery is used to handle all inbound | ||||
|                 # connections to us such that if the TCP server | ||||
|                 # is killed, connections can continue to process | ||||
|  |  | |||
|  | @ -327,9 +327,10 @@ async def soft_kill( | |||
|     uid: tuple[str, str] = portal.channel.uid | ||||
|     try: | ||||
|         log.cancel( | ||||
|             'Soft killing sub-actor via portal request\n' | ||||
|             f'c)> {portal.chan.uid}\n' | ||||
|             f' |_{proc}\n' | ||||
|             f'Soft killing sub-actor via portal request\n' | ||||
|             f'\n' | ||||
|             f'(c=> {portal.chan.uid}\n' | ||||
|             f'  |_{proc}\n' | ||||
|         ) | ||||
|         # wait on sub-proc to signal termination | ||||
|         await wait_func(proc) | ||||
|  |  | |||
|  | @ -44,6 +44,8 @@ _runtime_vars: dict[str, Any] = { | |||
|     '_root_mailbox': (None, None), | ||||
|     '_registry_addrs': [], | ||||
| 
 | ||||
|     '_is_infected_aio': False, | ||||
| 
 | ||||
|     # for `tractor.pause_from_sync()` & `breakpoint()` support | ||||
|     'use_greenback': False, | ||||
| } | ||||
|  | @ -70,7 +72,8 @@ def current_actor( | |||
|     ''' | ||||
|     if ( | ||||
|         err_on_no_runtime | ||||
|         and _current_actor is None | ||||
|         and | ||||
|         _current_actor is None | ||||
|     ): | ||||
|         msg: str = 'No local actor has been initialized yet?\n' | ||||
|         from ._exceptions import NoRuntime | ||||
|  | @ -105,6 +108,7 @@ def is_main_process() -> bool: | |||
|     return mp.current_process().name == 'MainProcess' | ||||
| 
 | ||||
| 
 | ||||
| # TODO, more verby name? | ||||
| def debug_mode() -> bool: | ||||
|     ''' | ||||
|     Bool determining if "debug mode" is on which enables | ||||
|  |  | |||
|  | @ -45,9 +45,11 @@ from .trionics import ( | |||
|     BroadcastReceiver, | ||||
| ) | ||||
| from tractor.msg import ( | ||||
|     # Return, | ||||
|     # Stop, | ||||
|     Error, | ||||
|     Return, | ||||
|     Stop, | ||||
|     MsgType, | ||||
|     PayloadT, | ||||
|     Yield, | ||||
| ) | ||||
| 
 | ||||
|  | @ -70,8 +72,7 @@ class MsgStream(trio.abc.Channel): | |||
|     A bidirectional message stream for receiving logically sequenced | ||||
|     values over an inter-actor IPC `Channel`. | ||||
| 
 | ||||
|     This is the type returned to a local task which entered either | ||||
|     `Portal.open_stream_from()` or `Context.open_stream()`. | ||||
| 
 | ||||
| 
 | ||||
|     Termination rules: | ||||
| 
 | ||||
|  | @ -94,6 +95,9 @@ class MsgStream(trio.abc.Channel): | |||
|         self._rx_chan = rx_chan | ||||
|         self._broadcaster = _broadcaster | ||||
| 
 | ||||
|         # any actual IPC msg which is effectively an `EndOfStream` | ||||
|         self._stop_msg: bool|Stop = False | ||||
| 
 | ||||
|         # flag to denote end of stream | ||||
|         self._eoc: bool|trio.EndOfChannel = False | ||||
|         self._closed: bool|trio.ClosedResourceError = False | ||||
|  | @ -125,16 +129,67 @@ class MsgStream(trio.abc.Channel): | |||
|     def receive_nowait( | ||||
|         self, | ||||
|         expect_msg: MsgType = Yield, | ||||
|     ): | ||||
|     ) -> PayloadT: | ||||
|         ctx: Context = self._ctx | ||||
|         return ctx._pld_rx.recv_pld_nowait( | ||||
|         ( | ||||
|             msg, | ||||
|             pld, | ||||
|         ) = ctx._pld_rx.recv_msg_nowait( | ||||
|             ipc=self, | ||||
|             expect_msg=expect_msg, | ||||
|         ) | ||||
| 
 | ||||
|         # ?TODO, maybe factor this into a hyper-common `unwrap_pld()` | ||||
|         # | ||||
|         match msg: | ||||
| 
 | ||||
|             # XXX, these never seems to ever hit? cool? | ||||
|             case Stop(): | ||||
|                 log.cancel( | ||||
|                     f'Msg-stream was ended via stop msg\n' | ||||
|                     f'{msg}' | ||||
|                 ) | ||||
|             case Error(): | ||||
|                 log.error( | ||||
|                     f'Msg-stream was ended via error msg\n' | ||||
|                     f'{msg}' | ||||
|                 ) | ||||
| 
 | ||||
|             # XXX NOTE, always set any final result on the ctx to | ||||
|             # avoid teardown race conditions where previously this msg | ||||
|             # would be consumed silently (by `.aclose()` doing its | ||||
|             # own "msg drain loop" but WITHOUT those `drained: lists[MsgType]` | ||||
|             # being post-close-processed! | ||||
|             # | ||||
|             # !!TODO, see the equiv todo-comment in `.receive()` | ||||
|             # around the `if drained:` where we should prolly | ||||
|             # ACTUALLY be doing this post-close processing?? | ||||
|             # | ||||
|             case Return(pld=pld): | ||||
|                 log.warning( | ||||
|                     f'Msg-stream final result msg for IPC ctx?\n' | ||||
|                     f'{msg}' | ||||
|                 ) | ||||
|                 # XXX TODO, this **should be covered** by higher | ||||
|                 # scoped runtime-side method calls such as | ||||
|                 # `Context._deliver_msg()`, so you should never | ||||
|                 # really see the warning above or else something | ||||
|                 # racy/out-of-order is likely going on between | ||||
|                 # actor-runtime-side push tasks and the user-app-side | ||||
|                 # consume tasks! | ||||
|                 # -[ ] figure out that set of race cases and fix! | ||||
|                 # -[ ] possibly return the `msg` given an input | ||||
|                 #     arg-flag is set so we can process the `Return` | ||||
|                 #     from the `.aclose()` caller? | ||||
|                 # | ||||
|                 # breakpoint()  # to debug this RACE CASE! | ||||
|                 ctx._result = pld | ||||
|                 ctx._outcome_msg = msg | ||||
| 
 | ||||
|         return pld | ||||
| 
 | ||||
|     async def receive( | ||||
|         self, | ||||
| 
 | ||||
|         hide_tb: bool = False, | ||||
|     ): | ||||
|         ''' | ||||
|  | @ -154,7 +209,7 @@ class MsgStream(trio.abc.Channel): | |||
|         #     except trio.EndOfChannel: | ||||
|         #         raise StopAsyncIteration | ||||
|         # | ||||
|         # see ``.aclose()`` for notes on the old behaviour prior to | ||||
|         # see `.aclose()` for notes on the old behaviour prior to | ||||
|         # introducing this | ||||
|         if self._eoc: | ||||
|             raise self._eoc | ||||
|  | @ -165,7 +220,11 @@ class MsgStream(trio.abc.Channel): | |||
|         src_err: Exception|None = None  # orig tb | ||||
|         try: | ||||
|             ctx: Context = self._ctx | ||||
|             return await ctx._pld_rx.recv_pld(ipc=self) | ||||
|             pld = await ctx._pld_rx.recv_pld( | ||||
|                 ipc=self, | ||||
|                 expect_msg=Yield, | ||||
|             ) | ||||
|             return pld | ||||
| 
 | ||||
|         # XXX: the stream terminates on either of: | ||||
|         # - `self._rx_chan.receive()` raising  after manual closure | ||||
|  | @ -174,7 +233,7 @@ class MsgStream(trio.abc.Channel): | |||
|         # - via a `Stop`-msg received from remote peer task. | ||||
|         #   NOTE | ||||
|         #   |_ previously this was triggered by calling | ||||
|         #   ``._rx_chan.aclose()`` on the send side of the channel | ||||
|         #   `._rx_chan.aclose()` on the send side of the channel | ||||
|         #   inside `Actor._deliver_ctx_payload()`, but now the 'stop' | ||||
|         #   message handling gets delegated to `PldRFx.recv_pld()` | ||||
|         #   internals. | ||||
|  | @ -198,11 +257,14 @@ class MsgStream(trio.abc.Channel): | |||
|         # terminated and signal this local iterator to stop | ||||
|         drained: list[Exception|dict] = await self.aclose() | ||||
|         if drained: | ||||
|             # ?TODO? pass these to the `._ctx._drained_msgs: deque` | ||||
|             # and then iterate them as part of any `.wait_for_result()` call? | ||||
|             # | ||||
|             # from .devx import pause | ||||
|             # await pause() | ||||
|         #  ^^^^^^^^TODO? pass these to the `._ctx._drained_msgs: | ||||
|         #  deque` and then iterate them as part of any | ||||
|         #  `.wait_for_result()` call? | ||||
|         # | ||||
|         # -[ ] move the match-case processing from | ||||
|         #     `.receive_nowait()` instead to right here, use it from | ||||
|         #     a for msg in drained:` post-proc loop? | ||||
|         # | ||||
|             log.warning( | ||||
|                 'Drained context msgs during closure\n\n' | ||||
|                 f'{drained}' | ||||
|  | @ -265,9 +327,6 @@ class MsgStream(trio.abc.Channel): | |||
|          - more or less we try to maintain adherance to trio's `.aclose()` semantics: | ||||
|            https://trio.readthedocs.io/en/stable/reference-io.html#trio.abc.AsyncResource.aclose | ||||
|         ''' | ||||
| 
 | ||||
|         # rx_chan = self._rx_chan | ||||
| 
 | ||||
|         # XXX NOTE XXX | ||||
|         # it's SUPER IMPORTANT that we ensure we don't DOUBLE | ||||
|         # DRAIN msgs on closure so avoid getting stuck handing on | ||||
|  | @ -279,15 +338,16 @@ class MsgStream(trio.abc.Channel): | |||
|             # this stream has already been closed so silently succeed as | ||||
|             # per ``trio.AsyncResource`` semantics. | ||||
|             # https://trio.readthedocs.io/en/stable/reference-io.html#trio.abc.AsyncResource.aclose | ||||
|             # import tractor | ||||
|             # await tractor.pause() | ||||
|             return [] | ||||
| 
 | ||||
|         ctx: Context = self._ctx | ||||
|         drained: list[Exception|dict] = [] | ||||
|         while not drained: | ||||
|             try: | ||||
|                 maybe_final_msg = self.receive_nowait( | ||||
|                     # allow_msgs=[Yield, Return], | ||||
|                     expect_msg=Yield, | ||||
|                 maybe_final_msg: Yield|Return = self.receive_nowait( | ||||
|                     expect_msg=Yield|Return, | ||||
|                 ) | ||||
|                 if maybe_final_msg: | ||||
|                     log.debug( | ||||
|  | @ -372,18 +432,30 @@ class MsgStream(trio.abc.Channel): | |||
|         #         await rx_chan.aclose() | ||||
| 
 | ||||
|         if not self._eoc: | ||||
|             this_side: str = self._ctx.side | ||||
|             peer_side: str = self._ctx.peer_side | ||||
|             message: str = ( | ||||
|                 f'Stream self-closed by {self._ctx.side!r}-side before EoC\n' | ||||
|                 f'Stream self-closed by {this_side!r}-side before EoC from {peer_side!r}\n' | ||||
|                 # } bc a stream is a "scope"/msging-phase inside an IPC | ||||
|                 f'x}}>\n' | ||||
|                 f'|_{self}\n' | ||||
|                 f'  |_{self}\n' | ||||
|             ) | ||||
|             log.cancel(message) | ||||
|             self._eoc = trio.EndOfChannel(message) | ||||
| 
 | ||||
|             if ( | ||||
|                 (rx_chan := self._rx_chan) | ||||
|                 and | ||||
|                 (stats := rx_chan.statistics()).tasks_waiting_receive | ||||
|             ): | ||||
|                 log.cancel( | ||||
|                     f'Msg-stream is closing but there is still reader tasks,\n' | ||||
|                     f'{stats}\n' | ||||
|                 ) | ||||
| 
 | ||||
|         # ?XXX WAIT, why do we not close the local mem chan `._rx_chan` XXX? | ||||
|         # => NO, DEFINITELY NOT! <= | ||||
|         # if we're a bi-dir ``MsgStream`` BECAUSE this same | ||||
|         # if we're a bi-dir `MsgStream` BECAUSE this same | ||||
|         # core-msg-loop mem recv-chan is used to deliver the | ||||
|         # potential final result from the surrounding inter-actor | ||||
|         # `Context` so we don't want to close it until that | ||||
|  |  | |||
|  | @ -158,6 +158,7 @@ class ActorNursery: | |||
|         # configure and pass runtime state | ||||
|         _rtv = _state._runtime_vars.copy() | ||||
|         _rtv['_is_root'] = False | ||||
|         _rtv['_is_infected_aio'] = infect_asyncio | ||||
| 
 | ||||
|         # allow setting debug policy per actor | ||||
|         if debug_mode is not None: | ||||
|  | @ -394,17 +395,23 @@ async def _open_and_supervise_one_cancels_all_nursery( | |||
|     # `ActorNursery.start_actor()`). | ||||
| 
 | ||||
|     # errors from this daemon actor nursery bubble up to caller | ||||
|     async with trio.open_nursery() as da_nursery: | ||||
|     async with trio.open_nursery( | ||||
|         strict_exception_groups=False, | ||||
|         # ^XXX^ TODO? instead unpack any RAE as per "loose" style? | ||||
|     ) as da_nursery: | ||||
|         try: | ||||
|             # This is the inner level "run in actor" nursery. It is | ||||
|             # awaited first since actors spawned in this way (using | ||||
|             # ``ActorNusery.run_in_actor()``) are expected to only | ||||
|             # `ActorNusery.run_in_actor()`) are expected to only | ||||
|             # return a single result and then complete (i.e. be canclled | ||||
|             # gracefully). Errors collected from these actors are | ||||
|             # immediately raised for handling by a supervisor strategy. | ||||
|             # As such if the strategy propagates any error(s) upwards | ||||
|             # the above "daemon actor" nursery will be notified. | ||||
|             async with trio.open_nursery() as ria_nursery: | ||||
|             async with trio.open_nursery( | ||||
|                 strict_exception_groups=False, | ||||
|                 # ^XXX^ TODO? instead unpack any RAE as per "loose" style? | ||||
|             ) as ria_nursery: | ||||
| 
 | ||||
|                 an = ActorNursery( | ||||
|                     actor, | ||||
|  | @ -471,8 +478,8 @@ async def _open_and_supervise_one_cancels_all_nursery( | |||
|                             ContextCancelled, | ||||
|                         }: | ||||
|                             log.cancel( | ||||
|                                 'Actor-nursery caught remote cancellation\n\n' | ||||
| 
 | ||||
|                                 'Actor-nursery caught remote cancellation\n' | ||||
|                                 '\n' | ||||
|                                 f'{inner_err.tb_str}' | ||||
|                             ) | ||||
|                         else: | ||||
|  | @ -564,7 +571,9 @@ async def _open_and_supervise_one_cancels_all_nursery( | |||
| @acm | ||||
| # @api_frame | ||||
| async def open_nursery( | ||||
|     hide_tb: bool = True, | ||||
|     **kwargs, | ||||
|     # ^TODO, paramspec for `open_root_actor()` | ||||
| 
 | ||||
| ) -> typing.AsyncGenerator[ActorNursery, None]: | ||||
|     ''' | ||||
|  | @ -582,7 +591,7 @@ async def open_nursery( | |||
|     which cancellation scopes correspond to each spawned subactor set. | ||||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = True | ||||
|     __tracebackhide__: bool = hide_tb | ||||
|     implicit_runtime: bool = False | ||||
|     actor: Actor = current_actor(err_on_no_runtime=False) | ||||
|     an: ActorNursery|None = None | ||||
|  | @ -598,7 +607,10 @@ async def open_nursery( | |||
|             # mark us for teardown on exit | ||||
|             implicit_runtime: bool = True | ||||
| 
 | ||||
|             async with open_root_actor(**kwargs) as actor: | ||||
|             async with open_root_actor( | ||||
|                 hide_tb=hide_tb, | ||||
|                 **kwargs, | ||||
|             ) as actor: | ||||
|                 assert actor is current_actor() | ||||
| 
 | ||||
|                 try: | ||||
|  | @ -636,8 +648,10 @@ async def open_nursery( | |||
|         # show frame on any internal runtime-scope error | ||||
|         if ( | ||||
|             an | ||||
|             and not an.cancelled | ||||
|             and an._scope_error | ||||
|             and | ||||
|             not an.cancelled | ||||
|             and | ||||
|             an._scope_error | ||||
|         ): | ||||
|             __tracebackhide__: bool = False | ||||
| 
 | ||||
|  |  | |||
|  | @ -19,10 +19,16 @@ Various helpers/utils for auditing your `tractor` app and/or the | |||
| core runtime. | ||||
| 
 | ||||
| ''' | ||||
| from contextlib import asynccontextmanager as acm | ||||
| from contextlib import ( | ||||
|     asynccontextmanager as acm, | ||||
| ) | ||||
| import os | ||||
| import pathlib | ||||
| 
 | ||||
| import tractor | ||||
| from tractor.devx._debug import ( | ||||
|     BoxedMaybeException, | ||||
| ) | ||||
| from .pytest import ( | ||||
|     tractor_test as tractor_test | ||||
| ) | ||||
|  | @ -54,6 +60,35 @@ def examples_dir() -> pathlib.Path: | |||
|     return repodir() / 'examples' | ||||
| 
 | ||||
| 
 | ||||
| def mk_cmd( | ||||
|     ex_name: str, | ||||
|     exs_subpath: str = 'debugging', | ||||
| ) -> str: | ||||
|     ''' | ||||
|     Generate a shell command suitable to pass to `pexpect.spawn()` | ||||
|     which runs the script as a python program's entrypoint. | ||||
| 
 | ||||
|     In particular ensure we disable the new tb coloring via unsetting | ||||
|     `$PYTHON_COLORS` so that `pexpect` can pattern match without | ||||
|     color-escape-codes. | ||||
| 
 | ||||
|     ''' | ||||
|     script_path: pathlib.Path = ( | ||||
|         examples_dir() | ||||
|         / exs_subpath | ||||
|         / f'{ex_name}.py' | ||||
|     ) | ||||
|     py_cmd: str = ' '.join([ | ||||
|         'python', | ||||
|         str(script_path) | ||||
|     ]) | ||||
|     # XXX, required for py 3.13+ | ||||
|     # https://docs.python.org/3/using/cmdline.html#using-on-controlling-color | ||||
|     # https://docs.python.org/3/using/cmdline.html#envvar-PYTHON_COLORS | ||||
|     os.environ['PYTHON_COLORS'] = '0' | ||||
|     return py_cmd | ||||
| 
 | ||||
| 
 | ||||
| @acm | ||||
| async def expect_ctxc( | ||||
|     yay: bool, | ||||
|  | @ -66,12 +101,13 @@ async def expect_ctxc( | |||
|     ''' | ||||
|     if yay: | ||||
|         try: | ||||
|             yield | ||||
|             yield (maybe_exc := BoxedMaybeException()) | ||||
|             raise RuntimeError('Never raised ctxc?') | ||||
|         except tractor.ContextCancelled: | ||||
|         except tractor.ContextCancelled as ctxc: | ||||
|             maybe_exc.value = ctxc | ||||
|             if reraise: | ||||
|                 raise | ||||
|             else: | ||||
|                 return | ||||
|     else: | ||||
|         yield | ||||
|         yield (maybe_exc := BoxedMaybeException()) | ||||
|  |  | |||
|  | @ -26,7 +26,7 @@ from ._debug import ( | |||
|     breakpoint as breakpoint, | ||||
|     pause as pause, | ||||
|     pause_from_sync as pause_from_sync, | ||||
|     shield_sigint_handler as shield_sigint_handler, | ||||
|     sigint_shield as sigint_shield, | ||||
|     open_crash_handler as open_crash_handler, | ||||
|     maybe_open_crash_handler as maybe_open_crash_handler, | ||||
|     maybe_init_greenback as maybe_init_greenback, | ||||
|  |  | |||
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							|  | @ -234,7 +234,7 @@ def find_caller_info( | |||
| _frame2callerinfo_cache: dict[FrameType, CallerInfo] = {} | ||||
| 
 | ||||
| 
 | ||||
| # TODO: -[x] move all this into new `.devx._code`! | ||||
| # TODO: -[x] move all this into new `.devx._frame_stack`! | ||||
| # -[ ] consider rename to _callstack? | ||||
| # -[ ] prolly create a `@runtime_api` dec? | ||||
| #   |_ @api_frame seems better? | ||||
|  | @ -286,3 +286,18 @@ def api_frame( | |||
|     wrapped._call_infos: dict[FrameType, CallerInfo] = _frame2callerinfo_cache | ||||
|     wrapped.__api_func__: bool = True | ||||
|     return wrapper(wrapped) | ||||
| 
 | ||||
| 
 | ||||
| # TODO: something like this instead of the adhoc frame-unhiding | ||||
| # blocks all over the runtime!! XD | ||||
| # -[ ] ideally we can expect a certain error (set) and if something | ||||
| #     else is raised then all frames below the wrapped one will be | ||||
| #     un-hidden via `__tracebackhide__: bool = False`. | ||||
| # |_ might need to dynamically mutate the code objs like | ||||
| #    `pdbp.hideframe()` does? | ||||
| # -[ ] use this as a `@acm` decorator as introed in 3.10? | ||||
| # @acm | ||||
| # async def unhide_frame_when_not( | ||||
| #     error_set: set[BaseException], | ||||
| # ) -> TracebackType: | ||||
| #     ... | ||||
|  |  | |||
|  | @ -24,19 +24,32 @@ disjoint, parallel executing tasks in separate actors. | |||
| 
 | ||||
| ''' | ||||
| from __future__ import annotations | ||||
| # from functools import partial | ||||
| from threading import ( | ||||
|     current_thread, | ||||
|     Thread, | ||||
|     RLock, | ||||
| ) | ||||
| import multiprocessing as mp | ||||
| from signal import ( | ||||
|     signal, | ||||
|     getsignal, | ||||
|     SIGUSR1, | ||||
|     SIGINT, | ||||
| ) | ||||
| # import traceback | ||||
| from types import ModuleType | ||||
| from typing import ( | ||||
|     Callable, | ||||
|     TYPE_CHECKING, | ||||
| ) | ||||
| import traceback | ||||
| from typing import TYPE_CHECKING | ||||
| 
 | ||||
| import trio | ||||
| from tractor import ( | ||||
|     _state, | ||||
|     log as logmod, | ||||
| ) | ||||
| from tractor.devx import _debug | ||||
| 
 | ||||
| log = logmod.get_logger(__name__) | ||||
| 
 | ||||
|  | @ -51,26 +64,68 @@ if TYPE_CHECKING: | |||
| 
 | ||||
| @trio.lowlevel.disable_ki_protection | ||||
| def dump_task_tree() -> None: | ||||
|     import stackscope | ||||
|     from tractor.log import get_console_log | ||||
|     ''' | ||||
|     Do a classic `stackscope.extract()` task-tree dump to console at | ||||
|     `.devx()` level. | ||||
| 
 | ||||
|     ''' | ||||
|     import stackscope | ||||
|     tree_str: str = str( | ||||
|         stackscope.extract( | ||||
|             trio.lowlevel.current_root_task(), | ||||
|             recurse_child_tasks=True | ||||
|         ) | ||||
|     ) | ||||
|     log = get_console_log( | ||||
|         name=__name__, | ||||
|         level='cancel', | ||||
|     ) | ||||
|     actor: Actor = _state.current_actor() | ||||
|     thr: Thread = current_thread() | ||||
|     current_sigint_handler: Callable = getsignal(SIGINT) | ||||
|     if ( | ||||
|         current_sigint_handler | ||||
|         is not | ||||
|         _debug.DebugStatus._trio_handler | ||||
|     ): | ||||
|         sigint_handler_report: str = ( | ||||
|             'The default `trio` SIGINT handler was replaced?!' | ||||
|         ) | ||||
|     else: | ||||
|         sigint_handler_report: str = ( | ||||
|             'The default `trio` SIGINT handler is in use?!' | ||||
|         ) | ||||
| 
 | ||||
|     # sclang symbology | ||||
|     # |_<object> | ||||
|     # |_(Task/Thread/Process/Actor | ||||
|     # |_{Supervisor/Scope | ||||
|     # |_[Storage/Memory/IPC-Stream/Data-Struct | ||||
| 
 | ||||
|     log.devx( | ||||
|         f'Dumping `stackscope` tree for actor\n' | ||||
|         f'{actor.name}: {actor}\n' | ||||
|         f' |_{mp.current_process()}\n\n' | ||||
|         f'{tree_str}\n' | ||||
|         f'(>: {actor.uid!r}\n' | ||||
|         f' |_{mp.current_process()}\n' | ||||
|         f'   |_{thr}\n' | ||||
|         f'     |_{actor}\n' | ||||
|         f'\n' | ||||
|         f'{sigint_handler_report}\n' | ||||
|         f'signal.getsignal(SIGINT) -> {current_sigint_handler!r}\n' | ||||
|         # f'\n' | ||||
|         # start-of-trace-tree delimiter (mostly for testing) | ||||
|         # f'------ {actor.uid!r} ------\n' | ||||
|         f'\n' | ||||
|         f'------ start-of-{actor.uid!r} ------\n' | ||||
|         f'|\n' | ||||
|         f'{tree_str}' | ||||
|         # end-of-trace-tree delimiter (mostly for testing) | ||||
|         f'|\n' | ||||
|         f'|_____ end-of-{actor.uid!r} ______\n' | ||||
|     ) | ||||
|     # TODO: can remove this right? | ||||
|     # -[ ] was original code from author | ||||
|     # | ||||
|     # print( | ||||
|     #     'DUMPING FROM PRINT\n' | ||||
|     #     + | ||||
|     #     content | ||||
|     # ) | ||||
|     # import logging | ||||
|     # try: | ||||
|     #     with open("/dev/tty", "w") as tty: | ||||
|  | @ -80,58 +135,130 @@ def dump_task_tree() -> None: | |||
|     #         "task_tree" | ||||
|     #     ).exception("Error printing task tree") | ||||
| 
 | ||||
| _handler_lock = RLock() | ||||
| _tree_dumped: bool = False | ||||
| 
 | ||||
| def signal_handler( | ||||
| 
 | ||||
| def dump_tree_on_sig( | ||||
|     sig: int, | ||||
|     frame: object, | ||||
| 
 | ||||
|     relay_to_subs: bool = True, | ||||
| 
 | ||||
| ) -> None: | ||||
|     try: | ||||
|         trio.lowlevel.current_trio_token( | ||||
|         ).run_sync_soon(dump_task_tree) | ||||
|     except RuntimeError: | ||||
|         # not in async context -- print a normal traceback | ||||
|         traceback.print_stack() | ||||
|     global _tree_dumped, _handler_lock | ||||
|     with _handler_lock: | ||||
|         # if _tree_dumped: | ||||
|         #     log.warning( | ||||
|         #         'Already dumped for this actor...??' | ||||
|         #     ) | ||||
|         #     return | ||||
| 
 | ||||
|         _tree_dumped = True | ||||
| 
 | ||||
|         # actor: Actor = _state.current_actor() | ||||
|         log.devx( | ||||
|             'Trying to dump `stackscope` tree..\n' | ||||
|         ) | ||||
|         try: | ||||
|             dump_task_tree() | ||||
|             # await actor._service_n.start_soon( | ||||
|             #     partial( | ||||
|             #         trio.to_thread.run_sync, | ||||
|             #         dump_task_tree, | ||||
|             #     ) | ||||
|             # ) | ||||
|             # trio.lowlevel.current_trio_token().run_sync_soon( | ||||
|             #     dump_task_tree | ||||
|             # ) | ||||
| 
 | ||||
|         except RuntimeError: | ||||
|             log.exception( | ||||
|                 'Failed to dump `stackscope` tree..\n' | ||||
|             ) | ||||
|             # not in async context -- print a normal traceback | ||||
|             # traceback.print_stack() | ||||
|             raise | ||||
| 
 | ||||
|         except BaseException: | ||||
|             log.exception( | ||||
|                 'Failed to dump `stackscope` tree..\n' | ||||
|             ) | ||||
|             raise | ||||
| 
 | ||||
|         # log.devx( | ||||
|         #     'Supposedly we dumped just fine..?' | ||||
|         # ) | ||||
| 
 | ||||
|     if not relay_to_subs: | ||||
|         return | ||||
| 
 | ||||
|     an: ActorNursery | ||||
|     for an in _state.current_actor()._actoruid2nursery.values(): | ||||
| 
 | ||||
|         subproc: ProcessType | ||||
|         subactor: Actor | ||||
|         for subactor, subproc, _ in an._children.values(): | ||||
|             log.devx( | ||||
|             log.warning( | ||||
|                 f'Relaying `SIGUSR1`[{sig}] to sub-actor\n' | ||||
|                 f'{subactor}\n' | ||||
|                 f' |_{subproc}\n' | ||||
|             ) | ||||
| 
 | ||||
|             if isinstance(subproc, trio.Process): | ||||
|                 subproc.send_signal(sig) | ||||
|             # bc of course stdlib can't have a std API.. XD | ||||
|             match subproc: | ||||
|                 case trio.Process(): | ||||
|                     subproc.send_signal(sig) | ||||
| 
 | ||||
|             elif isinstance(subproc, mp.Process): | ||||
|                 subproc._send_signal(sig) | ||||
|                 case mp.Process(): | ||||
|                     subproc._send_signal(sig) | ||||
| 
 | ||||
| 
 | ||||
| def enable_stack_on_sig( | ||||
|     sig: int = SIGUSR1 | ||||
| ) -> None: | ||||
|     sig: int = SIGUSR1, | ||||
| ) -> ModuleType: | ||||
|     ''' | ||||
|     Enable `stackscope` tracing on reception of a signal; by | ||||
|     default this is SIGUSR1. | ||||
| 
 | ||||
|     HOT TIP: a task/ctx-tree dump can be triggered from a shell with | ||||
|     fancy cmds. | ||||
| 
 | ||||
|     For ex. from `bash` using `pgrep` and cmd-sustitution | ||||
|     (https://www.gnu.org/software/bash/manual/bash.html#Command-Substitution) | ||||
|     you could use: | ||||
| 
 | ||||
|     >> kill -SIGUSR1 $(pgrep -f <part-of-cmd: str>) | ||||
| 
 | ||||
|     OR without a sub-shell, | ||||
| 
 | ||||
|     >> pkill --signal SIGUSR1 -f <part-of-cmd: str> | ||||
| 
 | ||||
|     ''' | ||||
|     try: | ||||
|         import stackscope | ||||
|     except ImportError: | ||||
|         log.warning( | ||||
|             '`stackscope` not installed for use in debug mode!' | ||||
|         ) | ||||
|         return None | ||||
| 
 | ||||
|     handler: Callable|int = getsignal(sig) | ||||
|     if handler is dump_tree_on_sig: | ||||
|         log.devx( | ||||
|             'A `SIGUSR1` handler already exists?\n' | ||||
|             f'|_ {handler!r}\n' | ||||
|         ) | ||||
|         return | ||||
| 
 | ||||
|     signal( | ||||
|         sig, | ||||
|         signal_handler, | ||||
|         dump_tree_on_sig, | ||||
|     ) | ||||
|     # NOTE: not the above can be triggered from | ||||
|     # a (xonsh) shell using: | ||||
|     # kill -SIGUSR1 @$(pgrep -f '<cmd>') | ||||
|     # | ||||
|     # for example if you were looking to trace a `pytest` run | ||||
|     # kill -SIGUSR1 @$(pgrep -f 'pytest') | ||||
|     log.devx( | ||||
|         'Enabling trace-trees on `SIGUSR1` ' | ||||
|         'since `stackscope` is installed @ \n' | ||||
|         f'{stackscope!r}\n\n' | ||||
|         f'With `SIGUSR1` handler\n' | ||||
|         f'|_{dump_tree_on_sig}\n' | ||||
|     ) | ||||
|     return stackscope | ||||
|  |  | |||
|  | @ -53,6 +53,7 @@ def pformat_boxed_tb( | |||
| 
 | ||||
|     tb_box_indent: int|None = None, | ||||
|     tb_body_indent: int = 1, | ||||
|     boxer_header: str = '-' | ||||
| 
 | ||||
| ) -> str: | ||||
|     ''' | ||||
|  | @ -88,10 +89,10 @@ def pformat_boxed_tb( | |||
| 
 | ||||
|     tb_box: str = ( | ||||
|         f'|\n' | ||||
|         f' ------ - ------\n' | ||||
|         f' ------ {boxer_header} ------\n' | ||||
|         f'{tb_body}' | ||||
|         f' ------ - ------\n' | ||||
|         f'_|\n' | ||||
|         f' ------ {boxer_header}- ------\n' | ||||
|         f'_|' | ||||
|     ) | ||||
|     tb_box_indent: str = ( | ||||
|         tb_box_indent | ||||
|  |  | |||
|  | @ -0,0 +1,26 @@ | |||
| # tractor: structured concurrent "actors". | ||||
| # Copyright 2024-eternity Tyler Goodlet. | ||||
| 
 | ||||
| # This program is free software: you can redistribute it and/or modify | ||||
| # it under the terms of the GNU Affero General Public License as published by | ||||
| # the Free Software Foundation, either version 3 of the License, or | ||||
| # (at your option) any later version. | ||||
| 
 | ||||
| # This program is distributed in the hope that it will be useful, | ||||
| # but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
| # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | ||||
| # GNU Affero General Public License for more details. | ||||
| 
 | ||||
| # You should have received a copy of the GNU Affero General Public License | ||||
| # along with this program.  If not, see <https://www.gnu.org/licenses/>. | ||||
| 
 | ||||
| ''' | ||||
| High level design patterns, APIs and runtime extensions built on top | ||||
| of the `tractor` runtime core. | ||||
| 
 | ||||
| ''' | ||||
| from ._service import ( | ||||
|     open_service_mngr as open_service_mngr, | ||||
|     get_service_mngr as get_service_mngr, | ||||
|     ServiceMngr as ServiceMngr, | ||||
| ) | ||||
|  | @ -0,0 +1,592 @@ | |||
| # tractor: structured concurrent "actors". | ||||
| # Copyright 2024-eternity Tyler Goodlet. | ||||
| 
 | ||||
| # This program is free software: you can redistribute it and/or modify | ||||
| # it under the terms of the GNU Affero General Public License as published by | ||||
| # the Free Software Foundation, either version 3 of the License, or | ||||
| # (at your option) any later version. | ||||
| 
 | ||||
| # This program is distributed in the hope that it will be useful, | ||||
| # but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
| # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | ||||
| # GNU Affero General Public License for more details. | ||||
| 
 | ||||
| # You should have received a copy of the GNU Affero General Public License | ||||
| # along with this program.  If not, see <https://www.gnu.org/licenses/>. | ||||
| 
 | ||||
| ''' | ||||
| Daemon subactor as service(s) management and supervision primitives | ||||
| and API. | ||||
| 
 | ||||
| ''' | ||||
| from __future__ import annotations | ||||
| from contextlib import ( | ||||
|     asynccontextmanager as acm, | ||||
|     # contextmanager as cm, | ||||
| ) | ||||
| from collections import defaultdict | ||||
| from dataclasses import ( | ||||
|     dataclass, | ||||
|     field, | ||||
| ) | ||||
| import functools | ||||
| import inspect | ||||
| from typing import ( | ||||
|     Callable, | ||||
|     Any, | ||||
| ) | ||||
| 
 | ||||
| import tractor | ||||
| import trio | ||||
| from trio import TaskStatus | ||||
| from tractor import ( | ||||
|     log, | ||||
|     ActorNursery, | ||||
|     current_actor, | ||||
|     ContextCancelled, | ||||
|     Context, | ||||
|     Portal, | ||||
| ) | ||||
| 
 | ||||
| log = log.get_logger('tractor') | ||||
| 
 | ||||
| 
 | ||||
| # TODO: implement a `@singleton` deco-API for wrapping the below | ||||
| # factory's impl for general actor-singleton use? | ||||
| # | ||||
| # -[ ] go through the options peeps on SO did? | ||||
| #  * https://stackoverflow.com/questions/6760685/what-is-the-best-way-of-implementing-singleton-in-python | ||||
| #  * including @mikenerone's answer | ||||
| #   |_https://stackoverflow.com/questions/6760685/what-is-the-best-way-of-implementing-singleton-in-python/39186313#39186313 | ||||
| # | ||||
| # -[ ] put it in `tractor.lowlevel._globals` ? | ||||
| #  * fits with our oustanding actor-local/global feat req? | ||||
| #   |_ https://github.com/goodboy/tractor/issues/55 | ||||
| #  * how can it relate to the `Actor.lifetime_stack` that was | ||||
| #    silently patched in? | ||||
| #   |_ we could implicitly call both of these in the same | ||||
| #     spot in the runtime using the lifetime stack? | ||||
| #    - `open_singleton_cm().__exit__()` | ||||
| #    -`del_singleton()` | ||||
| #   |_ gives SC fixtue semantics to sync code oriented around | ||||
| #     sub-process lifetime? | ||||
| #  * what about with `trio.RunVar`? | ||||
| #   |_https://trio.readthedocs.io/en/stable/reference-lowlevel.html#trio.lowlevel.RunVar | ||||
| #    - which we'll need for no-GIL cpython (right?) presuming | ||||
| #      multiple `trio.run()` calls in process? | ||||
| # | ||||
| # | ||||
| # @singleton | ||||
| # async def open_service_mngr( | ||||
| #     **init_kwargs, | ||||
| # ) -> ServiceMngr: | ||||
| #     ''' | ||||
| #     Note this function body is invoke IFF no existing singleton instance already | ||||
| #     exists in this proc's memory. | ||||
| 
 | ||||
| #     ''' | ||||
| #     # setup | ||||
| #     yield ServiceMngr(**init_kwargs) | ||||
| #     # teardown | ||||
| 
 | ||||
| 
 | ||||
| # a deletion API for explicit instance de-allocation? | ||||
| # @open_service_mngr.deleter | ||||
| # def del_service_mngr() -> None: | ||||
| #     mngr = open_service_mngr._singleton[0] | ||||
| #     open_service_mngr._singleton[0] = None | ||||
| #     del mngr | ||||
| 
 | ||||
| 
 | ||||
| 
 | ||||
| # TODO: implement a singleton deco-API for wrapping the below | ||||
| # factory's impl for general actor-singleton use? | ||||
| # | ||||
| # @singleton | ||||
| # async def open_service_mngr( | ||||
| #     **init_kwargs, | ||||
| # ) -> ServiceMngr: | ||||
| #     ''' | ||||
| #     Note this function body is invoke IFF no existing singleton instance already | ||||
| #     exists in this proc's memory. | ||||
| 
 | ||||
| #     ''' | ||||
| #     # setup | ||||
| #     yield ServiceMngr(**init_kwargs) | ||||
| #     # teardown | ||||
| 
 | ||||
| 
 | ||||
| 
 | ||||
| # TODO: singleton factory API instead of a class API | ||||
| @acm | ||||
| async def open_service_mngr( | ||||
|     *, | ||||
|     debug_mode: bool = False, | ||||
| 
 | ||||
|     # NOTE; since default values for keyword-args are effectively | ||||
|     # module-vars/globals as per the note from, | ||||
|     # https://docs.python.org/3/tutorial/controlflow.html#default-argument-values | ||||
|     # | ||||
|     # > "The default value is evaluated only once. This makes | ||||
|     #   a difference when the default is a mutable object such as | ||||
|     #   a list, dictionary, or instances of most classes" | ||||
|     # | ||||
|     _singleton: list[ServiceMngr|None] = [None], | ||||
|     **init_kwargs, | ||||
| 
 | ||||
| ) -> ServiceMngr: | ||||
|     ''' | ||||
|     Open an actor-global "service-manager" for supervising a tree | ||||
|     of subactors and/or actor-global tasks. | ||||
| 
 | ||||
|     The delivered `ServiceMngr` is singleton instance for each | ||||
|     actor-process, that is, allocated on first open and never | ||||
|     de-allocated unless explicitly deleted by al call to | ||||
|     `del_service_mngr()`. | ||||
| 
 | ||||
|     ''' | ||||
|     # TODO: factor this an allocation into | ||||
|     # a `._mngr.open_service_mngr()` and put in the | ||||
|     # once-n-only-once setup/`.__aenter__()` part! | ||||
|     # -[ ] how to make this only happen on the `mngr == None` case? | ||||
|     #  |_ use `.trionics.maybe_open_context()` (for generic | ||||
|     #     async-with-style-only-once of the factory impl, though | ||||
|     #     what do we do for the allocation case? | ||||
|     #    / `.maybe_open_nursery()` (since for this specific case | ||||
|     #    it's simpler?) to activate | ||||
|     async with ( | ||||
|         tractor.open_nursery() as an, | ||||
|         trio.open_nursery() as tn, | ||||
|     ): | ||||
|         # impl specific obvi.. | ||||
|         init_kwargs.update({ | ||||
|             'an': an, | ||||
|             'tn': tn, | ||||
|         }) | ||||
| 
 | ||||
|         mngr: ServiceMngr|None | ||||
|         if (mngr := _singleton[0]) is None: | ||||
| 
 | ||||
|             log.info('Allocating a new service mngr!') | ||||
|             mngr = _singleton[0] = ServiceMngr(**init_kwargs) | ||||
| 
 | ||||
|             # TODO: put into `.__aenter__()` section of | ||||
|             # eventual `@singleton_acm` API wrapper. | ||||
|             # | ||||
|             # assign globally for future daemon/task creation | ||||
|             mngr.an = an | ||||
|             mngr.tn = tn | ||||
| 
 | ||||
|         else: | ||||
|             assert (mngr.an and mngr.tn) | ||||
|             log.info( | ||||
|                 'Using extant service mngr!\n\n' | ||||
|                 f'{mngr!r}\n'  # it has a nice `.__repr__()` of services state | ||||
|             ) | ||||
| 
 | ||||
|         try: | ||||
|             # NOTE: this is a singleton factory impl specific detail | ||||
|             # which should be supported in the condensed | ||||
|             # `@singleton_acm` API? | ||||
|             mngr.debug_mode = debug_mode | ||||
| 
 | ||||
|             yield mngr | ||||
|         finally: | ||||
|             # TODO: is this more clever/efficient? | ||||
|             # if 'samplerd' in mngr.service_ctxs: | ||||
|             #     await mngr.cancel_service('samplerd') | ||||
|             tn.cancel_scope.cancel() | ||||
| 
 | ||||
| 
 | ||||
| 
 | ||||
| def get_service_mngr() -> ServiceMngr: | ||||
|     ''' | ||||
|     Try to get the singleton service-mngr for this actor presuming it | ||||
|     has already been allocated using, | ||||
| 
 | ||||
|     .. code:: python | ||||
| 
 | ||||
|         async with open_<@singleton_acm(func)>() as mngr` | ||||
|             ... this block kept open ... | ||||
| 
 | ||||
|     If not yet allocated raise a `ServiceError`. | ||||
| 
 | ||||
|     ''' | ||||
|     # https://stackoverflow.com/a/12627202 | ||||
|     # https://docs.python.org/3/library/inspect.html#inspect.Signature | ||||
|     maybe_mngr: ServiceMngr|None = inspect.signature( | ||||
|         open_service_mngr | ||||
|     ).parameters['_singleton'].default[0] | ||||
| 
 | ||||
|     if maybe_mngr is None: | ||||
|         raise RuntimeError( | ||||
|             'Someone must allocate a `ServiceMngr` using\n\n' | ||||
|             '`async with open_service_mngr()` beforehand!!\n' | ||||
|         ) | ||||
| 
 | ||||
|     return maybe_mngr | ||||
| 
 | ||||
| 
 | ||||
| async def _open_and_supervise_service_ctx( | ||||
|     serman: ServiceMngr, | ||||
|     name: str, | ||||
|     ctx_fn: Callable,  # TODO, type for `@tractor.context` requirement | ||||
|     portal: Portal, | ||||
| 
 | ||||
|     allow_overruns: bool = False, | ||||
|     task_status: TaskStatus[ | ||||
|         tuple[ | ||||
|             trio.CancelScope, | ||||
|             Context, | ||||
|             trio.Event, | ||||
|             Any, | ||||
|         ] | ||||
|     ] = trio.TASK_STATUS_IGNORED, | ||||
|     **ctx_kwargs, | ||||
| 
 | ||||
| ) -> Any: | ||||
|     ''' | ||||
|     Open a remote IPC-context defined by `ctx_fn` in the | ||||
|     (service) actor accessed via `portal` and supervise the | ||||
|     (local) parent task to termination at which point the remote | ||||
|     actor runtime is cancelled alongside it. | ||||
| 
 | ||||
|     The main application is for allocating long-running | ||||
|     "sub-services" in a main daemon and explicitly controlling | ||||
|     their lifetimes from an actor-global singleton. | ||||
| 
 | ||||
|     ''' | ||||
|     # TODO: use the ctx._scope directly here instead? | ||||
|     # -[ ] actually what semantics do we expect for this | ||||
|     #   usage!? | ||||
|     with trio.CancelScope() as cs: | ||||
|         try: | ||||
|             async with portal.open_context( | ||||
|                 ctx_fn, | ||||
|                 allow_overruns=allow_overruns, | ||||
|                 **ctx_kwargs, | ||||
| 
 | ||||
|             ) as (ctx, started): | ||||
| 
 | ||||
|                 # unblock once the remote context has started | ||||
|                 complete = trio.Event() | ||||
|                 task_status.started(( | ||||
|                     cs, | ||||
|                     ctx, | ||||
|                     complete, | ||||
|                     started, | ||||
|                 )) | ||||
|                 log.info( | ||||
|                     f'`pikerd` service {name} started with value {started}' | ||||
|                 ) | ||||
|                 # wait on any context's return value | ||||
|                 # and any final portal result from the | ||||
|                 # sub-actor. | ||||
|                 ctx_res: Any = await ctx.wait_for_result() | ||||
| 
 | ||||
|                 # NOTE: blocks indefinitely until cancelled | ||||
|                 # either by error from the target context | ||||
|                 # function or by being cancelled here by the | ||||
|                 # surrounding cancel scope. | ||||
|                 return ( | ||||
|                     await portal.wait_for_result(), | ||||
|                     ctx_res, | ||||
|                 ) | ||||
| 
 | ||||
|         except ContextCancelled as ctxe: | ||||
|             canceller: tuple[str, str] = ctxe.canceller | ||||
|             our_uid: tuple[str, str] = current_actor().uid | ||||
|             if ( | ||||
|                 canceller != portal.chan.uid | ||||
|                 and | ||||
|                 canceller != our_uid | ||||
|             ): | ||||
|                 log.cancel( | ||||
|                     f'Actor-service `{name}` was remotely cancelled by a peer?\n' | ||||
| 
 | ||||
|                     # TODO: this would be a good spot to use | ||||
|                     # a respawn feature Bo | ||||
|                     f'-> Keeping `pikerd` service manager alive despite this inter-peer cancel\n\n' | ||||
| 
 | ||||
|                     f'cancellee: {portal.chan.uid}\n' | ||||
|                     f'canceller: {canceller}\n' | ||||
|                 ) | ||||
|             else: | ||||
|                 raise | ||||
| 
 | ||||
|         finally: | ||||
|             # NOTE: the ctx MUST be cancelled first if we | ||||
|             # don't want the above `ctx.wait_for_result()` to | ||||
|             # raise a self-ctxc. WHY, well since from the ctx's | ||||
|             # perspective the cancel request will have | ||||
|             # arrived out-out-of-band at the `Actor.cancel()` | ||||
|             # level, thus `Context.cancel_called == False`, | ||||
|             # meaning `ctx._is_self_cancelled() == False`. | ||||
|             # with trio.CancelScope(shield=True): | ||||
|             # await ctx.cancel() | ||||
|             await portal.cancel_actor()  # terminate (remote) sub-actor | ||||
|             complete.set()  # signal caller this task is done | ||||
|             serman.service_ctxs.pop(name)  # remove mngr entry | ||||
| 
 | ||||
| 
 | ||||
| # TODO: we need remote wrapping and a general soln: | ||||
| # - factor this into a ``tractor.highlevel`` extension # pack for the | ||||
| #   library. | ||||
| # - wrap a "remote api" wherein you can get a method proxy | ||||
| #   to the pikerd actor for starting services remotely! | ||||
| # - prolly rename this to ActorServicesNursery since it spawns | ||||
| #   new actors and supervises them to completion? | ||||
| @dataclass | ||||
| class ServiceMngr: | ||||
|     ''' | ||||
|     A multi-subactor-as-service manager. | ||||
| 
 | ||||
|     Spawn, supervise and monitor service/daemon subactors in a SC | ||||
|     process tree. | ||||
| 
 | ||||
|     ''' | ||||
|     an: ActorNursery | ||||
|     tn: trio.Nursery | ||||
|     debug_mode: bool = False # tractor sub-actor debug mode flag | ||||
| 
 | ||||
|     service_tasks: dict[ | ||||
|         str, | ||||
|         tuple[ | ||||
|             trio.CancelScope, | ||||
|             trio.Event, | ||||
|         ] | ||||
|     ] = field(default_factory=dict) | ||||
| 
 | ||||
|     service_ctxs: dict[ | ||||
|         str, | ||||
|         tuple[ | ||||
|             trio.CancelScope, | ||||
|             Context, | ||||
|             Portal, | ||||
|             trio.Event, | ||||
|         ] | ||||
|     ] = field(default_factory=dict) | ||||
| 
 | ||||
|     # internal per-service task mutexs | ||||
|     _locks = defaultdict(trio.Lock) | ||||
| 
 | ||||
|     # TODO, unify this interface with our `TaskManager` PR! | ||||
|     # | ||||
|     # | ||||
|     async def start_service_task( | ||||
|         self, | ||||
|         name: str, | ||||
|         # TODO: typevar for the return type of the target and then | ||||
|         # use it below for `ctx_res`? | ||||
|         fn: Callable, | ||||
| 
 | ||||
|         allow_overruns: bool = False, | ||||
|         **ctx_kwargs, | ||||
| 
 | ||||
|     ) -> tuple[ | ||||
|         trio.CancelScope, | ||||
|         Any, | ||||
|         trio.Event, | ||||
|     ]: | ||||
|         async def _task_manager_start( | ||||
|             task_status: TaskStatus[ | ||||
|                 tuple[ | ||||
|                     trio.CancelScope, | ||||
|                     trio.Event, | ||||
|                 ] | ||||
|             ] = trio.TASK_STATUS_IGNORED, | ||||
|         ) -> Any: | ||||
| 
 | ||||
|             task_cs = trio.CancelScope() | ||||
|             task_complete = trio.Event() | ||||
| 
 | ||||
|             with task_cs as cs: | ||||
|                 task_status.started(( | ||||
|                     cs, | ||||
|                     task_complete, | ||||
|                 )) | ||||
|                 try: | ||||
|                     await fn() | ||||
|                 except trio.Cancelled as taskc: | ||||
|                     log.cancel( | ||||
|                         f'Service task for `{name}` was cancelled!\n' | ||||
|                         # TODO: this would be a good spot to use | ||||
|                         # a respawn feature Bo | ||||
|                     ) | ||||
|                     raise taskc | ||||
|                 finally: | ||||
|                     task_complete.set() | ||||
|         ( | ||||
|             cs, | ||||
|             complete, | ||||
|         ) = await self.tn.start(_task_manager_start) | ||||
| 
 | ||||
|         # store the cancel scope and portal for later cancellation or | ||||
|         # retstart if needed. | ||||
|         self.service_tasks[name] = ( | ||||
|             cs, | ||||
|             complete, | ||||
|         ) | ||||
|         return ( | ||||
|             cs, | ||||
|             complete, | ||||
|         ) | ||||
| 
 | ||||
|     async def cancel_service_task( | ||||
|         self, | ||||
|         name: str, | ||||
| 
 | ||||
|     ) -> Any: | ||||
|         log.info(f'Cancelling `pikerd` service {name}') | ||||
|         cs, complete = self.service_tasks[name] | ||||
| 
 | ||||
|         cs.cancel() | ||||
|         await complete.wait() | ||||
|         # TODO, if we use the `TaskMngr` from #346 | ||||
|         # we can also get the return value from the task! | ||||
| 
 | ||||
|         if name in self.service_tasks: | ||||
|             # TODO: custom err? | ||||
|             # raise ServiceError( | ||||
|             raise RuntimeError( | ||||
|                 f'Service task {name!r} not terminated!?\n' | ||||
|             ) | ||||
| 
 | ||||
|     async def start_service_ctx( | ||||
|         self, | ||||
|         name: str, | ||||
|         portal: Portal, | ||||
|         # TODO: typevar for the return type of the target and then | ||||
|         # use it below for `ctx_res`? | ||||
|         ctx_fn: Callable, | ||||
|         **ctx_kwargs, | ||||
| 
 | ||||
|     ) -> tuple[ | ||||
|         trio.CancelScope, | ||||
|         Context, | ||||
|         Any, | ||||
|     ]: | ||||
|         ''' | ||||
|         Start a remote IPC-context defined by `ctx_fn` in a background | ||||
|         task and immediately return supervision primitives to manage it: | ||||
| 
 | ||||
|         - a `cs: CancelScope` for the newly allocated bg task | ||||
|         - the `ipc_ctx: Context` to manage the remotely scheduled | ||||
|           `trio.Task`. | ||||
|         - the `started: Any` value returned by the remote endpoint | ||||
|           task's `Context.started(<value>)` call. | ||||
| 
 | ||||
|         The bg task supervises the ctx such that when it terminates the supporting | ||||
|         actor runtime is also cancelled, see `_open_and_supervise_service_ctx()` | ||||
|         for details. | ||||
| 
 | ||||
|         ''' | ||||
|         cs, ipc_ctx, complete, started = await self.tn.start( | ||||
|             functools.partial( | ||||
|                 _open_and_supervise_service_ctx, | ||||
|                 serman=self, | ||||
|                 name=name, | ||||
|                 ctx_fn=ctx_fn, | ||||
|                 portal=portal, | ||||
|                 **ctx_kwargs, | ||||
|             ) | ||||
|         ) | ||||
| 
 | ||||
|         # store the cancel scope and portal for later cancellation or | ||||
|         # retstart if needed. | ||||
|         self.service_ctxs[name] = (cs, ipc_ctx, portal, complete) | ||||
|         return ( | ||||
|             cs, | ||||
|             ipc_ctx, | ||||
|             started, | ||||
|         ) | ||||
| 
 | ||||
|     async def start_service( | ||||
|         self, | ||||
|         daemon_name: str, | ||||
|         ctx_ep: Callable,  # kwargs must `partial`-ed in! | ||||
|         # ^TODO, type for `@tractor.context` deco-ed funcs! | ||||
| 
 | ||||
|         debug_mode: bool = False, | ||||
|         **start_actor_kwargs, | ||||
| 
 | ||||
|     ) -> Context: | ||||
|         ''' | ||||
|         Start new subactor and schedule a supervising "service task" | ||||
|         in it which explicitly defines the sub's lifetime. | ||||
| 
 | ||||
|         "Service daemon subactors" are cancelled (and thus | ||||
|         terminated) using the paired `.cancel_service()`. | ||||
| 
 | ||||
|         Effectively this API can be used to manage "service daemons" | ||||
|         spawned under a single parent actor with supervision | ||||
|         semantics equivalent to a one-cancels-one style actor-nursery | ||||
|         or "(subactor) task manager" where each subprocess's (and | ||||
|         thus its embedded actor runtime) lifetime is synced to that | ||||
|         of the remotely spawned task defined by `ctx_ep`. | ||||
| 
 | ||||
|         The funcionality can be likened to a "daemonized" version of | ||||
|         `.hilevel.worker.run_in_actor()` but with supervision | ||||
|         controls offered by `tractor.Context` where the main/root | ||||
|         remotely scheduled `trio.Task` invoking `ctx_ep` determines | ||||
|         the underlying subactor's lifetime. | ||||
| 
 | ||||
|         ''' | ||||
|         entry: tuple|None = self.service_ctxs.get(daemon_name) | ||||
|         if entry: | ||||
|             (cs, sub_ctx, portal, complete) = entry | ||||
|             return sub_ctx | ||||
| 
 | ||||
|         if daemon_name not in self.service_ctxs: | ||||
|             portal: Portal = await self.an.start_actor( | ||||
|                 daemon_name, | ||||
|                 debug_mode=(  # maybe set globally during allocate | ||||
|                     debug_mode | ||||
|                     or | ||||
|                     self.debug_mode | ||||
|                 ), | ||||
|                 **start_actor_kwargs, | ||||
|             ) | ||||
|             ctx_kwargs: dict[str, Any] = {} | ||||
|             if isinstance(ctx_ep, functools.partial): | ||||
|                 ctx_kwargs: dict[str, Any] = ctx_ep.keywords | ||||
|                 ctx_ep: Callable = ctx_ep.func | ||||
| 
 | ||||
|             ( | ||||
|                 cs, | ||||
|                 sub_ctx, | ||||
|                 started, | ||||
|             ) = await self.start_service_ctx( | ||||
|                 name=daemon_name, | ||||
|                 portal=portal, | ||||
|                 ctx_fn=ctx_ep, | ||||
|                 **ctx_kwargs, | ||||
|             ) | ||||
| 
 | ||||
|             return sub_ctx | ||||
| 
 | ||||
|     async def cancel_service( | ||||
|         self, | ||||
|         name: str, | ||||
| 
 | ||||
|     ) -> Any: | ||||
|         ''' | ||||
|         Cancel the service task and actor for the given ``name``. | ||||
| 
 | ||||
|         ''' | ||||
|         log.info(f'Cancelling `pikerd` service {name}') | ||||
|         cs, sub_ctx, portal, complete = self.service_ctxs[name] | ||||
| 
 | ||||
|         # cs.cancel() | ||||
|         await sub_ctx.cancel() | ||||
|         await complete.wait() | ||||
| 
 | ||||
|         if name in self.service_ctxs: | ||||
|             # TODO: custom err? | ||||
|             # raise ServiceError( | ||||
|             raise RuntimeError( | ||||
|                 f'Service actor for {name} not terminated and/or unknown?' | ||||
|             ) | ||||
| 
 | ||||
|         # assert name not in self.service_ctxs, \ | ||||
|         #     f'Serice task for {name} not terminated?' | ||||
|  | @ -258,20 +258,28 @@ class ActorContextInfo(Mapping): | |||
| 
 | ||||
| 
 | ||||
| def get_logger( | ||||
| 
 | ||||
|     name: str | None = None, | ||||
|     name: str|None = None, | ||||
|     _root_name: str = _proj_name, | ||||
| 
 | ||||
|     logger: Logger|None = None, | ||||
| 
 | ||||
|     # TODO, using `.config.dictConfig()` api? | ||||
|     # -[ ] SO answer with docs links | ||||
|     #  |_https://stackoverflow.com/questions/7507825/where-is-a-complete-example-of-logging-config-dictconfig | ||||
|     #  |_https://docs.python.org/3/library/logging.config.html#configuration-dictionary-schema | ||||
|     subsys_spec: str|None = None, | ||||
| 
 | ||||
| ) -> StackLevelAdapter: | ||||
|     '''Return the package log or a sub-logger for ``name`` if provided. | ||||
| 
 | ||||
|     ''' | ||||
|     log: Logger | ||||
|     log = rlog = logging.getLogger(_root_name) | ||||
|     log = rlog = logger or logging.getLogger(_root_name) | ||||
| 
 | ||||
|     if ( | ||||
|         name | ||||
|         and name != _proj_name | ||||
|         and | ||||
|         name != _proj_name | ||||
|     ): | ||||
| 
 | ||||
|         # NOTE: for handling for modules that use ``get_logger(__name__)`` | ||||
|  | @ -283,7 +291,7 @@ def get_logger( | |||
|         #   since in python the {filename} is always this same | ||||
|         #   module-file. | ||||
| 
 | ||||
|         sub_name: None | str = None | ||||
|         sub_name: None|str = None | ||||
|         rname, _, sub_name = name.partition('.') | ||||
|         pkgpath, _, modfilename = sub_name.rpartition('.') | ||||
| 
 | ||||
|  | @ -306,7 +314,10 @@ def get_logger( | |||
| 
 | ||||
|     # add our actor-task aware adapter which will dynamically look up | ||||
|     # the actor and task names at each log emit | ||||
|     logger = StackLevelAdapter(log, ActorContextInfo()) | ||||
|     logger = StackLevelAdapter( | ||||
|         log, | ||||
|         ActorContextInfo(), | ||||
|     ) | ||||
| 
 | ||||
|     # additional levels | ||||
|     for name, val in CUSTOM_LEVELS.items(): | ||||
|  | @ -319,15 +330,25 @@ def get_logger( | |||
| 
 | ||||
| 
 | ||||
| def get_console_log( | ||||
|     level: str | None = None, | ||||
|     level: str|None = None, | ||||
|     logger: Logger|None = None, | ||||
|     **kwargs, | ||||
| ) -> LoggerAdapter: | ||||
|     '''Get the package logger and enable a handler which writes to stderr. | ||||
| 
 | ||||
|     Yeah yeah, i know we can use ``DictConfig``. You do it. | ||||
| ) -> LoggerAdapter: | ||||
|     ''' | ||||
|     log = get_logger(**kwargs)  # our root logger | ||||
|     logger = log.logger | ||||
|     Get a `tractor`-style logging instance: a `Logger` wrapped in | ||||
|     a `StackLevelAdapter` which injects various concurrency-primitive | ||||
|     (process, thread, task) fields and enables a `StreamHandler` that | ||||
|     writes on stderr using `colorlog` formatting. | ||||
| 
 | ||||
|     Yeah yeah, i know we can use `logging.config.dictConfig()`. You do it. | ||||
| 
 | ||||
|     ''' | ||||
|     log = get_logger( | ||||
|         logger=logger, | ||||
|         **kwargs | ||||
|     )  # set a root logger | ||||
|     logger: Logger = log.logger | ||||
| 
 | ||||
|     if not level: | ||||
|         return log | ||||
|  | @ -346,9 +367,13 @@ def get_console_log( | |||
|             None, | ||||
|         ) | ||||
|     ): | ||||
|         fmt = LOG_FORMAT | ||||
|         # if logger: | ||||
|         #     fmt = None | ||||
| 
 | ||||
|         handler = StreamHandler() | ||||
|         formatter = colorlog.ColoredFormatter( | ||||
|             LOG_FORMAT, | ||||
|             fmt=fmt, | ||||
|             datefmt=DATE_FORMAT, | ||||
|             log_colors=STD_PALETTE, | ||||
|             secondary_log_colors=BOLD_PALETTE, | ||||
|  | @ -365,7 +390,7 @@ def get_loglevel() -> str: | |||
| 
 | ||||
| 
 | ||||
| # global module logger for tractor itself | ||||
| log = get_logger('tractor') | ||||
| log: StackLevelAdapter = get_logger('tractor') | ||||
| 
 | ||||
| 
 | ||||
| def at_least_level( | ||||
|  |  | |||
|  | @ -33,6 +33,7 @@ from ._codec import ( | |||
| 
 | ||||
|     apply_codec as apply_codec, | ||||
|     mk_codec as mk_codec, | ||||
|     mk_dec as mk_dec, | ||||
|     MsgCodec as MsgCodec, | ||||
|     MsgDec as MsgDec, | ||||
|     current_codec as current_codec, | ||||
|  |  | |||
|  | @ -41,8 +41,10 @@ import textwrap | |||
| from typing import ( | ||||
|     Any, | ||||
|     Callable, | ||||
|     Protocol, | ||||
|     Type, | ||||
|     TYPE_CHECKING, | ||||
|     TypeVar, | ||||
|     Union, | ||||
| ) | ||||
| from types import ModuleType | ||||
|  | @ -59,6 +61,7 @@ from tractor.msg.pretty_struct import Struct | |||
| from tractor.msg.types import ( | ||||
|     mk_msg_spec, | ||||
|     MsgType, | ||||
|     PayloadMsg, | ||||
| ) | ||||
| from tractor.log import get_logger | ||||
| 
 | ||||
|  | @ -78,6 +81,7 @@ class MsgDec(Struct): | |||
| 
 | ||||
|     ''' | ||||
|     _dec: msgpack.Decoder | ||||
|     # _ext_types_box: Struct|None = None | ||||
| 
 | ||||
|     @property | ||||
|     def dec(self) -> msgpack.Decoder: | ||||
|  | @ -177,19 +181,126 @@ class MsgDec(Struct): | |||
| 
 | ||||
| 
 | ||||
| def mk_dec( | ||||
|     spec: Union[Type[Struct]]|Any = Any, | ||||
|     spec: Union[Type[Struct]]|Type|None, | ||||
| 
 | ||||
|     # NOTE, required for ad-hoc type extensions to the underlying | ||||
|     # serialization proto (which is default `msgpack`), | ||||
|     # https://jcristharif.com/msgspec/extending.html#mapping-to-from-native-types | ||||
|     dec_hook: Callable|None = None, | ||||
|     ext_types: list[Type]|None = None, | ||||
| 
 | ||||
| ) -> MsgDec: | ||||
|     ''' | ||||
|     Create an IPC msg decoder, a slightly higher level wrapper around | ||||
|     a `msgspec.msgpack.Decoder` which provides, | ||||
| 
 | ||||
|     - easier introspection of the underlying type spec via | ||||
|       the `.spec` and `.spec_str` attrs, | ||||
|     - `.hook` access to the `Decoder.dec_hook()`, | ||||
|     - automatic custom extension-types decode support when | ||||
|       `dec_hook()` is provided such that any `PayloadMsg.pld` tagged | ||||
|       as a type from from `ext_types` (presuming the `MsgCodec.encode()` also used | ||||
|       a `.enc_hook()`) is processed and constructed by a `PldRx` implicitily. | ||||
| 
 | ||||
|     NOTE, as mentioned a `MsgDec` is normally used for `PayloadMsg.pld: PayloadT` field | ||||
|     decoding inside an IPC-ctx-oriented `PldRx`. | ||||
| 
 | ||||
|     ''' | ||||
|     if ( | ||||
|         spec is None | ||||
|         and | ||||
|         ext_types is None | ||||
|     ): | ||||
|         raise TypeError( | ||||
|             f'MIssing type-`spec` for msg decoder!\n' | ||||
|             f'\n' | ||||
|             f'`spec=None` is **only** permitted is if custom extension types ' | ||||
|             f'are provided via `ext_types`, meaning it must be non-`None`.\n' | ||||
|             f'\n' | ||||
|             f'In this case it is presumed that only the `ext_types`, ' | ||||
|             f'which much be handled by a paired `dec_hook()`, ' | ||||
|             f'will be permitted within the payload type-`spec`!\n' | ||||
|             f'\n' | ||||
|             f'spec = {spec!r}\n' | ||||
|             f'dec_hook = {dec_hook!r}\n' | ||||
|             f'ext_types = {ext_types!r}\n' | ||||
|         ) | ||||
| 
 | ||||
|     if dec_hook: | ||||
|         if ext_types is None: | ||||
|             raise TypeError( | ||||
|                 f'If extending the serializable types with a custom decode hook (`dec_hook()`), ' | ||||
|                 f'you must also provide the expected type set that the hook will handle ' | ||||
|                 f'via a `ext_types: Union[Type]|None = None` argument!\n' | ||||
|                 f'\n' | ||||
|                 f'dec_hook = {dec_hook!r}\n' | ||||
|                 f'ext_types = {ext_types!r}\n' | ||||
|             ) | ||||
| 
 | ||||
|         # XXX, i *thought* we would require a boxing struct as per docs, | ||||
|         # https://jcristharif.com/msgspec/extending.html#mapping-to-from-native-types | ||||
|         # |_ see comment, | ||||
|         #  > Note that typed deserialization is required for | ||||
|         #  > successful roundtripping here, so we pass `MyMessage` to | ||||
|         #  > `Decoder`. | ||||
|         # | ||||
|         # BUT, turns out as long as you spec a union with `Raw` it | ||||
|         # will work? kk B) | ||||
|         # | ||||
|         # maybe_box_struct = mk_boxed_ext_struct(ext_types) | ||||
|         spec = Raw | Union[*ext_types] | ||||
| 
 | ||||
|     return MsgDec( | ||||
|         _dec=msgpack.Decoder( | ||||
|             type=spec,  # like `MsgType[Any]` | ||||
|             dec_hook=dec_hook, | ||||
|         ) | ||||
|         ), | ||||
|     ) | ||||
| 
 | ||||
| 
 | ||||
| # TODO? remove since didn't end up needing this? | ||||
| def mk_boxed_ext_struct( | ||||
|     ext_types: list[Type], | ||||
| ) -> Struct: | ||||
|     # NOTE, originally was to wrap non-msgpack-supported "extension | ||||
|     # types" in a field-typed boxing struct, see notes around the | ||||
|     # `dec_hook()` branch in `mk_dec()`. | ||||
|     ext_types_union = Union[*ext_types] | ||||
|     repr_ext_types_union: str = ( | ||||
|         str(ext_types_union) | ||||
|         or | ||||
|         "|".join(ext_types) | ||||
|     ) | ||||
|     BoxedExtType = msgspec.defstruct( | ||||
|         f'BoxedExts[{repr_ext_types_union}]', | ||||
|         fields=[ | ||||
|             ('boxed', ext_types_union), | ||||
|         ], | ||||
|     ) | ||||
|     return BoxedExtType | ||||
| 
 | ||||
| 
 | ||||
| def unpack_spec_types( | ||||
|     spec: Union[Type]|Type, | ||||
| ) -> set[Type]: | ||||
|     ''' | ||||
|     Given an input type-`spec`, either a lone type | ||||
|     or a `Union` of types (like `str|int|MyThing`), | ||||
|     return a set of individual types. | ||||
| 
 | ||||
|     When `spec` is not a type-union returns `{spec,}`. | ||||
| 
 | ||||
|     ''' | ||||
|     spec_subtypes: set[Union[Type]] = set( | ||||
|          getattr( | ||||
|              spec, | ||||
|              '__args__', | ||||
|              {spec,}, | ||||
|          ) | ||||
|     ) | ||||
|     return spec_subtypes | ||||
| 
 | ||||
| 
 | ||||
| def mk_msgspec_table( | ||||
|     dec: msgpack.Decoder, | ||||
|     msg: MsgType|None = None, | ||||
|  | @ -227,6 +338,13 @@ def pformat_msgspec( | |||
|     join_char: str = '\n', | ||||
| 
 | ||||
| ) -> str: | ||||
|     ''' | ||||
|     Pretty `str` format the `msgspec.msgpack.Decoder.type` attribute | ||||
|     for display in (console) log messages as a nice (maybe multiline) | ||||
|     presentation of all supported `Struct`s (subtypes) available for | ||||
|     typed decoding. | ||||
| 
 | ||||
|     ''' | ||||
|     dec: msgpack.Decoder = getattr(codec, 'dec', codec) | ||||
|     return join_char.join( | ||||
|         mk_msgspec_table( | ||||
|  | @ -260,6 +378,8 @@ class MsgCodec(Struct): | |||
|     _dec: msgpack.Decoder | ||||
|     _pld_spec: Type[Struct]|Raw|Any | ||||
| 
 | ||||
|     # _ext_types_box: Struct|None = None | ||||
| 
 | ||||
|     def __repr__(self) -> str: | ||||
|         speclines: str = textwrap.indent( | ||||
|             pformat_msgspec(codec=self), | ||||
|  | @ -326,12 +446,15 @@ class MsgCodec(Struct): | |||
| 
 | ||||
|     def encode( | ||||
|         self, | ||||
|         py_obj: Any, | ||||
|         py_obj: Any|PayloadMsg, | ||||
| 
 | ||||
|         use_buf: bool = False, | ||||
|         # ^-XXX-^ uhh why am i getting this? | ||||
|         # |_BufferError: Existing exports of data: object cannot be re-sized | ||||
| 
 | ||||
|         as_ext_type: bool = False, | ||||
|         hide_tb: bool = True, | ||||
| 
 | ||||
|     ) -> bytes: | ||||
|         ''' | ||||
|         Encode input python objects to `msgpack` bytes for | ||||
|  | @ -341,11 +464,46 @@ class MsgCodec(Struct): | |||
|         https://jcristharif.com/msgspec/perf-tips.html#reusing-an-output-buffer | ||||
| 
 | ||||
|         ''' | ||||
|         __tracebackhide__: bool = hide_tb | ||||
|         if use_buf: | ||||
|             self._enc.encode_into(py_obj, self._buf) | ||||
|             return self._buf | ||||
|         else: | ||||
|             return self._enc.encode(py_obj) | ||||
| 
 | ||||
|         return self._enc.encode(py_obj) | ||||
|         # try: | ||||
|         #     return self._enc.encode(py_obj) | ||||
|         # except TypeError as typerr: | ||||
|         #     typerr.add_note( | ||||
|         #         '|_src error from `msgspec`' | ||||
|         #         # f'|_{self._enc.encode!r}' | ||||
|         #     ) | ||||
|         #     raise typerr | ||||
| 
 | ||||
|         # TODO! REMOVE once i'm confident we won't ever need it! | ||||
|         # | ||||
|         # box: Struct = self._ext_types_box | ||||
|         # if ( | ||||
|         #     as_ext_type | ||||
|         #     or | ||||
|         #     ( | ||||
|         #         # XXX NOTE, auto-detect if the input type | ||||
|         #         box | ||||
|         #         and | ||||
|         #         (ext_types := unpack_spec_types( | ||||
|         #             spec=box.__annotations__['boxed']) | ||||
|         #         ) | ||||
|         #     ) | ||||
|         # ): | ||||
|         #     match py_obj: | ||||
|         #         # case PayloadMsg(pld=pld) if ( | ||||
|         #         #     type(pld) in ext_types | ||||
|         #         # ): | ||||
|         #         #     py_obj.pld = box(boxed=py_obj) | ||||
|         #         #     breakpoint() | ||||
|         #         case _ if ( | ||||
|         #             type(py_obj) in ext_types | ||||
|         #         ): | ||||
|         #             py_obj = box(boxed=py_obj) | ||||
| 
 | ||||
|     @property | ||||
|     def dec(self) -> msgpack.Decoder: | ||||
|  | @ -365,21 +523,30 @@ class MsgCodec(Struct): | |||
|         return self._dec.decode(msg) | ||||
| 
 | ||||
| 
 | ||||
| # [x] TODO: a sub-decoder system as well? => No! | ||||
| # ?TODO? time to remove this finally? | ||||
| # | ||||
| # -[x] TODO: a sub-decoder system as well? | ||||
| # => No! already re-architected to include a "payload-receiver" | ||||
| #   now found in `._ops`. | ||||
| # | ||||
| # -[x] do we still want to try and support the sub-decoder with | ||||
| # `.Raw` technique in the case that the `Generic` approach gives | ||||
| # future grief? | ||||
| # => NO, since we went with the `PldRx` approach instead B) | ||||
| # => well YES but NO, since we went with the `PldRx` approach | ||||
| #   instead! | ||||
| # | ||||
| # IF however you want to see the code that was staged for this | ||||
| # from wayyy back, see the pure removal commit. | ||||
| 
 | ||||
| 
 | ||||
| def mk_codec( | ||||
|     # struct type unions set for `Decoder` | ||||
|     # https://jcristharif.com/msgspec/structs.html#tagged-unions | ||||
|     ipc_pld_spec: Union[Type[Struct]]|Any = Any, | ||||
|     ipc_pld_spec: Union[Type[Struct]]|Any|Raw = Raw, | ||||
|     # tagged-struct-types-union set for `Decoder`ing of payloads, as | ||||
|     # per https://jcristharif.com/msgspec/structs.html#tagged-unions. | ||||
|     # NOTE that the default `Raw` here **is very intentional** since | ||||
|     # the `PldRx._pld_dec: MsgDec` is responsible for per ipc-ctx-task | ||||
|     # decoding of msg-specs defined by the user as part of **their** | ||||
|     # `tractor` "app's" type-limited IPC msg-spec. | ||||
| 
 | ||||
|     # TODO: offering a per-msg(-field) type-spec such that | ||||
|     # the fields can be dynamically NOT decoded and left as `Raw` | ||||
|  | @ -392,13 +559,18 @@ def mk_codec( | |||
| 
 | ||||
|     libname: str = 'msgspec', | ||||
| 
 | ||||
|     # proxy as `Struct(**kwargs)` for ad-hoc type extensions | ||||
|     # settings for encoding-to-send extension-types, | ||||
|     # https://jcristharif.com/msgspec/extending.html#mapping-to-from-native-types | ||||
|     # ------ - ------ | ||||
|     dec_hook: Callable|None = None, | ||||
|     # dec_hook: Callable|None = None, | ||||
|     enc_hook: Callable|None = None, | ||||
|     # ------ - ------ | ||||
|     ext_types: list[Type]|None = None, | ||||
| 
 | ||||
|     # optionally provided msg-decoder from which we pull its, | ||||
|     # |_.dec_hook() | ||||
|     # |_.type | ||||
|     ext_dec: MsgDec|None = None | ||||
|     # | ||||
|     # ?TODO? other params we might want to support | ||||
|     # Encoder: | ||||
|     # write_buffer_size=write_buffer_size, | ||||
|     # | ||||
|  | @ -412,26 +584,44 @@ def mk_codec( | |||
|     `msgspec` ;). | ||||
| 
 | ||||
|     ''' | ||||
|     # (manually) generate a msg-payload-spec for all relevant | ||||
|     # god-boxing-msg subtypes, parameterizing the `PayloadMsg.pld: PayloadT` | ||||
|     # for the decoder such that all sub-type msgs in our SCIPP | ||||
|     # will automatically decode to a type-"limited" payload (`Struct`) | ||||
|     # object (set). | ||||
|     pld_spec = ipc_pld_spec | ||||
|     if enc_hook: | ||||
|         if not ext_types: | ||||
|             raise TypeError( | ||||
|                 f'If extending the serializable types with a custom encode hook (`enc_hook()`), ' | ||||
|                 f'you must also provide the expected type set that the hook will handle ' | ||||
|                 f'via a `ext_types: Union[Type]|None = None` argument!\n' | ||||
|                 f'\n' | ||||
|                 f'enc_hook = {enc_hook!r}\n' | ||||
|                 f'ext_types = {ext_types!r}\n' | ||||
|             ) | ||||
| 
 | ||||
|     dec_hook: Callable|None = None | ||||
|     if ext_dec: | ||||
|         dec: msgspec.Decoder = ext_dec.dec | ||||
|         dec_hook = dec.dec_hook | ||||
|         pld_spec |= dec.type | ||||
|         if ext_types: | ||||
|             pld_spec |= Union[*ext_types] | ||||
| 
 | ||||
|     # (manually) generate a msg-spec (how appropes) for all relevant | ||||
|     # payload-boxing-struct-msg-types, parameterizing the | ||||
|     # `PayloadMsg.pld: PayloadT` for the decoder such that all msgs | ||||
|     # in our SC-RPC-protocol will automatically decode to | ||||
|     # a type-"limited" payload (`Struct`) object (set). | ||||
|     ( | ||||
|         ipc_msg_spec, | ||||
|         msg_types, | ||||
|     ) = mk_msg_spec( | ||||
|         payload_type_union=ipc_pld_spec, | ||||
|         payload_type_union=pld_spec, | ||||
|     ) | ||||
|     assert len(ipc_msg_spec.__args__) == len(msg_types) | ||||
|     assert ipc_msg_spec | ||||
| 
 | ||||
|     # TODO: use this shim instead? | ||||
|     # bc.. unification, err somethin? | ||||
|     # dec: MsgDec = mk_dec( | ||||
|     #     spec=ipc_msg_spec, | ||||
|     #     dec_hook=dec_hook, | ||||
|     # ) | ||||
|     msg_spec_types: set[Type] = unpack_spec_types(ipc_msg_spec) | ||||
|     assert ( | ||||
|         len(ipc_msg_spec.__args__) == len(msg_types) | ||||
|         and | ||||
|         len(msg_spec_types) == len(msg_types) | ||||
|     ) | ||||
| 
 | ||||
|     dec = msgpack.Decoder( | ||||
|         type=ipc_msg_spec, | ||||
|  | @ -440,22 +630,29 @@ def mk_codec( | |||
|     enc = msgpack.Encoder( | ||||
|        enc_hook=enc_hook, | ||||
|     ) | ||||
| 
 | ||||
|     codec = MsgCodec( | ||||
|         _enc=enc, | ||||
|         _dec=dec, | ||||
|         _pld_spec=ipc_pld_spec, | ||||
|         _pld_spec=pld_spec, | ||||
|     ) | ||||
| 
 | ||||
|     # sanity on expected backend support | ||||
|     assert codec.lib.__name__ == libname | ||||
| 
 | ||||
|     return codec | ||||
| 
 | ||||
| 
 | ||||
| # instance of the default `msgspec.msgpack` codec settings, i.e. | ||||
| # no custom structs, hooks or other special types. | ||||
| _def_msgspec_codec: MsgCodec = mk_codec(ipc_pld_spec=Any) | ||||
| # | ||||
| # XXX NOTE XXX, this will break our `Context.start()` call! | ||||
| # | ||||
| # * by default we roundtrip the started pld-`value` and if you apply | ||||
| #   this codec (globally anyway with `apply_codec()`) then the | ||||
| #   `roundtripped` value will include a non-`.pld: Raw` which will | ||||
| #   then type-error on the consequent `._ops.validte_payload_msg()`.. | ||||
| # | ||||
| _def_msgspec_codec: MsgCodec = mk_codec( | ||||
|     ipc_pld_spec=Any, | ||||
| ) | ||||
| 
 | ||||
| # The built-in IPC `Msg` spec. | ||||
| # Our composing "shuttle" protocol which allows `tractor`-app code | ||||
|  | @ -463,13 +660,13 @@ _def_msgspec_codec: MsgCodec = mk_codec(ipc_pld_spec=Any) | |||
| # https://jcristharif.com/msgspec/supported-types.html | ||||
| # | ||||
| _def_tractor_codec: MsgCodec = mk_codec( | ||||
|     # TODO: use this for debug mode locking prot? | ||||
|     # ipc_pld_spec=Any, | ||||
|     ipc_pld_spec=Raw, | ||||
|     ipc_pld_spec=Raw,  # XXX should be default righ!? | ||||
| ) | ||||
| # TODO: IDEALLY provides for per-`trio.Task` specificity of the | ||||
| 
 | ||||
| # -[x] TODO, IDEALLY provides for per-`trio.Task` specificity of the | ||||
| # IPC msging codec used by the transport layer when doing | ||||
| # `Channel.send()/.recv()` of wire data. | ||||
| # => impled as our `PldRx` which is `Context` scoped B) | ||||
| 
 | ||||
| # ContextVar-TODO: DIDN'T WORK, kept resetting in every new task to default!? | ||||
| # _ctxvar_MsgCodec: ContextVar[MsgCodec] = ContextVar( | ||||
|  | @ -546,17 +743,6 @@ def apply_codec( | |||
|     ) | ||||
|     token: Token = var.set(codec) | ||||
| 
 | ||||
|     # ?TODO? for TreeVar approach which copies from the | ||||
|     # cancel-scope of the prior value, NOT the prior task | ||||
|     # See the docs: | ||||
|     # - https://tricycle.readthedocs.io/en/latest/reference.html#tree-variables | ||||
|     # - https://github.com/oremanj/tricycle/blob/master/tricycle/_tests/test_tree_var.py | ||||
|     #   ^- see docs for @cm `.being()` API | ||||
|     # with _ctxvar_MsgCodec.being(codec): | ||||
|     #     new = _ctxvar_MsgCodec.get() | ||||
|     #     assert new is codec | ||||
|     #     yield codec | ||||
| 
 | ||||
|     try: | ||||
|         yield var.get() | ||||
|     finally: | ||||
|  | @ -567,6 +753,19 @@ def apply_codec( | |||
|         ) | ||||
|         assert var.get() is orig | ||||
| 
 | ||||
|     # ?TODO? for TreeVar approach which copies from the | ||||
|     # cancel-scope of the prior value, NOT the prior task | ||||
|     # | ||||
|     # See the docs: | ||||
|     # - https://tricycle.readthedocs.io/en/latest/reference.html#tree-variables | ||||
|     # - https://github.com/oremanj/tricycle/blob/master/tricycle/_tests/test_tree_var.py | ||||
|     #   ^- see docs for @cm `.being()` API | ||||
|     # | ||||
|     # with _ctxvar_MsgCodec.being(codec): | ||||
|     #     new = _ctxvar_MsgCodec.get() | ||||
|     #     assert new is codec | ||||
|     #     yield codec | ||||
| 
 | ||||
| 
 | ||||
| def current_codec() -> MsgCodec: | ||||
|     ''' | ||||
|  | @ -586,6 +785,7 @@ def limit_msg_spec( | |||
|     # -> related to the `MsgCodec._payload_decs` stuff above.. | ||||
|     # tagged_structs: list[Struct]|None = None, | ||||
| 
 | ||||
|     hide_tb: bool = True, | ||||
|     **codec_kwargs, | ||||
| 
 | ||||
| ) -> MsgCodec: | ||||
|  | @ -596,7 +796,7 @@ def limit_msg_spec( | |||
|     for all IPC contexts in use by the current `trio.Task`. | ||||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = True | ||||
|     __tracebackhide__: bool = hide_tb | ||||
|     curr_codec: MsgCodec = current_codec() | ||||
|     msgspec_codec: MsgCodec = mk_codec( | ||||
|         ipc_pld_spec=payload_spec, | ||||
|  | @ -630,31 +830,57 @@ def limit_msg_spec( | |||
| #         # import pdbp; pdbp.set_trace() | ||||
| #         assert ext_codec.pld_spec == extended_spec | ||||
| #         yield ext_codec | ||||
| # | ||||
| # ^-TODO-^ is it impossible to make something like this orr!? | ||||
| 
 | ||||
| # TODO: make an auto-custom hook generator from a set of input custom | ||||
| # types? | ||||
| # -[ ] below is a proto design using a `TypeCodec` idea? | ||||
| # | ||||
| # type var for the expected interchange-lib's | ||||
| # IPC-transport type when not available as a built-in | ||||
| # serialization output. | ||||
| WireT = TypeVar('WireT') | ||||
| 
 | ||||
| 
 | ||||
| # TODO: make something similar to this inside `._codec` such that | ||||
| # user can just pass a type table of some sort? | ||||
| # -[ ] we would need to decode all msgs to `pretty_struct.Struct` | ||||
| #     and then call `.to_dict()` on them? | ||||
| # -[x] we're going to need to re-impl all the stuff changed in the | ||||
| #    runtime port such that it can handle dicts or `Msg`s? | ||||
| # | ||||
| # def mk_dict_msg_codec_hooks() -> tuple[Callable, Callable]: | ||||
| #     ''' | ||||
| #     Deliver a `enc_hook()`/`dec_hook()` pair which does | ||||
| #     manual convertion from our above native `Msg` set | ||||
| #     to `dict` equivalent (wire msgs) in order to keep legacy compat | ||||
| #     with the original runtime implementation. | ||||
| # | ||||
| #     Note: this is is/was primarly used while moving the core | ||||
| #     runtime over to using native `Msg`-struct types wherein we | ||||
| #     start with the send side emitting without loading | ||||
| #     a typed-decoder and then later flipping the switch over to | ||||
| #     load to the native struct types once all runtime usage has | ||||
| #     been adjusted appropriately. | ||||
| # | ||||
| #     ''' | ||||
| #     return ( | ||||
| #         # enc_to_dict, | ||||
| #         dec_from_dict, | ||||
| #     ) | ||||
| # TODO: some kinda (decorator) API for built-in subtypes | ||||
| # that builds this implicitly by inspecting the `mro()`? | ||||
| class TypeCodec(Protocol): | ||||
|     ''' | ||||
|     A per-custom-type wire-transport serialization translator | ||||
|     description type. | ||||
| 
 | ||||
|     ''' | ||||
|     src_type: Type | ||||
|     wire_type: WireT | ||||
| 
 | ||||
|     def encode(obj: Type) -> WireT: | ||||
|         ... | ||||
| 
 | ||||
|     def decode( | ||||
|         obj_type: Type[WireT], | ||||
|         obj: WireT, | ||||
|     ) -> Type: | ||||
|         ... | ||||
| 
 | ||||
| 
 | ||||
| class MsgpackTypeCodec(TypeCodec): | ||||
|     ... | ||||
| 
 | ||||
| 
 | ||||
| def mk_codec_hooks( | ||||
|     type_codecs: list[TypeCodec], | ||||
| 
 | ||||
| ) -> tuple[Callable, Callable]: | ||||
|     ''' | ||||
|     Deliver a `enc_hook()`/`dec_hook()` pair which handle | ||||
|     manual convertion from an input `Type` set such that whenever | ||||
|     the `TypeCodec.filter()` predicate matches the | ||||
|     `TypeCodec.decode()` is called on the input native object by | ||||
|     the `dec_hook()` and whenever the | ||||
|     `isiinstance(obj, TypeCodec.type)` matches against an | ||||
|     `enc_hook(obj=obj)` the return value is taken from a | ||||
|     `TypeCodec.encode(obj)` callback. | ||||
| 
 | ||||
|     ''' | ||||
|     ... | ||||
|  |  | |||
|  | @ -0,0 +1,94 @@ | |||
| # tractor: structured concurrent "actors". | ||||
| # Copyright 2018-eternity Tyler Goodlet. | ||||
| 
 | ||||
| # This program is free software: you can redistribute it and/or modify | ||||
| # it under the terms of the GNU Affero General Public License as published by | ||||
| # the Free Software Foundation, either version 3 of the License, or | ||||
| # (at your option) any later version. | ||||
| 
 | ||||
| # This program is distributed in the hope that it will be useful, | ||||
| # but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
| # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | ||||
| # GNU Affero General Public License for more details. | ||||
| 
 | ||||
| # You should have received a copy of the GNU Affero General Public License | ||||
| # along with this program.  If not, see <https://www.gnu.org/licenses/>. | ||||
| 
 | ||||
| ''' | ||||
| Type-extension-utils for codec-ing (python) objects not | ||||
| covered by the `msgspec.msgpack` protocol. | ||||
| 
 | ||||
| See the various API docs from `msgspec`. | ||||
| 
 | ||||
| extending from native types, | ||||
| - https://jcristharif.com/msgspec/extending.html#mapping-to-from-native-types | ||||
| 
 | ||||
| converters, | ||||
| - https://jcristharif.com/msgspec/converters.html | ||||
| - https://jcristharif.com/msgspec/api.html#msgspec.convert | ||||
| 
 | ||||
| `Raw` fields, | ||||
| - https://jcristharif.com/msgspec/api.html#raw | ||||
| - support for `.convert()` and `Raw`, | ||||
|   |_ https://jcristharif.com/msgspec/changelog.html | ||||
| 
 | ||||
| ''' | ||||
| from types import ( | ||||
|     ModuleType, | ||||
| ) | ||||
| import typing | ||||
| from typing import ( | ||||
|     Type, | ||||
|     Union, | ||||
| ) | ||||
| 
 | ||||
| def dec_type_union( | ||||
|     type_names: list[str], | ||||
|     mods: list[ModuleType] = [] | ||||
| ) -> Type|Union[Type]: | ||||
|     ''' | ||||
|     Look up types by name, compile into a list and then create and | ||||
|     return a `typing.Union` from the full set. | ||||
| 
 | ||||
|     ''' | ||||
|     # import importlib | ||||
|     types: list[Type] = [] | ||||
|     for type_name in type_names: | ||||
|         for mod in [ | ||||
|             typing, | ||||
|             # importlib.import_module(__name__), | ||||
|         ] + mods: | ||||
|             if type_ref := getattr( | ||||
|                 mod, | ||||
|                 type_name, | ||||
|                 False, | ||||
|             ): | ||||
|                 types.append(type_ref) | ||||
| 
 | ||||
|     # special case handling only.. | ||||
|     # ipc_pld_spec: Union[Type] = eval( | ||||
|     #     pld_spec_str, | ||||
|     #     {},  # globals | ||||
|     #     {'typing': typing},  # locals | ||||
|     # ) | ||||
| 
 | ||||
|     return Union[*types] | ||||
| 
 | ||||
| 
 | ||||
| def enc_type_union( | ||||
|     union_or_type: Union[Type]|Type, | ||||
| ) -> list[str]: | ||||
|     ''' | ||||
|     Encode a type-union or single type to a list of type-name-strings | ||||
|     ready for IPC interchange. | ||||
| 
 | ||||
|     ''' | ||||
|     type_strs: list[str] = [] | ||||
|     for typ in getattr( | ||||
|         union_or_type, | ||||
|         '__args__', | ||||
|         {union_or_type,}, | ||||
|     ): | ||||
|         type_strs.append(typ.__qualname__) | ||||
| 
 | ||||
|     return type_strs | ||||
|  | @ -50,7 +50,9 @@ from tractor._exceptions import ( | |||
|     _mk_recv_mte, | ||||
|     pack_error, | ||||
| ) | ||||
| from tractor._state import current_ipc_ctx | ||||
| from tractor._state import ( | ||||
|     current_ipc_ctx, | ||||
| ) | ||||
| from ._codec import ( | ||||
|     mk_dec, | ||||
|     MsgDec, | ||||
|  | @ -78,7 +80,7 @@ if TYPE_CHECKING: | |||
| log = get_logger(__name__) | ||||
| 
 | ||||
| 
 | ||||
| _def_any_pldec: MsgDec[Any] = mk_dec() | ||||
| _def_any_pldec: MsgDec[Any] = mk_dec(spec=Any) | ||||
| 
 | ||||
| 
 | ||||
| class PldRx(Struct): | ||||
|  | @ -108,33 +110,11 @@ class PldRx(Struct): | |||
|     # TODO: better to bind it here? | ||||
|     # _rx_mc: trio.MemoryReceiveChannel | ||||
|     _pld_dec: MsgDec | ||||
|     _ctx: Context|None = None | ||||
|     _ipc: Context|MsgStream|None = None | ||||
| 
 | ||||
|     @property | ||||
|     def pld_dec(self) -> MsgDec: | ||||
|         return self._pld_dec | ||||
| 
 | ||||
|     # TODO: a better name? | ||||
|     # -[ ] when would this be used as it avoids needingn to pass the | ||||
|     #   ipc prim to every method | ||||
|     @cm | ||||
|     def wraps_ipc( | ||||
|         self, | ||||
|         ipc_prim: Context|MsgStream, | ||||
| 
 | ||||
|     ) -> PldRx: | ||||
|         ''' | ||||
|         Apply this payload receiver to an IPC primitive type, one | ||||
|         of `Context` or `MsgStream`. | ||||
| 
 | ||||
|         ''' | ||||
|         self._ipc = ipc_prim | ||||
|         try: | ||||
|             yield self | ||||
|         finally: | ||||
|             self._ipc = None | ||||
| 
 | ||||
|     @cm | ||||
|     def limit_plds( | ||||
|         self, | ||||
|  | @ -148,6 +128,10 @@ class PldRx(Struct): | |||
|         exit. | ||||
| 
 | ||||
|         ''' | ||||
|         # TODO, ensure we pull the current `MsgCodec`'s custom | ||||
|         # dec/enc_hook settings as well ? | ||||
|         # -[ ] see `._codec.mk_codec()` inputs | ||||
|         # | ||||
|         orig_dec: MsgDec = self._pld_dec | ||||
|         limit_dec: MsgDec = mk_dec( | ||||
|             spec=spec, | ||||
|  | @ -163,7 +147,7 @@ class PldRx(Struct): | |||
|     def dec(self) -> msgpack.Decoder: | ||||
|         return self._pld_dec.dec | ||||
| 
 | ||||
|     def recv_pld_nowait( | ||||
|     def recv_msg_nowait( | ||||
|         self, | ||||
|         # TODO: make this `MsgStream` compat as well, see above^ | ||||
|         # ipc_prim: Context|MsgStream, | ||||
|  | @ -174,34 +158,95 @@ class PldRx(Struct): | |||
|         hide_tb: bool = False, | ||||
|         **dec_pld_kwargs, | ||||
| 
 | ||||
|     ) -> Any|Raw: | ||||
|     ) -> tuple[ | ||||
|         MsgType[PayloadT], | ||||
|         PayloadT, | ||||
|     ]: | ||||
|         ''' | ||||
|         Attempt to non-blocking receive a message from the `._rx_chan` and | ||||
|         unwrap it's payload delivering the pair to the caller. | ||||
| 
 | ||||
|         ''' | ||||
|         __tracebackhide__: bool = hide_tb | ||||
| 
 | ||||
|         msg: MsgType = ( | ||||
|             ipc_msg | ||||
|             or | ||||
| 
 | ||||
|             # sync-rx msg from underlying IPC feeder (mem-)chan | ||||
|             ipc._rx_chan.receive_nowait() | ||||
|         ) | ||||
|         return self.decode_pld( | ||||
|         pld: PayloadT = self.decode_pld( | ||||
|             msg, | ||||
|             ipc=ipc, | ||||
|             expect_msg=expect_msg, | ||||
|             hide_tb=hide_tb, | ||||
|             **dec_pld_kwargs, | ||||
|         ) | ||||
|         return ( | ||||
|             msg, | ||||
|             pld, | ||||
|         ) | ||||
| 
 | ||||
|     async def recv_msg( | ||||
|         self, | ||||
|         ipc: Context|MsgStream, | ||||
|         expect_msg: MsgType, | ||||
| 
 | ||||
|         # NOTE: ONLY for handling `Stop`-msgs that arrive during | ||||
|         # a call to `drain_to_final_msg()` above! | ||||
|         passthrough_non_pld_msgs: bool = True, | ||||
|         hide_tb: bool = True, | ||||
| 
 | ||||
|         **decode_pld_kwargs, | ||||
| 
 | ||||
|     ) -> tuple[MsgType, PayloadT]: | ||||
|         ''' | ||||
|         Retrieve the next avail IPC msg, decode its payload, and | ||||
|         return the (msg, pld) pair. | ||||
| 
 | ||||
|         ''' | ||||
|         __tracebackhide__: bool = hide_tb | ||||
|         msg: MsgType = await ipc._rx_chan.receive() | ||||
|         match msg: | ||||
|             case Return()|Error(): | ||||
|                 log.runtime( | ||||
|                     f'Rxed final outcome msg\n' | ||||
|                     f'{msg}\n' | ||||
|                 ) | ||||
|             case Stop(): | ||||
|                 log.runtime( | ||||
|                     f'Rxed stream stopped msg\n' | ||||
|                     f'{msg}\n' | ||||
|                 ) | ||||
|                 if passthrough_non_pld_msgs: | ||||
|                     return msg, None | ||||
| 
 | ||||
|         # TODO: is there some way we can inject the decoded | ||||
|         # payload into an existing output buffer for the original | ||||
|         # msg instance? | ||||
|         pld: PayloadT = self.decode_pld( | ||||
|             msg, | ||||
|             ipc=ipc, | ||||
|             expect_msg=expect_msg, | ||||
|             hide_tb=hide_tb, | ||||
| 
 | ||||
|             **decode_pld_kwargs, | ||||
|         ) | ||||
|         return ( | ||||
|             msg, | ||||
|             pld, | ||||
|         ) | ||||
| 
 | ||||
|     async def recv_pld( | ||||
|         self, | ||||
|         ipc: Context|MsgStream, | ||||
|         ipc_msg: MsgType|None = None, | ||||
|         ipc_msg: MsgType[PayloadT]|None = None, | ||||
|         expect_msg: Type[MsgType]|None = None, | ||||
|         hide_tb: bool = True, | ||||
| 
 | ||||
|         **dec_pld_kwargs, | ||||
| 
 | ||||
|     ) -> Any|Raw: | ||||
|     ) -> PayloadT: | ||||
|         ''' | ||||
|         Receive a `MsgType`, then decode and return its `.pld` field. | ||||
| 
 | ||||
|  | @ -213,6 +258,13 @@ class PldRx(Struct): | |||
|             # async-rx msg from underlying IPC feeder (mem-)chan | ||||
|             await ipc._rx_chan.receive() | ||||
|         ) | ||||
|         if ( | ||||
|             type(msg) is Return | ||||
|         ): | ||||
|             log.info( | ||||
|                 f'Rxed final result msg\n' | ||||
|                 f'{msg}\n' | ||||
|             ) | ||||
|         return self.decode_pld( | ||||
|             msg=msg, | ||||
|             ipc=ipc, | ||||
|  | @ -258,6 +310,9 @@ class PldRx(Struct): | |||
|                         f'|_pld={pld!r}\n' | ||||
|                     ) | ||||
|                     return pld | ||||
|                 except TypeError as typerr: | ||||
|                     __tracebackhide__: bool = False | ||||
|                     raise typerr | ||||
| 
 | ||||
|                 # XXX pld-value type failure | ||||
|                 except ValidationError as valerr: | ||||
|  | @ -398,45 +453,6 @@ class PldRx(Struct): | |||
|             __tracebackhide__: bool = False | ||||
|             raise | ||||
| 
 | ||||
|     dec_msg = decode_pld | ||||
| 
 | ||||
|     async def recv_msg_w_pld( | ||||
|         self, | ||||
|         ipc: Context|MsgStream, | ||||
|         expect_msg: MsgType, | ||||
| 
 | ||||
|         # NOTE: generally speaking only for handling `Stop`-msgs that | ||||
|         # arrive during a call to `drain_to_final_msg()` above! | ||||
|         passthrough_non_pld_msgs: bool = True, | ||||
|         hide_tb: bool = True, | ||||
|         **kwargs, | ||||
| 
 | ||||
|     ) -> tuple[MsgType, PayloadT]: | ||||
|         ''' | ||||
|         Retrieve the next avail IPC msg, decode it's payload, and return | ||||
|         the pair of refs. | ||||
| 
 | ||||
|         ''' | ||||
|         __tracebackhide__: bool = hide_tb | ||||
|         msg: MsgType = await ipc._rx_chan.receive() | ||||
| 
 | ||||
|         if passthrough_non_pld_msgs: | ||||
|             match msg: | ||||
|                 case Stop(): | ||||
|                     return msg, None | ||||
| 
 | ||||
|         # TODO: is there some way we can inject the decoded | ||||
|         # payload into an existing output buffer for the original | ||||
|         # msg instance? | ||||
|         pld: PayloadT = self.decode_pld( | ||||
|             msg, | ||||
|             ipc=ipc, | ||||
|             expect_msg=expect_msg, | ||||
|             hide_tb=hide_tb, | ||||
|             **kwargs, | ||||
|         ) | ||||
|         return msg, pld | ||||
| 
 | ||||
| 
 | ||||
| @cm | ||||
| def limit_plds( | ||||
|  | @ -452,11 +468,16 @@ def limit_plds( | |||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = True | ||||
|     curr_ctx: Context|None = current_ipc_ctx() | ||||
|     if curr_ctx is None: | ||||
|         raise RuntimeError( | ||||
|             'No IPC `Context` is active !?\n' | ||||
|             'Did you open `limit_plds()` from outside ' | ||||
|             'a `Portal.open_context()` scope-block?' | ||||
|         ) | ||||
|     try: | ||||
|         curr_ctx: Context = current_ipc_ctx() | ||||
|         rx: PldRx = curr_ctx._pld_rx | ||||
|         orig_pldec: MsgDec = rx.pld_dec | ||||
| 
 | ||||
|         with rx.limit_plds( | ||||
|             spec=spec, | ||||
|             **dec_kwargs, | ||||
|  | @ -466,6 +487,11 @@ def limit_plds( | |||
|                 f'{pldec}\n' | ||||
|             ) | ||||
|             yield pldec | ||||
| 
 | ||||
|     except BaseException: | ||||
|         __tracebackhide__: bool = False | ||||
|         raise | ||||
| 
 | ||||
|     finally: | ||||
|         log.runtime( | ||||
|             'Reverted to previous payload-decoder\n\n' | ||||
|  | @ -519,8 +545,8 @@ async def maybe_limit_plds( | |||
| async def drain_to_final_msg( | ||||
|     ctx: Context, | ||||
| 
 | ||||
|     hide_tb: bool = True, | ||||
|     msg_limit: int = 6, | ||||
|     hide_tb: bool = True, | ||||
| 
 | ||||
| ) -> tuple[ | ||||
|     Return|None, | ||||
|  | @ -549,8 +575,8 @@ async def drain_to_final_msg( | |||
|     even after ctx closure and the `.open_context()` block exit. | ||||
| 
 | ||||
|     ''' | ||||
|     __tracebackhide__: bool = hide_tb | ||||
|     raise_overrun: bool = not ctx._allow_overruns | ||||
|     parent_never_opened_stream: bool = ctx._stream is None | ||||
| 
 | ||||
|     # wait for a final context result by collecting (but | ||||
|     # basically ignoring) any bi-dir-stream msgs still in transit | ||||
|  | @ -559,13 +585,14 @@ async def drain_to_final_msg( | |||
|     result_msg: Return|Error|None = None | ||||
|     while not ( | ||||
|         ctx.maybe_error | ||||
|         and not ctx._final_result_is_set() | ||||
|         and | ||||
|         not ctx._final_result_is_set() | ||||
|     ): | ||||
|         try: | ||||
|             # receive all msgs, scanning for either a final result | ||||
|             # or error; the underlying call should never raise any | ||||
|             # remote error directly! | ||||
|             msg, pld = await ctx._pld_rx.recv_msg_w_pld( | ||||
|             msg, pld = await ctx._pld_rx.recv_msg( | ||||
|                 ipc=ctx, | ||||
|                 expect_msg=Return, | ||||
|                 raise_error=False, | ||||
|  | @ -612,6 +639,11 @@ async def drain_to_final_msg( | |||
|                     ) | ||||
|                     __tracebackhide__: bool = False | ||||
| 
 | ||||
|             else: | ||||
|                 log.cancel( | ||||
|                     f'IPC ctx cancelled externally during result drain ?\n' | ||||
|                     f'{ctx}' | ||||
|                 ) | ||||
|             # CASE 2: mask the local cancelled-error(s) | ||||
|             # only when we are sure the remote error is | ||||
|             # the source cause of this local task's | ||||
|  | @ -643,17 +675,24 @@ async def drain_to_final_msg( | |||
|             case Yield(): | ||||
|                 pre_result_drained.append(msg) | ||||
|                 if ( | ||||
|                     (ctx._stream.closed | ||||
|                      and (reason := 'stream was already closed') | ||||
|                     ) | ||||
|                     or (ctx.cancel_acked | ||||
|                         and (reason := 'ctx cancelled other side') | ||||
|                     ) | ||||
|                     or (ctx._cancel_called | ||||
|                         and (reason := 'ctx called `.cancel()`') | ||||
|                     ) | ||||
|                     or (len(pre_result_drained) > msg_limit | ||||
|                         and (reason := f'"yield" limit={msg_limit}') | ||||
|                     not parent_never_opened_stream | ||||
|                     and ( | ||||
|                         (ctx._stream.closed | ||||
|                          and | ||||
|                          (reason := 'stream was already closed') | ||||
|                         ) or | ||||
|                         (ctx.cancel_acked | ||||
|                             and | ||||
|                             (reason := 'ctx cancelled other side') | ||||
|                         ) | ||||
|                         or (ctx._cancel_called | ||||
|                             and | ||||
|                             (reason := 'ctx called `.cancel()`') | ||||
|                         ) | ||||
|                         or (len(pre_result_drained) > msg_limit | ||||
|                             and | ||||
|                             (reason := f'"yield" limit={msg_limit}') | ||||
|                         ) | ||||
|                     ) | ||||
|                 ): | ||||
|                     log.cancel( | ||||
|  | @ -671,7 +710,7 @@ async def drain_to_final_msg( | |||
|                 # drain up to the `msg_limit` hoping to get | ||||
|                 # a final result or error/ctxc. | ||||
|                 else: | ||||
|                     log.warning( | ||||
|                     report: str = ( | ||||
|                         'Ignoring "yield" msg during `ctx.result()` drain..\n' | ||||
|                         f'<= {ctx.chan.uid}\n' | ||||
|                         f'  |_{ctx._nsf}()\n\n' | ||||
|  | @ -680,6 +719,14 @@ async def drain_to_final_msg( | |||
| 
 | ||||
|                         f'{pretty_struct.pformat(msg)}\n' | ||||
|                     ) | ||||
|                     if parent_never_opened_stream: | ||||
|                         report = ( | ||||
|                             f'IPC ctx never opened stream on {ctx.side!r}-side!\n' | ||||
|                             f'\n' | ||||
|                             # f'{ctx}\n' | ||||
|                         ) + report | ||||
| 
 | ||||
|                     log.warning(report) | ||||
|                     continue | ||||
| 
 | ||||
|             # stream terminated, but no result yet.. | ||||
|  | @ -771,6 +818,7 @@ async def drain_to_final_msg( | |||
|             f'{ctx.outcome}\n' | ||||
|         ) | ||||
| 
 | ||||
|     __tracebackhide__: bool = hide_tb | ||||
|     return ( | ||||
|         result_msg, | ||||
|         pre_result_drained, | ||||
|  | @ -796,8 +844,14 @@ def validate_payload_msg( | |||
|     __tracebackhide__: bool = hide_tb | ||||
|     codec: MsgCodec = current_codec() | ||||
|     msg_bytes: bytes = codec.encode(pld_msg) | ||||
|     roundtripped: Started|None = None | ||||
|     try: | ||||
|         roundtripped: Started = codec.decode(msg_bytes) | ||||
|     except TypeError as typerr: | ||||
|         __tracebackhide__: bool = False | ||||
|         raise typerr | ||||
| 
 | ||||
|     try: | ||||
|         ctx: Context = getattr(ipc, 'ctx', ipc) | ||||
|         pld: PayloadT = ctx.pld_rx.decode_pld( | ||||
|             msg=roundtripped, | ||||
|  | @ -822,6 +876,11 @@ def validate_payload_msg( | |||
|             ) | ||||
|             raise ValidationError(complaint) | ||||
| 
 | ||||
|     # usually due to `.decode()` input type | ||||
|     except TypeError as typerr: | ||||
|         __tracebackhide__: bool = False | ||||
|         raise typerr | ||||
| 
 | ||||
|     # raise any msg type error NO MATTER WHAT! | ||||
|     except ValidationError as verr: | ||||
|         try: | ||||
|  | @ -832,9 +891,13 @@ def validate_payload_msg( | |||
|                 verb_header='Trying to send ', | ||||
|                 is_invalid_payload=True, | ||||
|             ) | ||||
|         except BaseException: | ||||
|         except BaseException as _be: | ||||
|             if not roundtripped: | ||||
|                 raise verr | ||||
| 
 | ||||
|             be = _be | ||||
|             __tracebackhide__: bool = False | ||||
|             raise | ||||
|             raise be | ||||
| 
 | ||||
|         if not raise_mte: | ||||
|             return mte | ||||
|  |  | |||
|  | @ -30,9 +30,9 @@ from msgspec import ( | |||
|     Struct as _Struct, | ||||
|     structs, | ||||
| ) | ||||
| from pprint import ( | ||||
|     saferepr, | ||||
| ) | ||||
| # from pprint import ( | ||||
| #     saferepr, | ||||
| # ) | ||||
| 
 | ||||
| from tractor.log import get_logger | ||||
| 
 | ||||
|  | @ -75,8 +75,8 @@ class DiffDump(UserList): | |||
|         for k, left, right in self: | ||||
|             repstr += ( | ||||
|                 f'({k},\n' | ||||
|                 f'\t{repr(left)},\n' | ||||
|                 f'\t{repr(right)},\n' | ||||
|                 f' |_{repr(left)},\n' | ||||
|                 f' |_{repr(right)},\n' | ||||
|                 ')\n' | ||||
|             ) | ||||
|         repstr += ']\n' | ||||
|  | @ -144,15 +144,22 @@ def pformat( | |||
|                 field_indent=indent + field_indent, | ||||
|             ) | ||||
| 
 | ||||
|         else:  # the `pprint` recursion-safe format: | ||||
|         else: | ||||
|             val_str: str = repr(v) | ||||
| 
 | ||||
|             # XXX LOL, below just seems to be f#$%in causing | ||||
|             # recursion errs.. | ||||
|             # | ||||
|             # the `pprint` recursion-safe format: | ||||
|             # https://docs.python.org/3.11/library/pprint.html#pprint.saferepr | ||||
|             try: | ||||
|                 val_str: str = saferepr(v) | ||||
|             except Exception: | ||||
|                 log.exception( | ||||
|                     'Failed to `saferepr({type(struct)})` !?\n' | ||||
|                 ) | ||||
|             return _Struct.__repr__(struct) | ||||
|             # try: | ||||
|             #     val_str: str = saferepr(v) | ||||
|             # except Exception: | ||||
|             #     log.exception( | ||||
|             #         'Failed to `saferepr({type(struct)})` !?\n' | ||||
|             #     ) | ||||
|                 # raise | ||||
|                 # return _Struct.__repr__(struct) | ||||
| 
 | ||||
|         # TODO: LOLOL use `textwrap.indent()` instead dawwwwwg! | ||||
|         obj_str += (field_ws + f'{k}: {typ_name} = {val_str},\n') | ||||
|  | @ -203,12 +210,7 @@ class Struct( | |||
|         return sin_props | ||||
| 
 | ||||
|     pformat = pformat | ||||
|     # __repr__ = pformat | ||||
|     # __str__ = __repr__ = pformat | ||||
|     # TODO: use a pprint.PrettyPrinter instance around ONLY rendering | ||||
|     # inside a known tty? | ||||
|     # def __repr__(self) -> str: | ||||
|     #     ... | ||||
| 
 | ||||
|     def __repr__(self) -> str: | ||||
|         try: | ||||
|             return pformat(self) | ||||
|  | @ -218,6 +220,13 @@ class Struct( | |||
|             ) | ||||
|             return _Struct.__repr__(self) | ||||
| 
 | ||||
|     # __repr__ = pformat | ||||
|     # __str__ = __repr__ = pformat | ||||
|     # TODO: use a pprint.PrettyPrinter instance around ONLY rendering | ||||
|     # inside a known tty? | ||||
|     # def __repr__(self) -> str: | ||||
|     #     ... | ||||
| 
 | ||||
|     def copy( | ||||
|         self, | ||||
|         update: dict | None = None, | ||||
|  | @ -267,13 +276,15 @@ class Struct( | |||
|                 fi.type(getattr(self, fi.name)), | ||||
|             ) | ||||
| 
 | ||||
|     # TODO: make a mod func instead and just point to it here for | ||||
|     # method impl? | ||||
|     def __sub__( | ||||
|         self, | ||||
|         other: Struct, | ||||
| 
 | ||||
|     ) -> DiffDump[tuple[str, Any, Any]]: | ||||
|         ''' | ||||
|         Compare fields/items key-wise and return a ``DiffDump`` | ||||
|         Compare fields/items key-wise and return a `DiffDump` | ||||
|         for easy visual REPL comparison B) | ||||
| 
 | ||||
|         ''' | ||||
|  | @ -290,3 +301,42 @@ class Struct( | |||
|                 )) | ||||
| 
 | ||||
|         return diffs | ||||
| 
 | ||||
|     @classmethod | ||||
|     def fields_diff( | ||||
|         cls, | ||||
|         other: dict|Struct, | ||||
| 
 | ||||
|     ) -> DiffDump[tuple[str, Any, Any]]: | ||||
|         ''' | ||||
|         Very similar to `PrettyStruct.__sub__()` except accepts an | ||||
|         input `other: dict` (presumably that would normally be called | ||||
|         like `Struct(**other)`) which returns a `DiffDump` of the | ||||
|         fields of the struct and the `dict`'s fields. | ||||
| 
 | ||||
|         ''' | ||||
|         nullish = object() | ||||
|         consumed: dict = other.copy() | ||||
|         diffs: DiffDump[tuple[str, Any, Any]] = DiffDump() | ||||
|         for fi in structs.fields(cls): | ||||
|             field_name: str = fi.name | ||||
|             # ours: Any = getattr(self, field_name) | ||||
|             theirs: Any = consumed.pop(field_name, nullish) | ||||
|             if theirs is nullish: | ||||
|                 diffs.append(( | ||||
|                     field_name, | ||||
|                     f'{fi.type!r}', | ||||
|                     'NOT-DEFINED in `other: dict`', | ||||
|                 )) | ||||
| 
 | ||||
|         # when there are lingering fields in `other` that this struct | ||||
|         # DOES NOT define we also append those. | ||||
|         if consumed: | ||||
|             for k, v in consumed.items(): | ||||
|                 diffs.append(( | ||||
|                     k, | ||||
|                     f'NOT-DEFINED for `{cls.__name__}`', | ||||
|                     f'`other: dict` has value = {v!r}', | ||||
|                 )) | ||||
| 
 | ||||
|         return diffs | ||||
|  |  | |||
|  | @ -599,15 +599,15 @@ def mk_msg_spec( | |||
|         Msg[payload_type_union], | ||||
|         Generic[PayloadT], | ||||
|     ) | ||||
|     defstruct_bases: tuple = ( | ||||
|         Msg, # [payload_type_union], | ||||
|         # Generic[PayloadT], | ||||
|         # ^-XXX-^: not allowed? lul.. | ||||
|     ) | ||||
|     # defstruct_bases: tuple = ( | ||||
|     #     Msg, # [payload_type_union], | ||||
|     #     # Generic[PayloadT], | ||||
|     #     # ^-XXX-^: not allowed? lul.. | ||||
|     # ) | ||||
|     ipc_msg_types: list[Msg] = [] | ||||
| 
 | ||||
|     idx_msg_types: list[Msg] = [] | ||||
|     defs_msg_types: list[Msg] = [] | ||||
|     # defs_msg_types: list[Msg] = [] | ||||
|     nc_msg_types: list[Msg] = [] | ||||
| 
 | ||||
|     for msgtype in __msg_types__: | ||||
|  | @ -625,7 +625,7 @@ def mk_msg_spec( | |||
|         # TODO: wait why do we need the dynamic version here? | ||||
|         # XXX ANSWER XXX -> BC INHERITANCE.. don't work w generics.. | ||||
|         # | ||||
|         # NOTE previously bc msgtypes WERE NOT inheritting | ||||
|         # NOTE previously bc msgtypes WERE NOT inheriting | ||||
|         # directly the `Generic[PayloadT]` type, the manual method | ||||
|         # of generic-paraming with `.__class_getitem__()` wasn't | ||||
|         # working.. | ||||
|  | @ -662,38 +662,35 @@ def mk_msg_spec( | |||
| 
 | ||||
|         # with `msgspec.structs.defstruct` | ||||
|         # XXX ALSO DOESN'T WORK | ||||
|         defstruct_msgtype = defstruct( | ||||
|             name=msgtype.__name__, | ||||
|             fields=[ | ||||
|                 ('cid', str), | ||||
|         # defstruct_msgtype = defstruct( | ||||
|         #     name=msgtype.__name__, | ||||
|         #     fields=[ | ||||
|         #         ('cid', str), | ||||
| 
 | ||||
|                 # XXX doesn't seem to work.. | ||||
|                 # ('pld', PayloadT), | ||||
| 
 | ||||
|                 ('pld', payload_type_union), | ||||
|             ], | ||||
|             bases=defstruct_bases, | ||||
|         ) | ||||
|         defs_msg_types.append(defstruct_msgtype) | ||||
|         #         # XXX doesn't seem to work.. | ||||
|         #         # ('pld', PayloadT), | ||||
| 
 | ||||
|         #         ('pld', payload_type_union), | ||||
|         #     ], | ||||
|         #     bases=defstruct_bases, | ||||
|         # ) | ||||
|         # defs_msg_types.append(defstruct_msgtype) | ||||
|         # assert index_paramed_msg_type == manual_paramed_msg_subtype | ||||
| 
 | ||||
|         # paramed_msg_type = manual_paramed_msg_subtype | ||||
| 
 | ||||
|         # ipc_payload_msgs_type_union |= index_paramed_msg_type | ||||
| 
 | ||||
|     idx_spec: Union[Type[Msg]] = Union[*idx_msg_types] | ||||
|     def_spec: Union[Type[Msg]] = Union[*defs_msg_types] | ||||
|     # def_spec: Union[Type[Msg]] = Union[*defs_msg_types] | ||||
|     nc_spec: Union[Type[Msg]] = Union[*nc_msg_types] | ||||
| 
 | ||||
|     specs: dict[str, Union[Type[Msg]]] = { | ||||
|         'indexed_generics': idx_spec, | ||||
|         'defstruct': def_spec, | ||||
|         # 'defstruct': def_spec, | ||||
|         'types_new_class': nc_spec, | ||||
|     } | ||||
|     msgtypes_table: dict[str, list[Msg]] = { | ||||
|         'indexed_generics': idx_msg_types, | ||||
|         'defstruct': defs_msg_types, | ||||
|         # 'defstruct': defs_msg_types, | ||||
|         'types_new_class': nc_msg_types, | ||||
|     } | ||||
| 
 | ||||
|  |  | |||
										
											
												File diff suppressed because it is too large
												Load Diff
											
										
									
								
							|  | @ -29,3 +29,6 @@ from ._broadcast import ( | |||
|     BroadcastReceiver as BroadcastReceiver, | ||||
|     Lagged as Lagged, | ||||
| ) | ||||
| from ._beg import ( | ||||
|     collapse_eg as collapse_eg, | ||||
| ) | ||||
|  |  | |||
|  | @ -0,0 +1,58 @@ | |||
| # tractor: structured concurrent "actors". | ||||
| # Copyright 2018-eternity Tyler Goodlet. | ||||
| 
 | ||||
| # This program is free software: you can redistribute it and/or modify | ||||
| # it under the terms of the GNU Affero General Public License as published by | ||||
| # the Free Software Foundation, either version 3 of the License, or | ||||
| # (at your option) any later version. | ||||
| 
 | ||||
| # This program is distributed in the hope that it will be useful, | ||||
| # but WITHOUT ANY WARRANTY; without even the implied warranty of | ||||
| # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the | ||||
| # GNU Affero General Public License for more details. | ||||
| 
 | ||||
| # You should have received a copy of the GNU Affero General Public License | ||||
| # along with this program.  If not, see <https://www.gnu.org/licenses/>. | ||||
| 
 | ||||
| ''' | ||||
| `BaseExceptionGroup` related utils and helpers pertaining to | ||||
| first-class-`trio` from a historical perspective B) | ||||
| 
 | ||||
| ''' | ||||
| from contextlib import ( | ||||
|     asynccontextmanager as acm, | ||||
| ) | ||||
| 
 | ||||
| 
 | ||||
| def maybe_collapse_eg( | ||||
|     beg: BaseExceptionGroup, | ||||
| ) -> BaseException: | ||||
|     ''' | ||||
|     If the input beg can collapse to a single non-eg sub-exception, | ||||
|     return it instead. | ||||
| 
 | ||||
|     ''' | ||||
|     if len(excs := beg.exceptions) == 1: | ||||
|         return excs[0] | ||||
| 
 | ||||
|     return beg | ||||
| 
 | ||||
| 
 | ||||
| @acm | ||||
| async def collapse_eg(): | ||||
|     ''' | ||||
|     If `BaseExceptionGroup` raised in the body scope is | ||||
|     "collapse-able" (in the same way that | ||||
|     `trio.open_nursery(strict_exception_groups=False)` works) then | ||||
|     only raise the lone emedded non-eg in in place. | ||||
| 
 | ||||
|     ''' | ||||
|     try: | ||||
|         yield | ||||
|     except* BaseException as beg: | ||||
|         if ( | ||||
|             exc := maybe_collapse_eg(beg) | ||||
|         ) is not beg: | ||||
|             raise exc | ||||
| 
 | ||||
|         raise beg | ||||
|  | @ -15,7 +15,7 @@ | |||
| # along with this program.  If not, see <https://www.gnu.org/licenses/>. | ||||
| 
 | ||||
| ''' | ||||
| ``tokio`` style broadcast channel. | ||||
| `tokio` style broadcast channel. | ||||
| https://docs.rs/tokio/1.11.0/tokio/sync/broadcast/index.html | ||||
| 
 | ||||
| ''' | ||||
|  | @ -382,7 +382,7 @@ class BroadcastReceiver(ReceiveChannel): | |||
|                         # likely it makes sense to unwind back to the | ||||
|                         # underlying? | ||||
|                         # import tractor | ||||
|                         # await tractor.breakpoint() | ||||
|                         # await tractor.pause() | ||||
|                         log.warning( | ||||
|                             f'Only one sub left for {self}?\n' | ||||
|                             'We can probably unwind from breceiver?' | ||||
|  |  | |||
|  | @ -57,6 +57,8 @@ async def maybe_open_nursery( | |||
|     shield: bool = False, | ||||
|     lib: ModuleType = trio, | ||||
| 
 | ||||
|     **kwargs,  # proxy thru | ||||
| 
 | ||||
| ) -> AsyncGenerator[trio.Nursery, Any]: | ||||
|     ''' | ||||
|     Create a new nursery if None provided. | ||||
|  | @ -67,7 +69,7 @@ async def maybe_open_nursery( | |||
|     if nursery is not None: | ||||
|         yield nursery | ||||
|     else: | ||||
|         async with lib.open_nursery() as nursery: | ||||
|         async with lib.open_nursery(**kwargs) as nursery: | ||||
|             nursery.cancel_scope.shield = shield | ||||
|             yield nursery | ||||
| 
 | ||||
|  | @ -143,9 +145,14 @@ async def gather_contexts( | |||
|             'Use a non-lazy iterator or sequence type intead!' | ||||
|         ) | ||||
| 
 | ||||
|     async with trio.open_nursery() as n: | ||||
|     async with trio.open_nursery( | ||||
|         strict_exception_groups=False, | ||||
|         # ^XXX^ TODO? soo roll our own then ?? | ||||
|         # -> since we kinda want the "if only one `.exception` then | ||||
|         # just raise that" interface? | ||||
|     ) as tn: | ||||
|         for mngr in mngrs: | ||||
|             n.start_soon( | ||||
|             tn.start_soon( | ||||
|                 _enter_and_wait, | ||||
|                 mngr, | ||||
|                 unwrapped, | ||||
|  |  | |||
							
								
								
									
										88
									
								
								uv.lock
								
								
								
								
							
							
						
						
									
										88
									
								
								uv.lock
								
								
								
								
							|  | @ -126,7 +126,31 @@ wheels = [ | |||
| [[package]] | ||||
| name = "msgspec" | ||||
| version = "0.19.0" | ||||
| source = { git = "https://github.com/jcrist/msgspec.git#dd965dce22e5278d4935bea923441ecde31b5325" } | ||||
| source = { registry = "https://pypi.org/simple" } | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/cf/9b/95d8ce458462b8b71b8a70fa94563b2498b89933689f3a7b8911edfae3d7/msgspec-0.19.0.tar.gz", hash = "sha256:604037e7cd475345848116e89c553aa9a233259733ab51986ac924ab1b976f8e", size = 216934 } | ||||
| wheels = [ | ||||
|     { url = "https://files.pythonhosted.org/packages/24/d4/2ec2567ac30dab072cce3e91fb17803c52f0a37aab6b0c24375d2b20a581/msgspec-0.19.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:aa77046904db764b0462036bc63ef71f02b75b8f72e9c9dd4c447d6da1ed8f8e", size = 187939 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/2b/c0/18226e4328897f4f19875cb62bb9259fe47e901eade9d9376ab5f251a929/msgspec-0.19.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:047cfa8675eb3bad68722cfe95c60e7afabf84d1bd8938979dd2b92e9e4a9551", size = 182202 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/81/25/3a4b24d468203d8af90d1d351b77ea3cffb96b29492855cf83078f16bfe4/msgspec-0.19.0-cp311-cp311-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:e78f46ff39a427e10b4a61614a2777ad69559cc8d603a7c05681f5a595ea98f7", size = 209029 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/85/2e/db7e189b57901955239f7689b5dcd6ae9458637a9c66747326726c650523/msgspec-0.19.0-cp311-cp311-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:6c7adf191e4bd3be0e9231c3b6dc20cf1199ada2af523885efc2ed218eafd011", size = 210682 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/03/97/7c8895c9074a97052d7e4a1cc1230b7b6e2ca2486714eb12c3f08bb9d284/msgspec-0.19.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:f04cad4385e20be7c7176bb8ae3dca54a08e9756cfc97bcdb4f18560c3042063", size = 214003 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/61/61/e892997bcaa289559b4d5869f066a8021b79f4bf8e955f831b095f47a4cd/msgspec-0.19.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:45c8fb410670b3b7eb884d44a75589377c341ec1392b778311acdbfa55187716", size = 216833 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/ce/3d/71b2dffd3a1c743ffe13296ff701ee503feaebc3f04d0e75613b6563c374/msgspec-0.19.0-cp311-cp311-win_amd64.whl", hash = "sha256:70eaef4934b87193a27d802534dc466778ad8d536e296ae2f9334e182ac27b6c", size = 186184 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/b2/5f/a70c24f075e3e7af2fae5414c7048b0e11389685b7f717bb55ba282a34a7/msgspec-0.19.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:f98bd8962ad549c27d63845b50af3f53ec468b6318400c9f1adfe8b092d7b62f", size = 190485 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/89/b0/1b9763938cfae12acf14b682fcf05c92855974d921a5a985ecc197d1c672/msgspec-0.19.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:43bbb237feab761b815ed9df43b266114203f53596f9b6e6f00ebd79d178cdf2", size = 183910 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/87/81/0c8c93f0b92c97e326b279795f9c5b956c5a97af28ca0fbb9fd86c83737a/msgspec-0.19.0-cp312-cp312-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:4cfc033c02c3e0aec52b71710d7f84cb3ca5eb407ab2ad23d75631153fdb1f12", size = 210633 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/d0/ef/c5422ce8af73928d194a6606f8ae36e93a52fd5e8df5abd366903a5ca8da/msgspec-0.19.0-cp312-cp312-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:d911c442571605e17658ca2b416fd8579c5050ac9adc5e00c2cb3126c97f73bc", size = 213594 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/19/2b/4137bc2ed45660444842d042be2cf5b18aa06efd2cda107cff18253b9653/msgspec-0.19.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:757b501fa57e24896cf40a831442b19a864f56d253679f34f260dcb002524a6c", size = 214053 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/9d/e6/8ad51bdc806aac1dc501e8fe43f759f9ed7284043d722b53323ea421c360/msgspec-0.19.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:5f0f65f29b45e2816d8bded36e6b837a4bf5fb60ec4bc3c625fa2c6da4124537", size = 219081 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/b1/ef/27dd35a7049c9a4f4211c6cd6a8c9db0a50647546f003a5867827ec45391/msgspec-0.19.0-cp312-cp312-win_amd64.whl", hash = "sha256:067f0de1c33cfa0b6a8206562efdf6be5985b988b53dd244a8e06f993f27c8c0", size = 187467 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/3c/cb/2842c312bbe618d8fefc8b9cedce37f773cdc8fa453306546dba2c21fd98/msgspec-0.19.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:f12d30dd6266557aaaf0aa0f9580a9a8fbeadfa83699c487713e355ec5f0bd86", size = 190498 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/58/95/c40b01b93465e1a5f3b6c7d91b10fb574818163740cc3acbe722d1e0e7e4/msgspec-0.19.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:82b2c42c1b9ebc89e822e7e13bbe9d17ede0c23c187469fdd9505afd5a481314", size = 183950 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/e8/f0/5b764e066ce9aba4b70d1db8b087ea66098c7c27d59b9dd8a3532774d48f/msgspec-0.19.0-cp313-cp313-manylinux_2_17_aarch64.manylinux2014_aarch64.whl", hash = "sha256:19746b50be214a54239aab822964f2ac81e38b0055cca94808359d779338c10e", size = 210647 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/9d/87/bc14f49bc95c4cb0dd0a8c56028a67c014ee7e6818ccdce74a4862af259b/msgspec-0.19.0-cp313-cp313-manylinux_2_17_x86_64.manylinux2014_x86_64.whl", hash = "sha256:60ef4bdb0ec8e4ad62e5a1f95230c08efb1f64f32e6e8dd2ced685bcc73858b5", size = 213563 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/53/2f/2b1c2b056894fbaa975f68f81e3014bb447516a8b010f1bed3fb0e016ed7/msgspec-0.19.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:ac7f7c377c122b649f7545810c6cd1b47586e3aa3059126ce3516ac7ccc6a6a9", size = 213996 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/aa/5a/4cd408d90d1417e8d2ce6a22b98a6853c1b4d7cb7669153e4424d60087f6/msgspec-0.19.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:a5bc1472223a643f5ffb5bf46ccdede7f9795078194f14edd69e3aab7020d327", size = 219087 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/23/d8/f15b40611c2d5753d1abb0ca0da0c75348daf1252220e5dda2867bd81062/msgspec-0.19.0-cp313-cp313-win_amd64.whl", hash = "sha256:317050bc0f7739cb30d257ff09152ca309bf5a369854bbf1e57dffc310c1f20f", size = 187432 }, | ||||
| ] | ||||
| 
 | ||||
| [[package]] | ||||
| name = "outcome" | ||||
|  | @ -240,7 +264,7 @@ wheels = [ | |||
| 
 | ||||
| [[package]] | ||||
| name = "pytest" | ||||
| version = "8.3.4" | ||||
| version = "8.3.5" | ||||
| source = { registry = "https://pypi.org/simple" } | ||||
| dependencies = [ | ||||
|     { name = "colorama", marker = "sys_platform == 'win32'" }, | ||||
|  | @ -248,9 +272,9 @@ dependencies = [ | |||
|     { name = "packaging" }, | ||||
|     { name = "pluggy" }, | ||||
| ] | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/05/35/30e0d83068951d90a01852cb1cef56e5d8a09d20c7f511634cc2f7e0372a/pytest-8.3.4.tar.gz", hash = "sha256:965370d062bce11e73868e0335abac31b4d3de0e82f4007408d242b4f8610761", size = 1445919 } | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/ae/3c/c9d525a414d506893f0cd8a8d0de7706446213181570cdbd766691164e40/pytest-8.3.5.tar.gz", hash = "sha256:f4efe70cc14e511565ac476b57c279e12a855b11f48f212af1080ef2263d3845", size = 1450891 } | ||||
| wheels = [ | ||||
|     { url = "https://files.pythonhosted.org/packages/11/92/76a1c94d3afee238333bc0a42b82935dd8f9cf8ce9e336ff87ee14d9e1cf/pytest-8.3.4-py3-none-any.whl", hash = "sha256:50e16d954148559c9a74109af1eaf0c945ba2d8f30f0a3d3335edde19788b6f6", size = 343083 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/30/3d/64ad57c803f1fa1e963a7946b6e0fea4a70df53c1a7fed304586539c2bac/pytest-8.3.5-py3-none-any.whl", hash = "sha256:c69214aa47deac29fad6c2a4f590b9c4a9fdb16a403176fe154b79c0b4d4d820", size = 343634 }, | ||||
| ] | ||||
| 
 | ||||
| [[package]] | ||||
|  | @ -314,17 +338,15 @@ dev = [ | |||
|     { name = "pytest" }, | ||||
|     { name = "stackscope" }, | ||||
|     { name = "xonsh" }, | ||||
|     { name = "xonsh-vox-tabcomplete" }, | ||||
|     { name = "xontrib-vox" }, | ||||
| ] | ||||
| 
 | ||||
| [package.metadata] | ||||
| requires-dist = [ | ||||
|     { name = "colorlog", specifier = ">=6.8.2,<7" }, | ||||
|     { name = "msgspec", git = "https://github.com/jcrist/msgspec.git" }, | ||||
|     { name = "pdbp", specifier = ">=1.5.0,<2" }, | ||||
|     { name = "msgspec", specifier = ">=0.19.0" }, | ||||
|     { name = "pdbp", specifier = ">=1.6,<2" }, | ||||
|     { name = "tricycle", specifier = ">=0.4.1,<0.5" }, | ||||
|     { name = "trio", specifier = ">=0.24,<0.25" }, | ||||
|     { name = "trio", specifier = ">0.27" }, | ||||
|     { name = "wrapt", specifier = ">=1.16.0,<2" }, | ||||
| ] | ||||
| 
 | ||||
|  | @ -332,13 +354,11 @@ requires-dist = [ | |||
| dev = [ | ||||
|     { name = "greenback", specifier = ">=1.2.1,<2" }, | ||||
|     { name = "pexpect", specifier = ">=4.9.0,<5" }, | ||||
|     { name = "prompt-toolkit", specifier = ">=3.0.43,<4" }, | ||||
|     { name = "prompt-toolkit", specifier = ">=3.0.50" }, | ||||
|     { name = "pyperclip", specifier = ">=1.9.0" }, | ||||
|     { name = "pytest", specifier = ">=8.2.0,<9" }, | ||||
|     { name = "pytest", specifier = ">=8.3.5" }, | ||||
|     { name = "stackscope", specifier = ">=0.2.2,<0.3" }, | ||||
|     { name = "xonsh", specifier = ">=0.19.1" }, | ||||
|     { name = "xonsh-vox-tabcomplete", specifier = ">=0.5,<0.6" }, | ||||
|     { name = "xontrib-vox", specifier = ">=0.0.1,<0.0.2" }, | ||||
|     { name = "xonsh", specifier = ">=0.19.2" }, | ||||
| ] | ||||
| 
 | ||||
| [[package]] | ||||
|  | @ -355,7 +375,7 @@ wheels = [ | |||
| 
 | ||||
| [[package]] | ||||
| name = "trio" | ||||
| version = "0.24.0" | ||||
| version = "0.29.0" | ||||
| source = { registry = "https://pypi.org/simple" } | ||||
| dependencies = [ | ||||
|     { name = "attrs" }, | ||||
|  | @ -365,9 +385,9 @@ dependencies = [ | |||
|     { name = "sniffio" }, | ||||
|     { name = "sortedcontainers" }, | ||||
| ] | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/8a/f3/07c152213222c615fe2391b8e1fea0f5af83599219050a549c20fcbd9ba2/trio-0.24.0.tar.gz", hash = "sha256:ffa09a74a6bf81b84f8613909fb0beaee84757450183a7a2e0b47b455c0cac5d", size = 545131 } | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/a1/47/f62e62a1a6f37909aed0bf8f5d5411e06fa03846cfcb64540cd1180ccc9f/trio-0.29.0.tar.gz", hash = "sha256:ea0d3967159fc130acb6939a0be0e558e364fee26b5deeecc893a6b08c361bdf", size = 588952 } | ||||
| wheels = [ | ||||
|     { url = "https://files.pythonhosted.org/packages/14/fb/9299cf74953f473a15accfdbe2c15218e766bae8c796f2567c83bae03e98/trio-0.24.0-py3-none-any.whl", hash = "sha256:c3bd3a4e3e3025cd9a2241eae75637c43fe0b9e88b4c97b9161a55b9e54cd72c", size = 460205 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/c9/55/c4d9bea8b3d7937901958f65124123512419ab0eb73695e5f382521abbfb/trio-0.29.0-py3-none-any.whl", hash = "sha256:d8c463f1a9cc776ff63e331aba44c125f423a5a13c684307e828d930e625ba66", size = 492920 }, | ||||
| ] | ||||
| 
 | ||||
| [[package]] | ||||
|  | @ -434,33 +454,13 @@ wheels = [ | |||
| 
 | ||||
| [[package]] | ||||
| name = "xonsh" | ||||
| version = "0.19.1" | ||||
| version = "0.19.2" | ||||
| source = { registry = "https://pypi.org/simple" } | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/98/6e/b54a0b2685535995ee50f655103c463f9d339455c9b08c4bce3e03e7bb17/xonsh-0.19.1.tar.gz", hash = "sha256:5d3de649c909f6d14bc69232219bcbdb8152c830e91ddf17ad169c672397fb97", size = 796468 } | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/68/4e/56e95a5e607eb3b0da37396f87cde70588efc8ef819ab16f02d5b8378dc4/xonsh-0.19.2.tar.gz", hash = "sha256:cfdd0680d954a2c3aefd6caddcc7143a3d06aa417ed18365a08219bb71b960b0", size = 799960 } | ||||
| wheels = [ | ||||
|     { url = "https://files.pythonhosted.org/packages/8c/e6/db44068c5725af9678e37980ae9503165393d51b80dc8517fa4ec74af1cf/xonsh-0.19.1-py310-none-any.whl", hash = "sha256:83eb6610ed3535f8542abd80af9554fb7e2805b0b3f96e445f98d4b5cf1f7046", size = 640686 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/77/4e/e487e82349866b245c559433c9ba626026a2e66bd17d7f9ac1045082f146/xonsh-0.19.1-py311-none-any.whl", hash = "sha256:c176e515b0260ab803963d1f0924f1e32f1064aa6fd5d791aa0cf6cda3a924ae", size = 640680 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/5d/88/09060815548219b8f6953a06c247cb5c92d03cbdf7a02a980bda1b5754db/xonsh-0.19.1-py312-none-any.whl", hash = "sha256:fe1266c86b117aced3bdc4d5972420bda715864435d0bd3722d63451e8001036", size = 640604 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/83/ff/7873cb8184cffeafddbf861712831c2baa2e9dbecdbfd33b1228f0db0019/xonsh-0.19.1-py313-none-any.whl", hash = "sha256:3f158b6fc0bba954e0b989004d4261bafc4bd94c68c2abd75b825da23e5a869c", size = 641166 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/cc/03/b9f8dd338df0a330011d104e63d4d0acd8bbbc1e990ff049487b6bdf585d/xonsh-0.19.1-py39-none-any.whl", hash = "sha256:a900a6eb87d881a7ef90b1ac8522ba3699582f0bcb1e9abd863d32f6d63faf04", size = 632912 }, | ||||
| ] | ||||
| 
 | ||||
| [[package]] | ||||
| name = "xonsh-vox-tabcomplete" | ||||
| version = "0.5" | ||||
| source = { registry = "https://pypi.org/simple" } | ||||
| wheels = [ | ||||
|     { url = "https://files.pythonhosted.org/packages/ab/fd/af0c2ee6c067c2a4dc64ec03598c94de1f6ec5984b3116af917f3add4a16/xonsh_vox_tabcomplete-0.5-py3-none-any.whl", hash = "sha256:9701b198180f167071234e77eab87b7befa97c1873b088d0b3fbbe6d6d8dcaad", size = 14381 }, | ||||
| ] | ||||
| 
 | ||||
| [[package]] | ||||
| name = "xontrib-vox" | ||||
| version = "0.0.1" | ||||
| source = { registry = "https://pypi.org/simple" } | ||||
| dependencies = [ | ||||
|     { name = "xonsh" }, | ||||
| ] | ||||
| sdist = { url = "https://files.pythonhosted.org/packages/6c/ac/a5db68a1f2e4036f7ff4c8546b1cbe29edee2ff40e0ff931836745988b79/xontrib-vox-0.0.1.tar.gz", hash = "sha256:c1f0b155992b4b0ebe6dcfd651084a8707ade7372f7e456c484d2a85339d9907", size = 16504 } | ||||
| wheels = [ | ||||
|     { url = "https://files.pythonhosted.org/packages/23/58/dcdf11849c8340033da00669527ce75d8292a4e8d82605c082ed236a081a/xontrib_vox-0.0.1-py3-none-any.whl", hash = "sha256:df2bbb815832db5b04d46684f540eac967ee40ef265add2662a95d6947d04c70", size = 13467 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/6c/13/281094759df87b23b3c02dc4a16603ab08ea54d7f6acfeb69f3341137c7a/xonsh-0.19.2-py310-none-any.whl", hash = "sha256:ec7f163fd3a4943782aa34069d4e72793328c916a5975949dbec8536cbfc089b", size = 642301 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/29/41/a51e4c3918fe9a293b150cb949b1b8c6d45eb17dfed480dcb76ea43df4e7/xonsh-0.19.2-py311-none-any.whl", hash = "sha256:53c45f7a767901f2f518f9b8dd60fc653e0498e56e89825e1710bb0859985049", size = 642286 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/0a/93/9a77b731f492fac27c577dea2afb5a2bcc2a6a1c79be0c86c95498060270/xonsh-0.19.2-py312-none-any.whl", hash = "sha256:b24c619aa52b59eae4d35c4195dba9b19a2c548fb5c42c6f85f2b8ccb96807b5", size = 642386 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/be/75/070324769c1ff88d971ce040f4f486339be98e0a365c8dd9991eb654265b/xonsh-0.19.2-py313-none-any.whl", hash = "sha256:c53ef6c19f781fbc399ed1b382b5c2aac2125010679a3b61d643978273c27df0", size = 642873 }, | ||||
|     { url = "https://files.pythonhosted.org/packages/fa/cb/2c7ccec54f5b0e73fdf7650e8336582ff0347d9001c5ef8271dc00c034fe/xonsh-0.19.2-py39-none-any.whl", hash = "sha256:bcc0225dc3847f1ed2f175dac6122fbcc54cea67d9c2dc2753d9615e2a5ff284", size = 634602 }, | ||||
| ] | ||||
|  |  | |||
		Loading…
	
		Reference in New Issue