Draft v2 after new `run_in_actor()` API

Revamp the docs after some feedback from @vodik.
See #24 #25 for additional details.
init_docs
Tyler Goodlet 2018-08-03 00:55:50 -04:00
parent 88e92aae43
commit 8d5a72d423
2 changed files with 216 additions and 176 deletions

View File

@ -1,36 +1,13 @@
tractor tractor
======= =======
A minimalist `actor model`_ built on multiprocessing_ and trio_. An async-native `actor model`_ built on trio_ and multiprocessing_.
|travis| |travis|
.. |travis| image:: https://img.shields.io/travis/tgoodlet/tractor/master.svg .. |travis| image:: https://img.shields.io/travis/tgoodlet/tractor/master.svg
:target: https://travis-ci.org/tgoodlet/tractor :target: https://travis-ci.org/tgoodlet/tractor
``tractor`` is an attempt to take trionic_ concurrency concepts and apply
them to distributed-multicore Python.
``tractor`` lets you run and spawn Python *actors*: separate processes which are internally
running a ``trio`` scheduler and task tree (also known as an `async sandwich`_).
Actors communicate with each other by sending *messages* over channels_, but the details of this
in ``tractor`` is by default hidden and *actors* can instead easily invoke remote asynchronous
functions using *portals*.
``tractor``'s tenets non-comprehensively include:
- no spawning of processes *willy-nilly*; causality_ is paramount!
- `shared nothing architecture`_
- remote errors `always propagate`_ back to the caller
- verbatim support for ``trio``'s cancellation_ system
- no use of *proxy* objects to wrap RPC calls
- an immersive debugging experience
- be simple, be small
.. warning:: ``tractor`` is in alpha-alpha and is expected to change rapidly!
Expect nothing to be set in stone and your ideas about where it should go
to be greatly appreciated!
.. _actor model: https://en.wikipedia.org/wiki/Actor_model .. _actor model: https://en.wikipedia.org/wiki/Actor_model
.. _trio: https://github.com/python-trio/trio .. _trio: https://github.com/python-trio/trio
.. _multiprocessing: https://docs.python.org/3/library/multiprocessing.html .. _multiprocessing: https://docs.python.org/3/library/multiprocessing.html
@ -41,6 +18,53 @@ functions using *portals*.
.. _shared nothing architecture: https://en.wikipedia.org/wiki/Shared-nothing_architecture .. _shared nothing architecture: https://en.wikipedia.org/wiki/Shared-nothing_architecture
.. _cancellation: https://trio.readthedocs.io/en/latest/reference-core.html#cancellation-and-timeouts .. _cancellation: https://trio.readthedocs.io/en/latest/reference-core.html#cancellation-and-timeouts
.. _channels: https://en.wikipedia.org/wiki/Channel_(programming) .. _channels: https://en.wikipedia.org/wiki/Channel_(programming)
.. _chaos engineering: http://principlesofchaos.org/
What's this? Spawning event loops in subprocesses?
--------------------------------------------------
Close, but not quite.
``tractor`` is an attempt to take trionic_ concurrency concepts and apply
them to distributed multi-core Python.
``tractor`` lets you run and spawn *actors*: separate processes which run a ``trio``
scheduler and task tree (also known as an `async sandwich`_).
*Actors* communicate by sending *messages* over channels_ and avoid sharing any state.
This `actor model`_ allows for highly distributed software architecture which works just well on
multiple cores as it does over many hosts.
``tractor`` takes much inspiration from pulsar_ and execnet_ but attempts to be much more
focussed on sophistication of the lower level distributed architecture
as well as have first class support for modern async Python.
``tractor`` does **not** use ``asyncio`` hence **no** event loops.
The first step to grok ``tractor`` is to get the basics of ``trio``
down. A great place to start is the `trio docs`_ and this `blog post`_
by njsmith_.
.. _trio docs: https://trio.readthedocs.io/en/latest/
.. _blog post: https://vorpus.org/blog/notes-on-structured-concurrency-or-go-statement-considered-harmful/
.. _njsmith: https://github.com/njsmith/
Philosophy
----------
``tractor``'s tenets non-comprehensively include:
- no spawning of processes *willy-nilly*; causality_ is paramount!
- `shared nothing architecture`_
- remote errors `always propagate`_ back to the caller
- verbatim support for ``trio``'s cancellation_ system
- no use of *proxy* objects to wrap RPC calls
- an immersive debugging experience
- anti-fragility through `chaos engineering`_
.. warning:: ``tractor`` is in alpha-alpha and is expected to change rapidly!
Expect nothing to be set in stone. Your ideas about where it should go
are greatly appreciated!
.. _pulsar: http://quantmind.github.io/pulsar/design.html
.. _execnet: https://codespeak.net/execnet/
Install Install
@ -52,28 +76,10 @@ No PyPi release yet!
pip install git+git://github.com/tgoodlet/tractor.git pip install git+git://github.com/tgoodlet/tractor.git
What's this? Spawning event loops in subprocesses?
--------------------------------------------------
Close, but not quite.
The first step to grok ``tractor`` is to get the basics of ``trio``
down. A great place to start is the `trio docs`_ and this `blog post`_
by njsmith_.
``tractor`` takes much inspiration from pulsar_ and execnet_ but attempts to be much more
minimal, focus on sophistication of the lower level distributed architecture,
and of course does **not** use ``asyncio``, hence **no** event loops.
.. _trio docs: https://trio.readthedocs.io/en/latest/
.. _pulsar: http://quantmind.github.io/pulsar/design.html
.. _execnet: https://codespeak.net/execnet/
.. _blog post: https://vorpus.org/blog/notes-on-structured-concurrency-or-go-statement-considered-harmful/
.. _njsmith: https://github.com/njsmith/
A trynamic first scene A trynamic first scene
---------------------- ----------------------
As a first example let's spawn a couple *actors* and have them run their lines: Let's direct a couple *actors* and have them run their lines for
the hip new film we're shooting:
.. code:: python .. code:: python
@ -101,51 +107,109 @@ As a first example let's spawn a couple *actors* and have them run their lines:
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
print("Alright... Action!") print("Alright... Action!")
donny = await n.start_actor( donny = await n.run_in_actor(
'donny', 'donny',
main=partial(say_hello, 'gretchen'), say_hello,
rpc_module_paths=[_this_module], other_actor='gretchen',
outlive_main=True
) )
gretchen = await n.start_actor( gretchen = await n.run_in_actor(
'gretchen', 'gretchen',
main=partial(say_hello, 'donny'), say_hello,
rpc_module_paths=[_this_module], other_actor='donny',
) )
print(await gretchen.result()) print(await gretchen.result())
print(await donny.result()) print(await donny.result())
await donny.cancel_actor() await donny.cancel_actor()
print("CUTTTT CUUTT CUT!!?! Donny!! You're supposed to say...") print("CUTTTT CUUTT CUT!!! Donny!! You're supposed to say...")
tractor.run(main) tractor.run(main)
Here, we've spawned two actors, *donny* and *gretchen* in separate We spawn two *actors*, *donny* and *gretchen*.
processes. Each starts up and begins executing their *main task* Each actor starts up and executes their *main task* defined by an
defined by an async function, ``say_hello()``. The function instructs async function, ``say_hello()``. The function instructs each actor
each actor to find their partner and say hello by calling their to find their partner and say hello by calling their partner's
partner's ``hi()`` function using a something called a *portal*. Each ``hi()`` function using something called a *portal*. Each actor
actor receives a response and relays that back to the parent actor (in receives a response and relays that back to the parent actor (in
this case our "director"). this case our "director" executing ``main()``).
To gain more insight as to how ``tractor`` accomplishes all this please
read on!
Actor spawning and causality Actor spawning and causality
---------------------------- ----------------------------
``tractor`` tries to take ``trio``'s concept of causal task lifetimes ``tractor`` tries to take ``trio``'s concept of causal task lifetimes
to multi-process land. Accordingly ``tractor``'s actor nursery behaves to multi-process land. Accordingly, ``tractor``'s *actor nursery* behaves
similar to the nursery_ in ``trio``. That is, an ``ActorNursery`` similar to ``trio``'s nursery_. That is, ``tractor.open_nursery()``
created with ``tractor.open_nursery()`` waits on spawned sub-actors to opens an ``ActorNursery`` which waits on spawned *actors* to complete
complete (or error) in the same causal_ way ``trio`` waits on spawned (or error) in the same causal_ way ``trio`` waits on spawned subtasks.
subtasks. This includes errors from any one sub-actor causing all other This includes errors from any one actor causing all other actors
actors spawned by the nursery to be cancelled_. spawned by the same nursery to be cancelled_.
To spawn an actor open a *nursery block* and use the ``start_actor()`` To spawn an actor and run a function in it, open a *nursery block*
and use the ``run_in_actor()`` method:
.. code:: python
import tractor
def cellar_door():
return "Dang that's beautiful"
async def main():
"""The main ``tractor`` routine.
"""
async with tractor.open_nursery() as n:
portal = await n.run_in_actor('frank', movie_theatre_question)
# The ``async with`` will unblock here since the 'frank'
# actor has completed its main task ``movie_theatre_question()``.
print(await portal.result())
tractor.run(main)
What's going on?
- an initial *actor* is started with ``tractor.run()`` and told to execute
its main task_: ``main()``
- inside ``main()`` an actor is *spawned* using an ``ActorNusery`` and is told
to run a single function: ``cellar_door()``
- a ``portal`` instance (we'll get to what it is shortly)
returned from ``nursery.run_in_actor()`` is used to communicate with
the newly spawned *sub-actor*
- the second actor, *frank*, in a new *process* running a new ``trio`` task_
then executes ``cellar_door()`` and returns its result over a *channel* back
to the parent actor
- the parent actor retrieves the subactor's (*frank*) *final result* using ``portal.result()``
much like you'd expect from a future_.
This ``run_in_actor()`` API should look very familiar to users of
``asyncio``'s run_in_executor_ which uses a ``concurrent.futures`` Executor_.
Since you might also want to spawn long running *worker* or *daemon*
actors, each actor's *lifetime* can be determined based on the spawn
method: method:
- if the actor is spawned using ``run_in_actor()`` it terminates when
its *main* task completes (i.e. when the (async) function submitted
to it *returns*). The ``with tractor.open_nursery()`` exits only once
all actors' main function/task complete (just like the nursery_ in ``trio``)
- actors can be spawned to *live forever* using the ``start_actor()``
method and act like an RPC daemon that runs indefinitely (the
``with tractor.open_nursery()`` wont' exit) until cancelled_
Had we wanted the latter form in our example it would have looked like:
.. code:: python .. code:: python
def movie_theatre_question(): def movie_theatre_question():
@ -159,15 +223,15 @@ method:
"""The main ``tractor`` routine. """The main ``tractor`` routine.
""" """
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
portal = await n.start_actor( portal = await n.start_actor(
'frank', 'frank',
# enable the actor to run funcs from this current module # enable the actor to run funcs from this current module
rpc_module_paths=[__name__], rpc_module_paths=[__name__],
outlive_main=True,
) )
print(await portal.run(__name__, 'movie_theatre_question')) print(await portal.run(__name__, 'movie_theatre_question'))
# calls the subactor a 2nd time # call the subactor a 2nd time
print(await portal.run(__name__, 'movie_theatre_question')) print(await portal.run(__name__, 'movie_theatre_question'))
# the async with will block here indefinitely waiting # the async with will block here indefinitely waiting
@ -175,46 +239,13 @@ method:
# "outlive_main" actor it will never end until cancelled # "outlive_main" actor it will never end until cancelled
await portal.cancel_actor() await portal.cancel_actor()
Notice the ``portal`` instance returned from ``nursery.start_actor()``,
we'll get to that shortly.
Spawned actor lifetimes can be configured in one of two ways:
- the actor terminates when its *main* task completes (the default if
the ``main`` kwarg is provided)
- the actor can be told to ``outlive_main=True`` and thus act like an RPC
daemon where it runs indefinitely until cancelled
Had we wanted the former in our example it would have been much simpler:
.. code:: python
def cellar_door():
return "Dang that's beautiful"
async def main():
"""The main ``tractor`` routine.
"""
async with tractor.open_nursery() as n:
portal = await n.start_actor('some_linguist', main=cellar_door)
# The ``async with`` will unblock here since the 'some_linguist'
# actor has completed its main task ``cellar_door``.
print(await portal.result())
Note that the main task's *final result(s)* (returned from the provided
``main`` function) is **always** accessed using ``Portal.result()`` much
like you'd expect from a future_.
The ``rpc_module_paths`` `kwarg` above is a list of module path The ``rpc_module_paths`` `kwarg` above is a list of module path
strings that will be loaded and made accessible for execution in the strings that will be loaded and made accessible for execution in the
remote actor through a call to ``Portal.run()``. For now this is remote actor through a call to ``Portal.run()``. For now this is
a simple mechanism to restrict the functionality of the remote a simple mechanism to restrict the functionality of the remote
(daemonized) actor and uses Python's module system to limit the (and possibly daemonized) actor and uses Python's module system to
allowed remote function namespace(s). limit the allowed remote function namespace(s).
``tractor`` is opinionated about the underlying threading model used for ``tractor`` is opinionated about the underlying threading model used for
each *actor*. Since Python has a GIL and an actor model by definition each *actor*. Since Python has a GIL and an actor model by definition
@ -223,15 +254,18 @@ shares no state between actors, it fits naturally to use a multiprocessing_
hardware but also distribute over many hardware hosts (each *actor* can talk hardware but also distribute over many hardware hosts (each *actor* can talk
to all others with ease over standard network protocols). to all others with ease over standard network protocols).
.. _task: https://trio.readthedocs.io/en/latest/reference-core.html#tasks-let-you-do-multiple-things-at-once
.. _nursery: https://trio.readthedocs.io/en/latest/reference-core.html#nurseries-and-spawning .. _nursery: https://trio.readthedocs.io/en/latest/reference-core.html#nurseries-and-spawning
.. _causal: https://vorpus.org/blog/some-thoughts-on-asynchronous-api-design-in-a-post-asyncawait-world/#causality .. _causal: https://vorpus.org/blog/some-thoughts-on-asynchronous-api-design-in-a-post-asyncawait-world/#causality
.. _cancelled: https://trio.readthedocs.io/en/latest/reference-core.html#child-tasks-and-cancellation .. _cancelled: https://trio.readthedocs.io/en/latest/reference-core.html#child-tasks-and-cancellation
.. _run_in_executor: https://docs.python.org/3/library/asyncio-eventloop.html#executor
.. _Executor: https://docs.python.org/3/library/concurrent.futures.html#concurrent.futures.Executor
Transparent function calling using *portals* Transparent remote function calling using *portals*
-------------------------------------------- ---------------------------------------------------
``tractor`` introdces the concept of a *portal* which is an API ``tractor`` introduces the concept of a *portal* which is an API
borrowed_ from ``trio``. A portal may seems similar to the idea of borrowed_ from ``trio``. A portal may seem similar to the idea of
a RPC future_ except a *portal* allows invoking remote *async* functions and a RPC future_ except a *portal* allows invoking remote *async* functions and
generators and intermittently blocking to receive responses. This allows generators and intermittently blocking to receive responses. This allows
for fully async-native IPC between actors. for fully async-native IPC between actors.
@ -255,11 +289,53 @@ This *portal* approach turns out to be paricularly exciting with the
introduction of `asynchronous generators`_ in Python 3.6! It means that introduction of `asynchronous generators`_ in Python 3.6! It means that
actors can compose nicely in a data processing pipeline. actors can compose nicely in a data processing pipeline.
As an example here's an actor that streams for 1 second from a remote async
generator function running in a separate actor:
.. code:: python
from itertools import repeat
import trio
import tractor
async def stream_forever():
for i in repeat("I can see these little future bubble things"):
# each yielded value is sent over the ``Channel`` to the
# parent actor
yield i
await trio.sleep(0.01)
async def main():
# stream for at most 1 seconds
with trio.move_on_after(1) as cancel_scope:
async with tractor.open_nursery() as n:
portal = await n.start_actor(
f'donny',
rpc_module_paths=[__name__],
)
# this async for loop streams values from the above
# async generator running in a separate process
async for letter in await portal.run(__name__, 'stream_forever'):
print(letter)
# we support trio's cancellation system
assert cancel_scope.cancelled_caught
assert n.cancelled
tractor.run(main)
Alright, let's get fancy.
Say you wanted to spawn two actors which each pulling data feeds from Say you wanted to spawn two actors which each pulling data feeds from
two different sources (and wanted this work spread across 2 cpus). two different sources (and wanted this work spread across 2 cpus).
You also want to aggregate these feeds, do some processing on them and then You also want to aggregate these feeds, do some processing on them and then
deliver the final result stream to a client (or in this case parent) deliver the final result stream to a client (or in this case parent) actor
actor and print the results to your screen: and print the results to your screen:
.. code:: python .. code:: python
@ -287,7 +363,6 @@ actor and print the results to your screen:
portal = await nursery.start_actor( portal = await nursery.start_actor(
name=f'streamer_{i}', name=f'streamer_{i}',
rpc_module_paths=[__name__], rpc_module_paths=[__name__],
outlive_main=True, # daemonize these actors
) )
portals.append(portal) portals.append(portal)
@ -337,15 +412,15 @@ actor and print the results to your screen:
import time import time
pre_start = time.time() pre_start = time.time()
portal = await nursery.start_actor( portal = await nursery.run_in_actor(
name='aggregator', 'aggregator',
# executed in the actor's "main task" immediately aggregate,
main=partial(aggregate, seed), seed=seed,
) )
start = time.time() start = time.time()
# the portal call returns exactly what you'd expect # the portal call returns exactly what you'd expect
# as if the remote "main" function was called locally # as if the remote "aggregate" function was called locally
result_stream = [] result_stream = []
async for value in await portal.result(): async for value in await portal.result():
result_stream.append(value) result_stream.append(value)
@ -360,16 +435,12 @@ actor and print the results to your screen:
Here there's four actors running in separate processes (using all the Here there's four actors running in separate processes (using all the
cores on you machine). Two are streaming (by **yielding** value in the cores on you machine). Two are streaming by *yielding* values from the
``stream_data()`` async generator, one is aggregating values from ``stream_data()`` async generator, one is aggregating values from
those two in ``aggregate()`` (also an async generator) and shipping the those two in ``aggregate()`` (also an async generator) and shipping the
single stream of unique values up the parent actor (the ``'MainProcess'`` single stream of unique values up the parent actor (the ``'MainProcess'``
as ``multiprocessing`` calls it) which is running ``main()``. as ``multiprocessing`` calls it) which is running ``main()``.
There has also been some discussion about adding support for reactive
programming primitives and native support for asyncitertools_ like libs -
so keep an eye out for that!
.. _future: https://en.wikipedia.org/wiki/Futures_and_promises .. _future: https://en.wikipedia.org/wiki/Futures_and_promises
.. _borrowed: .. _borrowed:
https://trio.readthedocs.io/en/latest/reference-core.html#getting-back-into-the-trio-thread-from-another-thread https://trio.readthedocs.io/en/latest/reference-core.html#getting-back-into-the-trio-thread-from-another-thread
@ -380,38 +451,7 @@ so keep an eye out for that!
Cancellation Cancellation
------------ ------------
``tractor`` supports ``trio``'s cancellation_ system verbatim: ``tractor`` supports ``trio``'s cancellation_ system verbatim.
.. code:: python
import trio
import tractor
from itertools import repeat
async def stream_forever():
for i in repeat("I can see these little future bubble things"):
yield i
await trio.sleep(0.01)
async def main():
# stream for at most 1 second
with trio.move_on_after(1) as cancel_scope:
async with tractor.open_nursery() as n:
portal = await n.start_actor(
f'donny',
rpc_module_paths=[__name__],
outlive_main=True
)
async for letter in await portal.run(__name__, 'stream_forever'):
print(letter)
assert cancel_scope.cancelled_caught
assert n.cancelled
tractor.run(main)
Cancelling a nursery block cancels all actors spawned by it. Cancelling a nursery block cancels all actors spawned by it.
Eventually ``tractor`` plans to support different `supervision strategies`_ like ``erlang``. Eventually ``tractor`` plans to support different `supervision strategies`_ like ``erlang``.
@ -421,7 +461,7 @@ Eventually ``tractor`` plans to support different `supervision strategies`_ like
Remote error propagation Remote error propagation
------------------------ ------------------------
Any task invoked in a remote actor should ship any error(s) back to the calling Any task invoked in a remote actor should ship any error(s) back to the calling
actor where it is raised and expected to be dealt with. This way remote actor's actor where it is raised and expected to be dealt with. This way remote actors
are never cancelled unless explicitly asked or there's a bug in ``tractor`` itself. are never cancelled unless explicitly asked or there's a bug in ``tractor`` itself.
.. code:: python .. code:: python
@ -429,6 +469,7 @@ are never cancelled unless explicitly asked or there's a bug in ``tractor`` itse
async def assert_err(): async def assert_err():
assert 0 assert 0
async def main(): async def main():
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
real_actors = [] real_actors = []
@ -436,11 +477,10 @@ are never cancelled unless explicitly asked or there's a bug in ``tractor`` itse
real_actors.append(await n.start_actor( real_actors.append(await n.start_actor(
f'actor_{i}', f'actor_{i}',
rpc_module_paths=[__name__], rpc_module_paths=[__name__],
outlive_main=True
)) ))
# start one actor that will fail immediately # start one actor that will fail immediately
await n.start_actor('extra', main=assert_err) await n.run_in_actor('extra', assert_err)
# should error here with a ``RemoteActorError`` containing # should error here with a ``RemoteActorError`` containing
# an ``AssertionError`` and all the other actors have been cancelled # an ``AssertionError`` and all the other actors have been cancelled
@ -482,8 +522,9 @@ multiple RPC calls to an actor can access global data using the per actor
async def main(): async def main():
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
await n.start_actor( await n.run_in_actor(
'checker', main=check_statespace, 'checker',
check_statespace,
statespace=statespace statespace=statespace
) )
@ -579,6 +620,8 @@ Stuff I'd like to see ``tractor`` do one day:
- a distributed log ledger for tracking cluster behaviour - a distributed log ledger for tracking cluster behaviour
- a slick multi-process aware debugger much like in celery_ - a slick multi-process aware debugger much like in celery_
but with better `pdb++`_ support but with better `pdb++`_ support
- an extensive `chaos engineering`_ test suite
- support for reactive programming primitives and native support for asyncitertools_ like libs
If you're interested in tackling any of these please do shout about it on the If you're interested in tackling any of these please do shout about it on the
`trio gitter channel`_! `trio gitter channel`_!

View File

@ -188,6 +188,8 @@ def test_remote_error():
async def stream_forever(): async def stream_forever():
for i in repeat("I can see these little future bubble things"): for i in repeat("I can see these little future bubble things"):
# each yielded value is sent over the ``Channel`` to the
# parent actor
yield i yield i
await trio.sleep(0.01) await trio.sleep(0.01)
@ -195,16 +197,20 @@ async def stream_forever():
@tractor_test @tractor_test
async def test_cancel_infinite_streamer(): async def test_cancel_infinite_streamer():
# stream for at most 5 seconds # stream for at most 1 seconds
with trio.move_on_after(1) as cancel_scope: with trio.move_on_after(1) as cancel_scope:
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
portal = await n.start_actor( portal = await n.start_actor(
f'donny', f'donny',
rpc_module_paths=[__name__], rpc_module_paths=[__name__],
) )
# this async for loop streams values from the above
# async generator running in a separate process
async for letter in await portal.run(__name__, 'stream_forever'): async for letter in await portal.run(__name__, 'stream_forever'):
print(letter) print(letter)
# we support trio's cancellation system
assert cancel_scope.cancelled_caught assert cancel_scope.cancelled_caught
assert n.cancelled assert n.cancelled
@ -272,7 +278,7 @@ async def test_trynamic_trio():
print(await gretchen.result()) print(await gretchen.result())
print(await donny.result()) print(await donny.result())
await donny.cancel_actor() await donny.cancel_actor()
print("CUTTTT CUUTT CUT!!?! Donny!! You're supposed to say...") print("CUTTTT CUUTT CUT!!! Donny!! You're supposed to say...")
def movie_theatre_question(): def movie_theatre_question():
@ -287,6 +293,7 @@ async def test_movie_theatre_convo():
"""The main ``tractor`` routine. """The main ``tractor`` routine.
""" """
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
portal = await n.start_actor( portal = await n.start_actor(
'frank', 'frank',
# enable the actor to run funcs from this current module # enable the actor to run funcs from this current module
@ -294,7 +301,7 @@ async def test_movie_theatre_convo():
) )
print(await portal.run(__name__, 'movie_theatre_question')) print(await portal.run(__name__, 'movie_theatre_question'))
# calls the subactor a 2nd time # call the subactor a 2nd time
print(await portal.run(__name__, 'movie_theatre_question')) print(await portal.run(__name__, 'movie_theatre_question'))
# the async with will block here indefinitely waiting # the async with will block here indefinitely waiting
@ -303,17 +310,6 @@ async def test_movie_theatre_convo():
await portal.cancel_actor() await portal.cancel_actor()
@tractor_test
async def test_movie_theatre_convo_main_task():
async with tractor.open_nursery() as n:
portal = await n.run_in_actor('frank', movie_theatre_question)
# The ``async with`` will unblock here since the 'frank'
# actor has completed its main task ``movie_theatre_question()``.
print(await portal.result())
def cellar_door(): def cellar_door():
return "Dang that's beautiful" return "Dang that's beautiful"
@ -323,6 +319,7 @@ async def test_most_beautiful_word():
"""The main ``tractor`` routine. """The main ``tractor`` routine.
""" """
async with tractor.open_nursery() as n: async with tractor.open_nursery() as n:
portal = await n.run_in_actor('some_linguist', cellar_door) portal = await n.run_in_actor('some_linguist', cellar_door)
# The ``async with`` will unblock here since the 'some_linguist' # The ``async with`` will unblock here since the 'some_linguist'
@ -427,7 +424,7 @@ async def a_quadruple_example():
start = time.time() start = time.time()
# the portal call returns exactly what you'd expect # the portal call returns exactly what you'd expect
# as if the remote "main" function was called locally # as if the remote "aggregate" function was called locally
result_stream = [] result_stream = []
async for value in await portal.result(): async for value in await portal.result():
result_stream.append(value) result_stream.append(value)