Port to new `tractor.trionics.maybe_open_context()` api
parent
5b368992f6
commit
422977d27a
|
@ -66,14 +66,14 @@ def async_lifo_cache(maxsize=128):
|
||||||
async def open_cached_client(
|
async def open_cached_client(
|
||||||
brokername: str,
|
brokername: str,
|
||||||
) -> 'Client': # noqa
|
) -> 'Client': # noqa
|
||||||
'''Get a cached broker client from the current actor's local vars.
|
'''
|
||||||
|
Get a cached broker client from the current actor's local vars.
|
||||||
|
|
||||||
If one has not been setup do it and cache it.
|
If one has not been setup do it and cache it.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
brokermod = get_brokermod(brokername)
|
brokermod = get_brokermod(brokername)
|
||||||
async with maybe_open_context(
|
async with maybe_open_context(
|
||||||
key=brokername,
|
acm_func=brokermod.get_client,
|
||||||
mngr=brokermod.get_client(),
|
|
||||||
) as (cache_hit, client):
|
) as (cache_hit, client):
|
||||||
yield client
|
yield client
|
||||||
|
|
|
@ -27,7 +27,6 @@ from types import ModuleType
|
||||||
from typing import (
|
from typing import (
|
||||||
Any, Sequence,
|
Any, Sequence,
|
||||||
AsyncIterator, Optional,
|
AsyncIterator, Optional,
|
||||||
Awaitable, Callable,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
import trio
|
import trio
|
||||||
|
@ -372,11 +371,12 @@ async def open_sample_step_stream(
|
||||||
# a lone broker-daemon per provider should be
|
# a lone broker-daemon per provider should be
|
||||||
# created for all practical purposes
|
# created for all practical purposes
|
||||||
async with maybe_open_context(
|
async with maybe_open_context(
|
||||||
key=delay_s,
|
acm_func=partial(
|
||||||
mngr=portal.open_stream_from(
|
portal.open_stream_from,
|
||||||
iter_ohlc_periods,
|
iter_ohlc_periods,
|
||||||
delay_s=delay_s, # must be kwarg
|
|
||||||
),
|
),
|
||||||
|
|
||||||
|
kwargs={'delay_s': delay_s},
|
||||||
) as (cache_hit, istream):
|
) as (cache_hit, istream):
|
||||||
if cache_hit:
|
if cache_hit:
|
||||||
# add a new broadcast subscription for the quote stream
|
# add a new broadcast subscription for the quote stream
|
||||||
|
@ -574,6 +574,7 @@ async def open_feed(
|
||||||
|
|
||||||
feed._max_sample_rate = max(ohlc_sample_rates)
|
feed._max_sample_rate = max(ohlc_sample_rates)
|
||||||
|
|
||||||
|
# yield feed
|
||||||
try:
|
try:
|
||||||
yield feed
|
yield feed
|
||||||
finally:
|
finally:
|
||||||
|
@ -599,16 +600,18 @@ async def maybe_open_feed(
|
||||||
sym = symbols[0].lower()
|
sym = symbols[0].lower()
|
||||||
|
|
||||||
async with maybe_open_context(
|
async with maybe_open_context(
|
||||||
key=(brokername, sym),
|
acm_func=open_feed,
|
||||||
mngr=open_feed(
|
kwargs={
|
||||||
brokername,
|
'brokername': brokername,
|
||||||
[sym],
|
'symbols': [sym],
|
||||||
loglevel=loglevel,
|
'loglevel': loglevel,
|
||||||
**kwargs,
|
'tick_throttle': kwargs.get('tick_throttle'),
|
||||||
),
|
},
|
||||||
|
key=sym,
|
||||||
) as (cache_hit, feed):
|
) as (cache_hit, feed):
|
||||||
|
|
||||||
if cache_hit:
|
if cache_hit:
|
||||||
|
print('USING CACHED FEED')
|
||||||
# add a new broadcast subscription for the quote stream
|
# add a new broadcast subscription for the quote stream
|
||||||
# if this feed is likely already in use
|
# if this feed is likely already in use
|
||||||
async with feed.stream.subscribe() as bstream:
|
async with feed.stream.subscribe() as bstream:
|
||||||
|
|
|
@ -572,14 +572,14 @@ async def maybe_open_fsp_cluster(
|
||||||
**kwargs,
|
**kwargs,
|
||||||
) -> AsyncGenerator[int, dict[str, tractor.Portal]]:
|
) -> AsyncGenerator[int, dict[str, tractor.Portal]]:
|
||||||
|
|
||||||
uid = tractor.current_actor().uid
|
kwargs.update(
|
||||||
|
{'workers': workers}
|
||||||
|
)
|
||||||
|
|
||||||
async with maybe_open_context(
|
async with maybe_open_context(
|
||||||
key=uid, # for now make a cluster per client?
|
# for now make a cluster per client?
|
||||||
mngr=open_fsp_cluster(
|
acm_func=open_fsp_cluster,
|
||||||
workers,
|
kwargs=kwargs,
|
||||||
# loglevel=loglevel,
|
|
||||||
**kwargs,
|
|
||||||
),
|
|
||||||
) as (cache_hit, cluster_map):
|
) as (cache_hit, cluster_map):
|
||||||
if cache_hit:
|
if cache_hit:
|
||||||
log.info('re-using existing fsp cluster')
|
log.info('re-using existing fsp cluster')
|
||||||
|
|
Loading…
Reference in New Issue