Compare commits
174 Commits
Author | SHA1 | Date |
---|---|---|
|
ac0f43dc98 | |
|
3977f1cc7e | |
|
e45cb9d08a | |
|
27c523ca74 | |
|
b8b76a32a6 | |
|
dcee0ddd55 | |
|
67eab85f06 | |
|
afc95b8592 | |
|
14c98d82ee | |
|
b87aa30031 | |
|
958f53d8e9 | |
|
ba43b54175 | |
|
de970755d7 | |
|
7ddebf6773 | |
|
8eb4a427da | |
|
da5dea9f99 | |
|
3074773662 | |
|
4099b53ea2 | |
|
633fa7cc3a | |
|
1345b250bc | |
|
e9f0ea3daa | |
|
569674517f | |
|
bf7397f031 | |
|
85c2f6e79f | |
|
1c1661b783 | |
|
99eabe34c9 | |
|
827b5f9c45 | |
|
41f24f3de6 | |
|
34975dfbd5 | |
|
f6b54f02c0 | |
|
44c242a794 | |
|
99965e7601 | |
|
e5f96391e3 | |
|
a66934a49d | |
|
55772efb34 | |
|
736178adfd | |
|
d770867163 | |
|
c518553aa9 | |
|
4138cef512 | |
|
0f4bfcdf22 | |
|
80835d4e04 | |
|
e6d03ba97f | |
|
b71e8c5e6d | |
|
064d185395 | |
|
363ba8f9ae | |
|
fc24f5efd1 | |
|
a7ff47158b | |
|
57acc3bd29 | |
|
8f1faf97ee | |
|
3ab91deaec | |
|
6f00617bd3 | |
|
2c2c453932 | |
|
360643b32f | |
|
ab0def22c1 | |
|
a9ec1a97dd | |
|
d61b636487 | |
|
88ac2fda52 | |
|
08c83afa90 | |
|
066b8df619 | |
|
d4f31f2b3c | |
|
04897fd402 | |
|
42572d3808 | |
|
8ce7e99210 | |
|
1b38628b09 | |
|
bbe1ff19ef | |
|
eca2401ab5 | |
|
5d91516b41 | |
|
b985b48eb3 | |
|
c256d3bdc0 | |
|
f5de361f49 | |
|
432d4545c2 | |
|
fa30df36ba | |
|
17456d96e0 | |
|
167ae96566 | |
|
aa0efe1523 | |
|
664a208ae5 | |
|
876add4fc2 | |
|
72e849c651 | |
|
b3ae562e4f | |
|
b5b9ecf4b1 | |
|
1dab77ca0b | |
|
4c7661fc23 | |
|
e258654c86 | |
|
81be0b4bd0 | |
|
df1c89e811 | |
|
f67fd11a29 | |
|
1f95ba4fd8 | |
|
1dca7766d2 | |
|
b236dc72e4 | |
|
27ee9fdc81 | |
|
5d294031f2 | |
|
537b725bf3 | |
|
ca5a25f921 | |
|
037300ced0 | |
|
9c5bc6deda | |
|
bc50db5925 | |
|
e8e26232ea | |
|
f6909ae395 | |
|
b609f46d26 | |
|
5d26609693 | |
|
09e988ec3e | |
|
5e602214be | |
|
cfc4198837 | |
|
c455df7fa8 | |
|
47cf4aa4f7 | |
|
4f36743f64 | |
|
1fcb9233b4 | |
|
fb38265199 | |
|
e163a7e336 | |
|
36a10155bc | |
|
7a3437348d | |
|
0744dd0415 | |
|
0770a39125 | |
|
2b6041465c | |
|
859eaffa29 | |
|
b12921678b | |
|
186658ab09 | |
|
12d60e6d9c | |
|
c5beecf8a1 | |
|
629ea8ba9d | |
|
ba0ba346ec | |
|
82b2d2ee3a | |
|
b2b31b8f84 | |
|
b97ec38baf | |
|
64c6287cd1 | |
|
69282a9924 | |
|
aee44fed46 | |
|
db727910be | |
|
64206543cd | |
|
af6aad4e9c | |
|
c94c53286b | |
|
2af4050e5e | |
|
df78e9ba96 | |
|
7e1ec7b5a7 | |
|
3dbce6f891 | |
|
239c9d701a | |
|
427a33654b | |
|
f4dc0fbab8 | |
|
e0a72a2174 | |
|
5a9bab0b69 | |
|
d0af280a59 | |
|
599c77ff84 | |
|
c097016fd2 | |
|
be7c4e70f0 | |
|
051680e259 | |
|
55a453a710 | |
|
88eccc1e15 | |
|
488506d8b8 | |
|
78b9333bcd | |
|
7229a39f47 | |
|
d870a09a4b | |
|
5d53ecb433 | |
|
06832b94d4 | |
|
8d6c5b214e | |
|
a5389beccd | |
|
26f47227d2 | |
|
b357a120b9 | |
|
aba8b05a33 | |
|
c3142aec81 | |
|
bff625725e | |
|
6f172479eb | |
|
a96f1dec3a | |
|
86caf5f6a3 | |
|
72b4273ddc | |
|
4281936ff4 | |
|
4ddf04f68b | |
|
339fcda727 | |
|
4b7d7d688e | |
|
7ae7b2f864 | |
|
fa9f8c78c3 | |
|
3bbbc21d2b | |
|
b03603a6b4 | |
|
81b77df544 | |
|
a79a99fc71 |
|
@ -138,7 +138,7 @@ provider support
|
||||||
****************
|
****************
|
||||||
for live data feeds the in-progress set of supported brokers is:
|
for live data feeds the in-progress set of supported brokers is:
|
||||||
|
|
||||||
- IB_ via ``ib_insync``
|
- IB_ via ``ib_insync``, also see our `container docs`_
|
||||||
- binance_ and kraken_ for crypto over their public websocket API
|
- binance_ and kraken_ for crypto over their public websocket API
|
||||||
- questrade_ (ish) which comes with effectively free L1
|
- questrade_ (ish) which comes with effectively free L1
|
||||||
|
|
||||||
|
@ -150,6 +150,7 @@ coming soon...
|
||||||
if you want your broker supported and they have an API let us know.
|
if you want your broker supported and they have an API let us know.
|
||||||
|
|
||||||
.. _IB: https://interactivebrokers.github.io/tws-api/index.html
|
.. _IB: https://interactivebrokers.github.io/tws-api/index.html
|
||||||
|
.. _container docs: https://github.com/pikers/piker/tree/master/dockering/ib
|
||||||
.. _questrade: https://www.questrade.com/api/documentation
|
.. _questrade: https://www.questrade.com/api/documentation
|
||||||
.. _kraken: https://www.kraken.com/features/api#public-market-data
|
.. _kraken: https://www.kraken.com/features/api#public-market-data
|
||||||
.. _binance: https://github.com/pikers/piker/pull/182
|
.. _binance: https://github.com/pikers/piker/pull/182
|
||||||
|
|
|
@ -12,7 +12,26 @@ api_key = ""
|
||||||
secret = ""
|
secret = ""
|
||||||
|
|
||||||
[ib]
|
[ib]
|
||||||
host = "127.0.0.1"
|
hosts = [
|
||||||
|
"127.0.0.1",
|
||||||
|
]
|
||||||
|
# XXX: the order in which ports will be scanned
|
||||||
|
# (by the `brokerd` daemon-actor)
|
||||||
|
# is determined # by the line order here.
|
||||||
|
# TODO: when we eventually spawn gateways in our
|
||||||
|
# container, we can just dynamically allocate these
|
||||||
|
# using IBC.
|
||||||
|
ports = [
|
||||||
|
4002, # gw
|
||||||
|
7497, # tws
|
||||||
|
]
|
||||||
|
|
||||||
|
# XXX: for a paper account the flex web query service
|
||||||
|
# is not supported so you have to manually download
|
||||||
|
# and XML report and put it in a location that can be
|
||||||
|
# accessed by the ``brokerd.ib`` backend code for parsing.
|
||||||
|
flex_token = '666666666666666666666666'
|
||||||
|
flex_trades_query_id = '666666' # live account
|
||||||
|
|
||||||
# when clients are being scanned this determines
|
# when clients are being scanned this determines
|
||||||
# which clients are preferred to be used for data
|
# which clients are preferred to be used for data
|
||||||
|
@ -24,16 +43,10 @@ prefer_data_account = [
|
||||||
'ira',
|
'ira',
|
||||||
]
|
]
|
||||||
|
|
||||||
# the order in which ports will be scanned
|
[ib.accounts]
|
||||||
# (by the `brokerd` daemon-actor)
|
|
||||||
# is determined # by the line order here.
|
|
||||||
ports.gw = 4002
|
|
||||||
ports.tws = 7497
|
|
||||||
ports.order = ["gw", "tws",]
|
|
||||||
|
|
||||||
# the order in which accounts will be selectable
|
# the order in which accounts will be selectable
|
||||||
# in the order mode UI (if found via clients during
|
# in the order mode UI (if found via clients during
|
||||||
# API-app scanning)when a new symbol is loaded.
|
# API-app scanning)when a new symbol is loaded.
|
||||||
accounts.paper = "XX0000000"
|
paper = "XX0000000"
|
||||||
accounts.margin = "X0000000"
|
margin = "X0000000"
|
||||||
accounts.ira = "X0000000"
|
ira = "X0000000"
|
||||||
|
|
|
@ -0,0 +1,30 @@
|
||||||
|
running ``ib`` gateway in ``docker``
|
||||||
|
------------------------------------
|
||||||
|
We have a config based on the (now defunct)
|
||||||
|
image from "waytrade":
|
||||||
|
|
||||||
|
https://github.com/waytrade/ib-gateway-docker
|
||||||
|
|
||||||
|
To startup this image with our custom settings
|
||||||
|
simply run the command::
|
||||||
|
|
||||||
|
docker compose up
|
||||||
|
|
||||||
|
And you should have the following socket-available services:
|
||||||
|
|
||||||
|
- ``x11vnc1@127.0.0.1:3003``
|
||||||
|
- ``ib-gw@127.0.0.1:4002``
|
||||||
|
|
||||||
|
You can attach to the container via a VNC client
|
||||||
|
without password auth.
|
||||||
|
|
||||||
|
SECURITY STUFF!?!?!
|
||||||
|
-------------------
|
||||||
|
Though "``ib``" claims they host filter connections outside
|
||||||
|
localhost (aka ``127.0.0.1``) it's probably better if you filter
|
||||||
|
the socket at the OS level using a stateless firewall rule::
|
||||||
|
|
||||||
|
ip rule add not unicast iif lo to 0.0.0.0/0 dport 4002
|
||||||
|
|
||||||
|
We will soon have this baked into our own custom image but for
|
||||||
|
now you'll have to do it urself dawgy.
|
|
@ -0,0 +1,64 @@
|
||||||
|
# rework from the original @
|
||||||
|
# https://github.com/waytrade/ib-gateway-docker/blob/master/docker-compose.yml
|
||||||
|
version: "3.5"
|
||||||
|
|
||||||
|
services:
|
||||||
|
ib-gateway:
|
||||||
|
# other image tags available:
|
||||||
|
# https://github.com/waytrade/ib-gateway-docker#supported-tags
|
||||||
|
image: waytrade/ib-gateway:981.3j
|
||||||
|
restart: always
|
||||||
|
network_mode: 'host'
|
||||||
|
|
||||||
|
volumes:
|
||||||
|
- type: bind
|
||||||
|
source: ./jts.ini
|
||||||
|
target: /root/Jts/jts.ini
|
||||||
|
# don't let IBC clobber this file for
|
||||||
|
# the main reason of not having a stupid
|
||||||
|
# timezone set..
|
||||||
|
read_only: true
|
||||||
|
|
||||||
|
# force our own IBC config
|
||||||
|
- type: bind
|
||||||
|
source: ./ibc.ini
|
||||||
|
target: /root/ibc/config.ini
|
||||||
|
|
||||||
|
# force our noop script - socat isn't needed in host mode.
|
||||||
|
- type: bind
|
||||||
|
source: ./fork_ports_delayed.sh
|
||||||
|
target: /root/scripts/fork_ports_delayed.sh
|
||||||
|
|
||||||
|
# force our noop script - socat isn't needed in host mode.
|
||||||
|
- type: bind
|
||||||
|
source: ./run_x11_vnc.sh
|
||||||
|
target: /root/scripts/run_x11_vnc.sh
|
||||||
|
read_only: true
|
||||||
|
|
||||||
|
# NOTE:to fill these out, define an `.env` file in the same dir as
|
||||||
|
# this compose file which looks something like:
|
||||||
|
# TWS_USERID='myuser'
|
||||||
|
# TWS_PASSWORD='guest'
|
||||||
|
# TRADING_MODE=paper (or live)
|
||||||
|
# VNC_SERVER_PASSWORD='diggity'
|
||||||
|
|
||||||
|
environment:
|
||||||
|
TWS_USERID: ${TWS_USERID}
|
||||||
|
TWS_PASSWORD: ${TWS_PASSWORD}
|
||||||
|
TRADING_MODE: ${TRADING_MODE:-paper}
|
||||||
|
VNC_SERVER_PASSWORD: ${VNC_SERVER_PASSWORD:-}
|
||||||
|
|
||||||
|
# ports:
|
||||||
|
# - target: 4002
|
||||||
|
# host_ip: 127.0.0.1
|
||||||
|
# published: 4002
|
||||||
|
# protocol: tcp
|
||||||
|
|
||||||
|
# original mappings for use in non-host-mode
|
||||||
|
# which we won't really need going forward since
|
||||||
|
# ideally we just pick the port to have ib-gw listen
|
||||||
|
# on **when** we spawn the container - i.e. everything
|
||||||
|
# will be driven by a ``brokers.toml`` def.
|
||||||
|
# - "127.0.0.1:4001:4001"
|
||||||
|
# - "127.0.0.1:4002:4002"
|
||||||
|
# - "127.0.0.1:5900:5900"
|
|
@ -0,0 +1,6 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
# we now just set this is to a noop script
|
||||||
|
# since we can just run the container in
|
||||||
|
# `network_mode: 'host'` and get literally
|
||||||
|
# the exact same behaviour XD
|
|
@ -0,0 +1,711 @@
|
||||||
|
# Note that in the comments in this file, TWS refers to both the Trader
|
||||||
|
# Workstation and the IB Gateway, unless explicitly stated otherwise.
|
||||||
|
#
|
||||||
|
# When referred to below, the default value for a setting is the value
|
||||||
|
# assumed if either the setting is included but no value is specified, or
|
||||||
|
# the setting is not included at all.
|
||||||
|
#
|
||||||
|
# IBC may also be used to start the FIX CTCI Gateway. All settings
|
||||||
|
# relating to this have names prefixed with FIX.
|
||||||
|
#
|
||||||
|
# The IB API Gateway and the FIX CTCI Gateway share the same code. Which
|
||||||
|
# gateway actually runs is governed by an option on the initial gateway
|
||||||
|
# login screen. The FIX setting described under IBC Startup
|
||||||
|
# Settings below controls this.
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 1. IBC Startup Settings
|
||||||
|
# =============================================================================
|
||||||
|
|
||||||
|
|
||||||
|
# IBC may be used to start the IB Gateway for the FIX CTCI. This
|
||||||
|
# setting must be set to 'yes' if you want to run the FIX CTCI gateway. The
|
||||||
|
# default is 'no'.
|
||||||
|
|
||||||
|
FIX=no
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 2. Authentication Settings
|
||||||
|
# =============================================================================
|
||||||
|
|
||||||
|
# TWS and the IB API gateway require a single username and password.
|
||||||
|
# You may specify the username and password using the following settings:
|
||||||
|
#
|
||||||
|
# IbLoginId
|
||||||
|
# IbPassword
|
||||||
|
#
|
||||||
|
# Alternatively, you can specify the username and password in the command
|
||||||
|
# files used to start TWS or the Gateway, but this is not recommended for
|
||||||
|
# security reasons.
|
||||||
|
#
|
||||||
|
# If you don't specify them, you will be prompted for them in the usual
|
||||||
|
# login dialog when TWS starts (but whatever you have specified will be
|
||||||
|
# included in the dialog automatically: for example you may specify the
|
||||||
|
# username but not the password, and then you will be prompted for the
|
||||||
|
# password via the login dialog). Note that if you specify either
|
||||||
|
# the username or the password (or both) in the command file, then
|
||||||
|
# IbLoginId and IbPassword settings defined in this file are ignored.
|
||||||
|
#
|
||||||
|
#
|
||||||
|
# The FIX CTCI gateway requires one username and password for FIX order
|
||||||
|
# routing, and optionally a separate username and password for market
|
||||||
|
# data connections. You may specify the usernames and passwords using
|
||||||
|
# the following settings:
|
||||||
|
#
|
||||||
|
# FIXLoginId
|
||||||
|
# FIXPassword
|
||||||
|
# IbLoginId (optional - for market data connections)
|
||||||
|
# IbPassword (optional - for market data connections)
|
||||||
|
#
|
||||||
|
# Alternatively you can specify the FIX username and password in the
|
||||||
|
# command file used to start the FIX CTCI Gateway, but this is not
|
||||||
|
# recommended for security reasons.
|
||||||
|
#
|
||||||
|
# If you don't specify them, you will be prompted for them in the usual
|
||||||
|
# login dialog when FIX CTCI gateway starts (but whatever you have
|
||||||
|
# specified will be included in the dialog automatically: for example
|
||||||
|
# you may specify the usernames but not the passwords, and then you will
|
||||||
|
# be prompted for the passwords via the login dialog). Note that if you
|
||||||
|
# specify either the FIX username or the FIX password (or both) on the
|
||||||
|
# command line, then FIXLoginId and FIXPassword settings defined in this
|
||||||
|
# file are ignored; he same applies to the market data username and
|
||||||
|
# password.
|
||||||
|
|
||||||
|
# IB API Authentication Settings
|
||||||
|
# ------------------------------
|
||||||
|
|
||||||
|
# Your TWS username:
|
||||||
|
|
||||||
|
IbLoginId=
|
||||||
|
|
||||||
|
|
||||||
|
# Your TWS password:
|
||||||
|
|
||||||
|
IbPassword=
|
||||||
|
|
||||||
|
|
||||||
|
# FIX CTCI Authentication Settings
|
||||||
|
# --------------------------------
|
||||||
|
|
||||||
|
# Your FIX CTCI username:
|
||||||
|
|
||||||
|
FIXLoginId=
|
||||||
|
|
||||||
|
|
||||||
|
# Your FIX CTCI password:
|
||||||
|
|
||||||
|
FIXPassword=
|
||||||
|
|
||||||
|
|
||||||
|
# Second Factor Authentication Settings
|
||||||
|
# -------------------------------------
|
||||||
|
|
||||||
|
# If you have enabled more than one second factor authentication
|
||||||
|
# device, TWS presents a list from which you must select the device
|
||||||
|
# you want to use for this login. You can use this setting to
|
||||||
|
# instruct IBC to select a particular item in the list on your
|
||||||
|
# behalf. Note that you must spell this value exactly as it appears
|
||||||
|
# in the list. If no value is set, you must manually select the
|
||||||
|
# relevant list entry.
|
||||||
|
|
||||||
|
SecondFactorDevice=
|
||||||
|
|
||||||
|
|
||||||
|
# If you use the IBKR Mobile app for second factor authentication,
|
||||||
|
# and you fail to complete the process before the time limit imposed
|
||||||
|
# by IBKR, you can use this setting to tell IBC to exit: arrangements
|
||||||
|
# can then be made to automatically restart IBC in order to initiate
|
||||||
|
# the login sequence afresh. Otherwise, manual intervention at TWS's
|
||||||
|
# Second Factor Authentication dialog is needed to complete the
|
||||||
|
# login.
|
||||||
|
#
|
||||||
|
# Permitted values are 'yes' and 'no'. The default is 'no'.
|
||||||
|
#
|
||||||
|
# Note that the scripts provided with the IBC zips for Windows and
|
||||||
|
# Linux provide options to automatically restart in these
|
||||||
|
# circumstances, but only if this setting is also set to 'yes'.
|
||||||
|
|
||||||
|
ExitAfterSecondFactorAuthenticationTimeout=no
|
||||||
|
|
||||||
|
|
||||||
|
# This setting is only relevant if
|
||||||
|
# ExitAfterSecondFactorAuthenticationTimeout is set to 'yes'.
|
||||||
|
#
|
||||||
|
# It controls how long (in seconds) IBC waits for login to complete
|
||||||
|
# after the user acknowledges the second factor authentication
|
||||||
|
# alert at the IBKR Mobile app. If login has not completed after
|
||||||
|
# this time, IBC terminates.
|
||||||
|
# The default value is 40.
|
||||||
|
|
||||||
|
SecondFactorAuthenticationExitInterval=
|
||||||
|
|
||||||
|
|
||||||
|
# Trading Mode
|
||||||
|
# ------------
|
||||||
|
#
|
||||||
|
# TWS 955 introduced a new Trading Mode combo box on its login
|
||||||
|
# dialog. This indicates whether the live account or the paper
|
||||||
|
# trading account corresponding to the supplied credentials is
|
||||||
|
# to be used. The allowed values are 'live' (the default) and
|
||||||
|
# 'paper'. For earlier versions of TWS this setting has no
|
||||||
|
# effect.
|
||||||
|
|
||||||
|
TradingMode=
|
||||||
|
|
||||||
|
|
||||||
|
# Paper-trading Account Warning
|
||||||
|
# -----------------------------
|
||||||
|
#
|
||||||
|
# Logging in to a paper-trading account results in TWS displaying
|
||||||
|
# a dialog asking the user to confirm that they are aware that this
|
||||||
|
# is not a brokerage account. Until this dialog has been accepted,
|
||||||
|
# TWS will not allow API connections to succeed. Setting this
|
||||||
|
# to 'yes' (the default) will cause IBC to automatically
|
||||||
|
# confirm acceptance. Setting it to 'no' will leave the dialog
|
||||||
|
# on display, and the user will have to deal with it manually.
|
||||||
|
|
||||||
|
AcceptNonBrokerageAccountWarning=yes
|
||||||
|
|
||||||
|
|
||||||
|
# Login Dialog Display Timeout
|
||||||
|
#-----------------------------
|
||||||
|
#
|
||||||
|
# In some circumstances, starting TWS may result in failure to display
|
||||||
|
# the login dialog. Restarting TWS may help to resolve this situation,
|
||||||
|
# and IBC does this automatically.
|
||||||
|
#
|
||||||
|
# This setting controls how long (in seconds) IBC waits for the login
|
||||||
|
# dialog to appear before restarting TWS.
|
||||||
|
#
|
||||||
|
# Note that in normal circumstances with a reasonably specified
|
||||||
|
# computer the time to displaying the login dialog is typically less
|
||||||
|
# than 20 seconds, and frequently much less. However many factors can
|
||||||
|
# influence this, and it is unwise to set this value too low.
|
||||||
|
#
|
||||||
|
# The default value is 60.
|
||||||
|
|
||||||
|
LoginDialogDisplayTimeout = 60
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 3. TWS Startup Settings
|
||||||
|
# =============================================================================
|
||||||
|
|
||||||
|
# Path to settings store
|
||||||
|
# ----------------------
|
||||||
|
#
|
||||||
|
# Path to the directory where TWS should store its settings. This is
|
||||||
|
# normally the folder in which TWS is installed. However you may set
|
||||||
|
# it to some other location if you wish (for example if you want to
|
||||||
|
# run multiple instances of TWS with different settings).
|
||||||
|
#
|
||||||
|
# It is recommended for clarity that you use an absolute path. The
|
||||||
|
# effect of using a relative path is undefined.
|
||||||
|
#
|
||||||
|
# Linux and macOS users should use the appropriate path syntax.
|
||||||
|
#
|
||||||
|
# Note that, for Windows users, you MUST use double separator
|
||||||
|
# characters to separate the elements of the folder path: for
|
||||||
|
# example, IbDir=C:\\IBLiveSettings is valid, but
|
||||||
|
# IbDir=C:\IBLiveSettings is NOT valid and will give unexpected
|
||||||
|
# results. Linux and macOS users need not use double separators,
|
||||||
|
# but they are acceptable.
|
||||||
|
#
|
||||||
|
# The default is the current working directory when IBC is
|
||||||
|
# started.
|
||||||
|
|
||||||
|
IbDir=/root/Jts
|
||||||
|
|
||||||
|
|
||||||
|
# Store settings on server
|
||||||
|
# ------------------------
|
||||||
|
#
|
||||||
|
# If you wish to store a copy of your TWS settings on IB's
|
||||||
|
# servers as well as locally on your computer, set this to
|
||||||
|
# 'yes': this enables you to run TWS on different computers
|
||||||
|
# with the same configuration, market data lines, etc. If set
|
||||||
|
# to 'no', running TWS on different computers will not share the
|
||||||
|
# same settings. If no value is specified, TWS will obtain its
|
||||||
|
# settings from the same place as the last time this user logged
|
||||||
|
# in (whether manually or using IBC).
|
||||||
|
|
||||||
|
StoreSettingsOnServer=
|
||||||
|
|
||||||
|
|
||||||
|
# Minimize TWS on startup
|
||||||
|
# -----------------------
|
||||||
|
#
|
||||||
|
# Set to 'yes' to minimize TWS when it starts:
|
||||||
|
|
||||||
|
MinimizeMainWindow=no
|
||||||
|
|
||||||
|
|
||||||
|
# Existing Session Detected Action
|
||||||
|
# --------------------------------
|
||||||
|
#
|
||||||
|
# When a user logs on to an IBKR account for trading purposes by any means, the
|
||||||
|
# IBKR account server checks to see whether the account is already logged in
|
||||||
|
# elsewhere. If so, a dialog is displayed to both the users that enables them
|
||||||
|
# to determine what happens next. The 'ExistingSessionDetectedAction' setting
|
||||||
|
# instructs TWS how to proceed when it displays this dialog:
|
||||||
|
#
|
||||||
|
# * If the new TWS session is set to 'secondary', the existing session continues
|
||||||
|
# and the new session terminates. Thus a secondary TWS session can never
|
||||||
|
# override any other session.
|
||||||
|
#
|
||||||
|
# * If the existing TWS session is set to 'primary', the existing session
|
||||||
|
# continues and the new session terminates (even if the new session is also
|
||||||
|
# set to primary). Thus a primary TWS session can never be overridden by
|
||||||
|
# any new session).
|
||||||
|
#
|
||||||
|
# * If both the existing and the new TWS sessions are set to 'primaryoverride',
|
||||||
|
# the existing session terminates and the new session proceeds.
|
||||||
|
#
|
||||||
|
# * If the existing TWS session is set to 'manual', the user must handle the
|
||||||
|
# dialog.
|
||||||
|
#
|
||||||
|
# The difference between 'primary' and 'primaryoverride' is that a
|
||||||
|
# 'primaryoverride' session can be overriden over by a new 'primary' session,
|
||||||
|
# but a 'primary' session cannot be overriden by any other session.
|
||||||
|
#
|
||||||
|
# When set to 'primary', if another TWS session is started and manually told to
|
||||||
|
# end the 'primary' session, the 'primary' session is automatically reconnected.
|
||||||
|
#
|
||||||
|
# The default is 'manual'.
|
||||||
|
|
||||||
|
ExistingSessionDetectedAction=primary
|
||||||
|
|
||||||
|
|
||||||
|
# Override TWS API Port Number
|
||||||
|
# ----------------------------
|
||||||
|
#
|
||||||
|
# If OverrideTwsApiPort is set to an integer, IBC changes the
|
||||||
|
# 'Socket port' in TWS's API configuration to that number shortly
|
||||||
|
# after startup. Leaving the setting blank will make no change to
|
||||||
|
# the current setting. This setting is only intended for use in
|
||||||
|
# certain specialized situations where the port number needs to
|
||||||
|
# be set dynamically at run-time: most users will never need it,
|
||||||
|
# so don't use it unless you know you need it.
|
||||||
|
|
||||||
|
OverrideTwsApiPort=4002
|
||||||
|
|
||||||
|
|
||||||
|
# Read-only Login
|
||||||
|
# ---------------
|
||||||
|
#
|
||||||
|
# If ReadOnlyLogin is set to 'yes', and the user is enrolled in IB's
|
||||||
|
# account security programme, the user will not be asked to perform
|
||||||
|
# the second factor authentication action, and login to TWS will
|
||||||
|
# occur automatically in read-only mode: in this mode, placing or
|
||||||
|
# managing orders is not allowed. If set to 'no', and the user is
|
||||||
|
# enrolled in IB's account security programme, the user must perform
|
||||||
|
# the relevant second factor authentication action to complete the
|
||||||
|
# login.
|
||||||
|
|
||||||
|
# If the user is not enrolled in IB's account security programme,
|
||||||
|
# this setting is ignored. The default is 'no'.
|
||||||
|
|
||||||
|
ReadOnlyLogin=no
|
||||||
|
|
||||||
|
|
||||||
|
# Read-only API
|
||||||
|
# -------------
|
||||||
|
#
|
||||||
|
# If ReadOnlyApi is set to 'yes', API programs cannot submit, modify
|
||||||
|
# or cancel orders. If set to 'no', API programs can do these things.
|
||||||
|
# If not set, the existing TWS/Gateway configuration is unchanged.
|
||||||
|
# NB: this setting is really only supplied for the benefit of new TWS
|
||||||
|
# or Gateway instances that are being automatically installed and
|
||||||
|
# started without user intervention (eg Docker containers). Where
|
||||||
|
# a user is involved, they should use the Global Configuration to
|
||||||
|
# set the relevant checkbox (this only needs to be done once) and
|
||||||
|
# not provide a value for this setting.
|
||||||
|
|
||||||
|
ReadOnlyApi=no
|
||||||
|
|
||||||
|
|
||||||
|
# Market data size for US stocks - lots or shares
|
||||||
|
# -----------------------------------------------
|
||||||
|
#
|
||||||
|
# Since IB introduced the option of market data for US stocks showing
|
||||||
|
# bid, ask and last sizes in shares rather than lots, TWS and Gateway
|
||||||
|
# display a dialog immediately after login notifying the user about
|
||||||
|
# this and requiring user input before allowing market data to be
|
||||||
|
# accessed. The user can request that the dialog not be shown again.
|
||||||
|
#
|
||||||
|
# It is recommended that the user should handle this dialog manually
|
||||||
|
# rather than using these settings, which are provided for situations
|
||||||
|
# where the user interface is not easily accessible, or where user
|
||||||
|
# settings are not preserved between sessions (eg some Docker images).
|
||||||
|
#
|
||||||
|
# - If this setting is set to 'accept', the dialog will be handled
|
||||||
|
# automatically and the option to not show it again will be
|
||||||
|
# selected.
|
||||||
|
#
|
||||||
|
# Note that in this case, the only way to allow the dialog to be
|
||||||
|
# displayed again is to manually enable the 'Bid, Ask and Last
|
||||||
|
# Size Display Update' message in the 'Messages' section of the TWS
|
||||||
|
# configuration dialog. So you should only use 'Accept' if you are
|
||||||
|
# sure you really don't want the dialog to be displayed again, or
|
||||||
|
# you have easy access to the user interface.
|
||||||
|
#
|
||||||
|
# - If set to 'defer', the dialog will be handled automatically (so
|
||||||
|
# that market data will start), but the option to not show it again
|
||||||
|
# will not be selected, and it will be shown again after the next
|
||||||
|
# login.
|
||||||
|
#
|
||||||
|
# - If set to 'ignore', the user has to deal with the dialog manually.
|
||||||
|
#
|
||||||
|
# The default value is 'ignore'.
|
||||||
|
#
|
||||||
|
# Note if set to 'accept' or 'defer', TWS also automatically sets
|
||||||
|
# the API settings checkbox labelled 'Send market data in lots for
|
||||||
|
# US stocks for dual-mode API clients'. IBC cannot prevent this.
|
||||||
|
# However you can change this immmediately by setting
|
||||||
|
# SendMarketDataInLotsForUSstocks (see below) to 'no' .
|
||||||
|
|
||||||
|
AcceptBidAskLastSizeDisplayUpdateNotification=accept
|
||||||
|
|
||||||
|
|
||||||
|
# This setting determines whether the API settings checkbox labelled
|
||||||
|
# 'Send market data in lots for US stocks for dual-mode API clients'
|
||||||
|
# is set or cleared. If set to 'yes', the checkbox is set. If set to
|
||||||
|
# 'no' the checkbox is cleared. If defaulted, the checkbox is
|
||||||
|
# unchanged.
|
||||||
|
|
||||||
|
SendMarketDataInLotsForUSstocks=
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 4. TWS Auto-Closedown
|
||||||
|
# =============================================================================
|
||||||
|
#
|
||||||
|
# IMPORTANT NOTE: Starting with TWS 974, this setting no longer
|
||||||
|
# works properly, because IB have changed the way TWS handles its
|
||||||
|
# autologoff mechanism.
|
||||||
|
#
|
||||||
|
# You should now configure the TWS autologoff time to something
|
||||||
|
# convenient for you, and restart IBC each day.
|
||||||
|
#
|
||||||
|
# Alternatively, discontinue use of IBC and use the auto-relogin
|
||||||
|
# mechanism within TWS 974 and later versions (note that the
|
||||||
|
# auto-relogin mechanism provided by IB is not available if you
|
||||||
|
# use IBC).
|
||||||
|
|
||||||
|
# Set to yes or no (lower case).
|
||||||
|
#
|
||||||
|
# yes means allow TWS to shut down automatically at its
|
||||||
|
# specified shutdown time, which is set via the TWS
|
||||||
|
# configuration menu.
|
||||||
|
#
|
||||||
|
# no means TWS never shuts down automatically.
|
||||||
|
#
|
||||||
|
# NB: IB recommends that you do not keep TWS running
|
||||||
|
# continuously. If you set this setting to 'no', you may
|
||||||
|
# experience incorrect TWS operation.
|
||||||
|
#
|
||||||
|
# NB: the default for this setting is 'no'. Since this will
|
||||||
|
# only work properly with TWS versions earlier than 974, you
|
||||||
|
# should explicitly set this to 'yes' for version 974 and later.
|
||||||
|
|
||||||
|
IbAutoClosedown=yes
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 5. TWS Tidy Closedown Time
|
||||||
|
# =============================================================================
|
||||||
|
#
|
||||||
|
# NB: starting with TWS 974 this is no longer a useful option
|
||||||
|
# because both TWS and Gateway now have the same auto-logoff
|
||||||
|
# mechanism, and IBC can no longer avoid this.
|
||||||
|
#
|
||||||
|
# Note that giving this setting a value does not change TWS's
|
||||||
|
# auto-logoff in any way: any setting will be additional to the
|
||||||
|
# TWS auto-logoff.
|
||||||
|
#
|
||||||
|
# To tell IBC to tidily close TWS at a specified time every
|
||||||
|
# day, set this value to <hh:mm>, for example:
|
||||||
|
# ClosedownAt=22:00
|
||||||
|
#
|
||||||
|
# To tell IBC to tidily close TWS at a specified day and time
|
||||||
|
# each week, set this value to <dayOfWeek hh:mm>, for example:
|
||||||
|
# ClosedownAt=Friday 22:00
|
||||||
|
#
|
||||||
|
# Note that the day of the week must be specified using your
|
||||||
|
# default locale. Also note that Java will only accept
|
||||||
|
# characters encoded to ISO 8859-1 (Latin-1). This means that
|
||||||
|
# if the day name in your default locale uses any non-Latin-1
|
||||||
|
# characters you need to encode them using Unicode escapes
|
||||||
|
# (see http://java.sun.com/docs/books/jls/third_edition/html/lexical.html#3.3
|
||||||
|
# for details). For example, to tidily close TWS at 12:00 on
|
||||||
|
# Saturday where the default locale is Simplified Chinese,
|
||||||
|
# use the following:
|
||||||
|
# #ClosedownAt=\u661F\u671F\u516D 12:00
|
||||||
|
|
||||||
|
ClosedownAt=
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 6. Other TWS Settings
|
||||||
|
# =============================================================================
|
||||||
|
|
||||||
|
# Accept Incoming Connection
|
||||||
|
# --------------------------
|
||||||
|
#
|
||||||
|
# If set to 'accept', IBC automatically accepts incoming
|
||||||
|
# API connection dialogs. If set to 'reject', IBC
|
||||||
|
# automatically rejects incoming API connection dialogs. If
|
||||||
|
# set to 'manual', the user must decide whether to accept or reject
|
||||||
|
# incoming API connection dialogs. The default is 'manual'.
|
||||||
|
# NB: it is recommended to set this to 'reject', and to explicitly
|
||||||
|
# configure which IP addresses can connect to the API in TWS's API
|
||||||
|
# configuration page, as this is much more secure (in this case, no
|
||||||
|
# incoming API connection dialogs will occur for those IP addresses).
|
||||||
|
|
||||||
|
AcceptIncomingConnectionAction=reject
|
||||||
|
|
||||||
|
|
||||||
|
# Allow Blind Trading
|
||||||
|
# -------------------
|
||||||
|
#
|
||||||
|
# If you attempt to place an order for a contract for which
|
||||||
|
# you have no market data subscription, TWS displays a dialog
|
||||||
|
# to warn you against such blind trading.
|
||||||
|
#
|
||||||
|
# yes means the dialog is dismissed as though the user had
|
||||||
|
# clicked the 'Ok' button: this means that you accept
|
||||||
|
# the risk and want the order to be submitted.
|
||||||
|
#
|
||||||
|
# no means the dialog remains on display and must be
|
||||||
|
# handled by the user.
|
||||||
|
|
||||||
|
AllowBlindTrading=yes
|
||||||
|
|
||||||
|
|
||||||
|
# Save Settings on a Schedule
|
||||||
|
# ---------------------------
|
||||||
|
#
|
||||||
|
# You can tell TWS to automatically save its settings on a schedule
|
||||||
|
# of your choosing. You can specify one or more specific times,
|
||||||
|
# like this:
|
||||||
|
#
|
||||||
|
# SaveTwsSettingsAt=HH:MM [ HH:MM]...
|
||||||
|
#
|
||||||
|
# for example:
|
||||||
|
# SaveTwsSettingsAt=08:00 12:30 17:30
|
||||||
|
#
|
||||||
|
# Or you can specify an interval at which settings are to be saved,
|
||||||
|
# optionally starting at a specific time and continuing until another
|
||||||
|
# time, like this:
|
||||||
|
#
|
||||||
|
#SaveTwsSettingsAt=Every n [{mins | hours}] [hh:mm] [hh:mm]
|
||||||
|
#
|
||||||
|
# where the first hh:mm is the start time and the second is the end
|
||||||
|
# time. If you don't specify the end time, settings are saved regularly
|
||||||
|
# from the start time till midnight. If you don't specify the start time.
|
||||||
|
# settings are saved regularly all day, beginning at 00:00. Note that
|
||||||
|
# settings will always be saved at the end time, even if that is not
|
||||||
|
# exactly one interval later than the previous time. If neither 'mins'
|
||||||
|
# nor 'hours' is specified, 'mins' is assumed. Examples:
|
||||||
|
#
|
||||||
|
# To save every 30 minutes all day starting at 00:00
|
||||||
|
#SaveTwsSettingsAt=Every 30
|
||||||
|
#SaveTwsSettingsAt=Every 30 mins
|
||||||
|
#
|
||||||
|
# To save every hour starting at 08:00 and ending at midnight
|
||||||
|
#SaveTwsSettingsAt=Every 1 hours 08:00
|
||||||
|
#SaveTwsSettingsAt=Every 1 hours 08:00 00:00
|
||||||
|
#
|
||||||
|
# To save every 90 minutes starting at 08:00 up to and including 17:43
|
||||||
|
#SaveTwsSettingsAt=Every 90 08:00 17:43
|
||||||
|
|
||||||
|
SaveTwsSettingsAt=
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 7. Settings Specific to Indian Versions of TWS
|
||||||
|
# =============================================================================
|
||||||
|
|
||||||
|
# Indian versions of TWS may display a password expiry
|
||||||
|
# notification dialog and a NSE Compliance dialog. These can be
|
||||||
|
# dismissed by setting the following to yes. By default the
|
||||||
|
# password expiry notice is not dismissed, but the NSE Compliance
|
||||||
|
# notice is dismissed.
|
||||||
|
|
||||||
|
# Warning: setting DismissPasswordExpiryWarning=yes will mean
|
||||||
|
# you will not be notified when your password is about to expire.
|
||||||
|
# You must then take other measures to ensure that your password
|
||||||
|
# is changed within the expiry period, otherwise IBC will
|
||||||
|
# not be able to login successfully.
|
||||||
|
|
||||||
|
DismissPasswordExpiryWarning=no
|
||||||
|
DismissNSEComplianceNotice=yes
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 8. IBC Command Server Settings
|
||||||
|
# =============================================================================
|
||||||
|
|
||||||
|
# Do NOT CHANGE THE FOLLOWING SETTINGS unless you
|
||||||
|
# intend to issue commands to IBC (for example
|
||||||
|
# using telnet). Note that these settings have nothing to
|
||||||
|
# do with running programs that use the TWS API.
|
||||||
|
|
||||||
|
# Command Server Port Number
|
||||||
|
# --------------------------
|
||||||
|
#
|
||||||
|
# The port number that IBC listens on for commands
|
||||||
|
# such as "STOP". DO NOT set this to the port number
|
||||||
|
# used for TWS API connections. There is no good reason
|
||||||
|
# to change this setting unless the port is used by
|
||||||
|
# some other application (typically another instance of
|
||||||
|
# IBC). The default value is 0, which tells IBC not to
|
||||||
|
# start the command server
|
||||||
|
|
||||||
|
#CommandServerPort=7462
|
||||||
|
|
||||||
|
|
||||||
|
# Permitted Command Sources
|
||||||
|
# -------------------------
|
||||||
|
#
|
||||||
|
# A comma separated list of IP addresses, or host names,
|
||||||
|
# which are allowed addresses for sending commands to
|
||||||
|
# IBC. Commands can always be sent from the
|
||||||
|
# same host as IBC is running on.
|
||||||
|
|
||||||
|
ControlFrom=127.0.0.1
|
||||||
|
|
||||||
|
|
||||||
|
# Address for Receiving Commands
|
||||||
|
# ------------------------------
|
||||||
|
#
|
||||||
|
# Specifies the IP address on which the Command Server
|
||||||
|
# is so listen. For a multi-homed host, this can be used
|
||||||
|
# to specify that connection requests are only to be
|
||||||
|
# accepted on the specified address. The default is to
|
||||||
|
# accept connection requests on all local addresses.
|
||||||
|
|
||||||
|
BindAddress=127.0.0.1
|
||||||
|
|
||||||
|
|
||||||
|
# Command Prompt
|
||||||
|
# --------------
|
||||||
|
#
|
||||||
|
# The specified string is output by the server when
|
||||||
|
# the connection is first opened and after the completion
|
||||||
|
# of each command. This can be useful if sending commands
|
||||||
|
# using an interactive program such as telnet. The default
|
||||||
|
# is that no prompt is output.
|
||||||
|
# For example:
|
||||||
|
#
|
||||||
|
# CommandPrompt=>
|
||||||
|
|
||||||
|
CommandPrompt=
|
||||||
|
|
||||||
|
|
||||||
|
# Suppress Command Server Info Messages
|
||||||
|
# -------------------------------------
|
||||||
|
#
|
||||||
|
# Some commands can return intermediate information about
|
||||||
|
# their progress. This setting controls whether such
|
||||||
|
# information is sent. The default is that such information
|
||||||
|
# is not sent.
|
||||||
|
|
||||||
|
SuppressInfoMessages=no
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
# =============================================================================
|
||||||
|
# 9. Diagnostic Settings
|
||||||
|
# =============================================================================
|
||||||
|
#
|
||||||
|
# IBC can log information about the structure of windows
|
||||||
|
# displayed by TWS. This information is useful when adding
|
||||||
|
# new features to IBC or when behaviour is not as expected.
|
||||||
|
#
|
||||||
|
# The logged information shows the hierarchical organisation
|
||||||
|
# of all the components of the window, and includes the
|
||||||
|
# current values of text boxes and labels.
|
||||||
|
#
|
||||||
|
# Note that this structure logging has a small performance
|
||||||
|
# impact, and depending on the settings can cause the logfile
|
||||||
|
# size to be significantly increased. It is therefore
|
||||||
|
# recommended that the LogStructureWhen setting be set to
|
||||||
|
# 'never' (the default) unless there is a specific reason
|
||||||
|
# that this information is needed.
|
||||||
|
|
||||||
|
|
||||||
|
# Scope of Structure Logging
|
||||||
|
# --------------------------
|
||||||
|
#
|
||||||
|
# The LogStructureScope setting indicates which windows are
|
||||||
|
# eligible for structure logging:
|
||||||
|
#
|
||||||
|
# - if set to 'known', only windows that IBC recognizes
|
||||||
|
# are eligible - these are windows that IBC has some
|
||||||
|
# interest in monitoring, usually to take some action
|
||||||
|
# on the user's behalf;
|
||||||
|
#
|
||||||
|
# - if set to 'unknown', only windows that IBC does not
|
||||||
|
# recognize are eligible. Most windows displayed by
|
||||||
|
# TWS fall into this category;
|
||||||
|
#
|
||||||
|
# - if set to 'untitled', only windows that IBC does not
|
||||||
|
# recognize and that have no title are eligible. These
|
||||||
|
# are usually message boxes or similar small windows,
|
||||||
|
#
|
||||||
|
# - if set to 'all', then every window displayed by TWS
|
||||||
|
# is eligible.
|
||||||
|
#
|
||||||
|
# The default value is 'known'.
|
||||||
|
|
||||||
|
LogStructureScope=all
|
||||||
|
|
||||||
|
|
||||||
|
# When to Log Window Structure
|
||||||
|
# ----------------------------
|
||||||
|
#
|
||||||
|
# The LogStructureWhen setting specifies the circumstances
|
||||||
|
# when eligible TWS windows have their structure logged:
|
||||||
|
#
|
||||||
|
# - if set to 'open' or 'yes' or 'true', IBC logs the
|
||||||
|
# structure of an eligible window the first time it
|
||||||
|
# is encountered;
|
||||||
|
#
|
||||||
|
# - if set to 'activate', the structure is logged every
|
||||||
|
# time an eligible window is made active;
|
||||||
|
#
|
||||||
|
# - if set to 'never' or 'no' or 'false', structure
|
||||||
|
# information is never logged.
|
||||||
|
#
|
||||||
|
# The default value is 'never'.
|
||||||
|
|
||||||
|
LogStructureWhen=never
|
||||||
|
|
||||||
|
|
||||||
|
# DEPRECATED SETTING
|
||||||
|
# ------------------
|
||||||
|
#
|
||||||
|
# LogComponents - THIS SETTING WILL BE REMOVED IN A FUTURE
|
||||||
|
# RELEASE
|
||||||
|
#
|
||||||
|
# If LogComponents is set to any value, this is equivalent
|
||||||
|
# to setting LogStructureWhen to that same value and
|
||||||
|
# LogStructureScope to 'all': the actual values of those
|
||||||
|
# settings are ignored. The default is that the values
|
||||||
|
# of LogStructureScope and LogStructureWhen are honoured.
|
||||||
|
|
||||||
|
#LogComponents=
|
||||||
|
|
||||||
|
|
||||||
|
|
|
@ -0,0 +1,33 @@
|
||||||
|
[IBGateway]
|
||||||
|
ApiOnly=true
|
||||||
|
LocalServerPort=4002
|
||||||
|
# NOTE: must be set if using IBC's "reject" mode
|
||||||
|
TrustedIPs=127.0.0.1
|
||||||
|
; RemoteHostOrderRouting=ndc1.ibllc.com
|
||||||
|
; WriteDebug=true
|
||||||
|
; RemotePortOrderRouting=4001
|
||||||
|
; useRemoteSettings=false
|
||||||
|
; tradingMode=p
|
||||||
|
; Steps=8
|
||||||
|
; colorPalletName=dark
|
||||||
|
|
||||||
|
# window geo, this may be useful for sending `xdotool` commands?
|
||||||
|
; MainWindow.Width=1986
|
||||||
|
; screenHeight=3960
|
||||||
|
|
||||||
|
|
||||||
|
[Logon]
|
||||||
|
Locale=en
|
||||||
|
# most markets are oriented around this zone
|
||||||
|
# so might as well hard code it.
|
||||||
|
TimeZone=America/New_York
|
||||||
|
UseSSL=true
|
||||||
|
displayedproxymsg=1
|
||||||
|
os_titlebar=true
|
||||||
|
s3store=true
|
||||||
|
useRemoteSettings=false
|
||||||
|
|
||||||
|
[Communication]
|
||||||
|
ctciAutoEncrypt=true
|
||||||
|
Region=usr
|
||||||
|
; Peer=cdc1.ibllc.com:4001
|
|
@ -0,0 +1,16 @@
|
||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
# start VNC server
|
||||||
|
x11vnc \
|
||||||
|
-ncache_cr \
|
||||||
|
-listen localhost \
|
||||||
|
-display :1 \
|
||||||
|
-forever \
|
||||||
|
-shared \
|
||||||
|
-logappend /var/log/x11vnc.log \
|
||||||
|
-bg \
|
||||||
|
-noipv6 \
|
||||||
|
-autoport 3003 \
|
||||||
|
# can't use this because of ``asyncvnc`` issue:
|
||||||
|
# https://github.com/barneygale/asyncvnc/issues/1
|
||||||
|
# -passwd 'ibcansmbz'
|
|
@ -35,7 +35,7 @@ log = get_logger(__name__)
|
||||||
|
|
||||||
_root_dname = 'pikerd'
|
_root_dname = 'pikerd'
|
||||||
|
|
||||||
_registry_addr = ('127.0.0.1', 1616)
|
_registry_addr = ('127.0.0.1', 6116)
|
||||||
_tractor_kwargs: dict[str, Any] = {
|
_tractor_kwargs: dict[str, Any] = {
|
||||||
# use a different registry addr then tractor's default
|
# use a different registry addr then tractor's default
|
||||||
'arbiter_addr': _registry_addr
|
'arbiter_addr': _registry_addr
|
||||||
|
@ -426,9 +426,19 @@ async def spawn_brokerd(
|
||||||
|
|
||||||
# ask `pikerd` to spawn a new sub-actor and manage it under its
|
# ask `pikerd` to spawn a new sub-actor and manage it under its
|
||||||
# actor nursery
|
# actor nursery
|
||||||
|
modpath = brokermod.__name__
|
||||||
|
broker_enable = [modpath]
|
||||||
|
for submodname in getattr(
|
||||||
|
brokermod,
|
||||||
|
'__enable_modules__',
|
||||||
|
[],
|
||||||
|
):
|
||||||
|
subpath = f'{modpath}.{submodname}'
|
||||||
|
broker_enable.append(subpath)
|
||||||
|
|
||||||
portal = await _services.actor_n.start_actor(
|
portal = await _services.actor_n.start_actor(
|
||||||
dname,
|
dname,
|
||||||
enable_modules=_data_mods + [brokermod.__name__],
|
enable_modules=_data_mods + broker_enable,
|
||||||
loglevel=loglevel,
|
loglevel=loglevel,
|
||||||
debug_mode=_services.debug_mode,
|
debug_mode=_services.debug_mode,
|
||||||
**tractor_kwargs
|
**tractor_kwargs
|
||||||
|
|
2617
piker/brokers/ib.py
2617
piker/brokers/ib.py
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1,67 @@
|
||||||
|
# piker: trading gear for hackers
|
||||||
|
# Copyright (C) Tyler Goodlet (in stewardship for pikers)
|
||||||
|
|
||||||
|
# This program is free software: you can redistribute it and/or modify
|
||||||
|
# it under the terms of the GNU Affero General Public License as published by
|
||||||
|
# the Free Software Foundation, either version 3 of the License, or
|
||||||
|
# (at your option) any later version.
|
||||||
|
|
||||||
|
# This program is distributed in the hope that it will be useful,
|
||||||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
# GNU Affero General Public License for more details.
|
||||||
|
|
||||||
|
# You should have received a copy of the GNU Affero General Public License
|
||||||
|
# along with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
"""
|
||||||
|
Interactive Brokers API backend.
|
||||||
|
|
||||||
|
Sub-modules within break into the core functionalities:
|
||||||
|
|
||||||
|
- ``broker.py`` part for orders / trading endpoints
|
||||||
|
- ``data.py`` for real-time data feed endpoints
|
||||||
|
|
||||||
|
- ``client.py`` for the core API machinery which is ``trio``-ized
|
||||||
|
wrapping around ``ib_insync``.
|
||||||
|
|
||||||
|
- ``report.py`` for the hackery to build manual pp calcs
|
||||||
|
to avoid ib's absolute bullshit FIFO style position
|
||||||
|
tracking..
|
||||||
|
|
||||||
|
"""
|
||||||
|
from .api import (
|
||||||
|
get_client,
|
||||||
|
)
|
||||||
|
from .feed import (
|
||||||
|
open_history_client,
|
||||||
|
open_symbol_search,
|
||||||
|
stream_quotes,
|
||||||
|
)
|
||||||
|
from .broker import trades_dialogue
|
||||||
|
|
||||||
|
__all__ = [
|
||||||
|
'get_client',
|
||||||
|
'trades_dialogue',
|
||||||
|
'open_history_client',
|
||||||
|
'open_symbol_search',
|
||||||
|
'stream_quotes',
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
# tractor RPC enable arg
|
||||||
|
__enable_modules__: list[str] = [
|
||||||
|
'api',
|
||||||
|
'feed',
|
||||||
|
'broker',
|
||||||
|
]
|
||||||
|
|
||||||
|
# passed to ``tractor.ActorNursery.start_actor()``
|
||||||
|
_spawn_kwargs = {
|
||||||
|
'infect_asyncio': True,
|
||||||
|
}
|
||||||
|
|
||||||
|
# annotation to let backend agnostic code
|
||||||
|
# know if ``brokerd`` should be spawned with
|
||||||
|
# ``tractor``'s aio mode.
|
||||||
|
_infect_asyncio: bool = True
|
File diff suppressed because it is too large
Load Diff
|
@ -0,0 +1,590 @@
|
||||||
|
# piker: trading gear for hackers
|
||||||
|
# Copyright (C) Tyler Goodlet (in stewardship for pikers)
|
||||||
|
|
||||||
|
# This program is free software: you can redistribute it and/or modify
|
||||||
|
# it under the terms of the GNU Affero General Public License as published by
|
||||||
|
# the Free Software Foundation, either version 3 of the License, or
|
||||||
|
# (at your option) any later version.
|
||||||
|
|
||||||
|
# This program is distributed in the hope that it will be useful,
|
||||||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
# GNU Affero General Public License for more details.
|
||||||
|
|
||||||
|
# You should have received a copy of the GNU Affero General Public License
|
||||||
|
# along with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
|
"""
|
||||||
|
Order and trades endpoints for use with ``piker``'s EMS.
|
||||||
|
|
||||||
|
"""
|
||||||
|
from __future__ import annotations
|
||||||
|
from dataclasses import asdict
|
||||||
|
from functools import partial
|
||||||
|
from pprint import pformat
|
||||||
|
import time
|
||||||
|
from typing import (
|
||||||
|
Any,
|
||||||
|
Optional,
|
||||||
|
AsyncIterator,
|
||||||
|
)
|
||||||
|
|
||||||
|
import trio
|
||||||
|
from trio_typing import TaskStatus
|
||||||
|
import tractor
|
||||||
|
from ib_insync.contract import (
|
||||||
|
Contract,
|
||||||
|
Option,
|
||||||
|
)
|
||||||
|
from ib_insync.order import (
|
||||||
|
Trade,
|
||||||
|
OrderStatus,
|
||||||
|
)
|
||||||
|
from ib_insync.objects import (
|
||||||
|
Fill,
|
||||||
|
Execution,
|
||||||
|
)
|
||||||
|
from ib_insync.objects import Position
|
||||||
|
|
||||||
|
from piker import config
|
||||||
|
from piker.log import get_console_log
|
||||||
|
from piker.clearing._messages import (
|
||||||
|
BrokerdOrder,
|
||||||
|
BrokerdOrderAck,
|
||||||
|
BrokerdStatus,
|
||||||
|
BrokerdPosition,
|
||||||
|
BrokerdCancel,
|
||||||
|
BrokerdFill,
|
||||||
|
BrokerdError,
|
||||||
|
)
|
||||||
|
from .api import (
|
||||||
|
_accounts2clients,
|
||||||
|
_adhoc_futes_set,
|
||||||
|
log,
|
||||||
|
get_config,
|
||||||
|
open_client_proxies,
|
||||||
|
Client,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def pack_position(
|
||||||
|
pos: Position
|
||||||
|
|
||||||
|
) -> dict[str, Any]:
|
||||||
|
con = pos.contract
|
||||||
|
|
||||||
|
if isinstance(con, Option):
|
||||||
|
# TODO: option symbol parsing and sane display:
|
||||||
|
symbol = con.localSymbol.replace(' ', '')
|
||||||
|
|
||||||
|
else:
|
||||||
|
# TODO: lookup fqsn even for derivs.
|
||||||
|
symbol = con.symbol.lower()
|
||||||
|
|
||||||
|
exch = (con.primaryExchange or con.exchange).lower()
|
||||||
|
symkey = '.'.join((symbol, exch))
|
||||||
|
if not exch:
|
||||||
|
# attempt to lookup the symbol from our
|
||||||
|
# hacked set..
|
||||||
|
for sym in _adhoc_futes_set:
|
||||||
|
if symbol in sym:
|
||||||
|
symkey = sym
|
||||||
|
break
|
||||||
|
|
||||||
|
expiry = con.lastTradeDateOrContractMonth
|
||||||
|
if expiry:
|
||||||
|
symkey += f'.{expiry}'
|
||||||
|
|
||||||
|
# TODO: options contracts into a sane format..
|
||||||
|
|
||||||
|
return BrokerdPosition(
|
||||||
|
broker='ib',
|
||||||
|
account=pos.account,
|
||||||
|
symbol=symkey,
|
||||||
|
currency=con.currency,
|
||||||
|
size=float(pos.position),
|
||||||
|
avg_price=float(pos.avgCost) / float(con.multiplier or 1.0),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
async def handle_order_requests(
|
||||||
|
|
||||||
|
ems_order_stream: tractor.MsgStream,
|
||||||
|
accounts_def: dict[str, str],
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
|
||||||
|
request_msg: dict
|
||||||
|
async for request_msg in ems_order_stream:
|
||||||
|
log.info(f'Received order request {request_msg}')
|
||||||
|
|
||||||
|
action = request_msg['action']
|
||||||
|
account = request_msg['account']
|
||||||
|
|
||||||
|
acct_number = accounts_def.get(account)
|
||||||
|
if not acct_number:
|
||||||
|
log.error(
|
||||||
|
f'An IB account number for name {account} is not found?\n'
|
||||||
|
'Make sure you have all TWS and GW instances running.'
|
||||||
|
)
|
||||||
|
await ems_order_stream.send(BrokerdError(
|
||||||
|
oid=request_msg['oid'],
|
||||||
|
symbol=request_msg['symbol'],
|
||||||
|
reason=f'No account found: `{account}` ?',
|
||||||
|
).dict())
|
||||||
|
continue
|
||||||
|
|
||||||
|
client = _accounts2clients.get(account)
|
||||||
|
if not client:
|
||||||
|
log.error(
|
||||||
|
f'An IB client for account name {account} is not found.\n'
|
||||||
|
'Make sure you have all TWS and GW instances running.'
|
||||||
|
)
|
||||||
|
await ems_order_stream.send(BrokerdError(
|
||||||
|
oid=request_msg['oid'],
|
||||||
|
symbol=request_msg['symbol'],
|
||||||
|
reason=f'No api client loaded for account: `{account}` ?',
|
||||||
|
).dict())
|
||||||
|
continue
|
||||||
|
|
||||||
|
if action in {'buy', 'sell'}:
|
||||||
|
# validate
|
||||||
|
order = BrokerdOrder(**request_msg)
|
||||||
|
|
||||||
|
# call our client api to submit the order
|
||||||
|
reqid = client.submit_limit(
|
||||||
|
oid=order.oid,
|
||||||
|
symbol=order.symbol,
|
||||||
|
price=order.price,
|
||||||
|
action=order.action,
|
||||||
|
size=order.size,
|
||||||
|
account=acct_number,
|
||||||
|
|
||||||
|
# XXX: by default 0 tells ``ib_insync`` methods that
|
||||||
|
# there is no existing order so ask the client to create
|
||||||
|
# a new one (which it seems to do by allocating an int
|
||||||
|
# counter - collision prone..)
|
||||||
|
reqid=order.reqid,
|
||||||
|
)
|
||||||
|
if reqid is None:
|
||||||
|
await ems_order_stream.send(BrokerdError(
|
||||||
|
oid=request_msg['oid'],
|
||||||
|
symbol=request_msg['symbol'],
|
||||||
|
reason='Order already active?',
|
||||||
|
).dict())
|
||||||
|
|
||||||
|
# deliver ack that order has been submitted to broker routing
|
||||||
|
await ems_order_stream.send(
|
||||||
|
BrokerdOrderAck(
|
||||||
|
# ems order request id
|
||||||
|
oid=order.oid,
|
||||||
|
# broker specific request id
|
||||||
|
reqid=reqid,
|
||||||
|
time_ns=time.time_ns(),
|
||||||
|
account=account,
|
||||||
|
).dict()
|
||||||
|
)
|
||||||
|
|
||||||
|
elif action == 'cancel':
|
||||||
|
msg = BrokerdCancel(**request_msg)
|
||||||
|
client.submit_cancel(reqid=msg.reqid)
|
||||||
|
|
||||||
|
else:
|
||||||
|
log.error(f'Unknown order command: {request_msg}')
|
||||||
|
|
||||||
|
|
||||||
|
async def recv_trade_updates(
|
||||||
|
|
||||||
|
client: Client,
|
||||||
|
to_trio: trio.abc.SendChannel,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
"""Stream a ticker using the std L1 api.
|
||||||
|
"""
|
||||||
|
client.inline_errors(to_trio)
|
||||||
|
|
||||||
|
# sync with trio task
|
||||||
|
to_trio.send_nowait(None)
|
||||||
|
|
||||||
|
def push_tradesies(eventkit_obj, obj, fill=None):
|
||||||
|
"""Push events to trio task.
|
||||||
|
|
||||||
|
"""
|
||||||
|
if fill is not None:
|
||||||
|
# execution details event
|
||||||
|
item = ('fill', (obj, fill))
|
||||||
|
|
||||||
|
elif eventkit_obj.name() == 'positionEvent':
|
||||||
|
item = ('position', obj)
|
||||||
|
|
||||||
|
else:
|
||||||
|
item = ('status', obj)
|
||||||
|
|
||||||
|
log.info(f'eventkit event ->\n{pformat(item)}')
|
||||||
|
|
||||||
|
try:
|
||||||
|
to_trio.send_nowait(item)
|
||||||
|
except trio.BrokenResourceError:
|
||||||
|
log.exception(f'Disconnected from {eventkit_obj} updates')
|
||||||
|
eventkit_obj.disconnect(push_tradesies)
|
||||||
|
|
||||||
|
# hook up to the weird eventkit object - event stream api
|
||||||
|
for ev_name in [
|
||||||
|
'orderStatusEvent', # all order updates
|
||||||
|
'execDetailsEvent', # all "fill" updates
|
||||||
|
'positionEvent', # avg price updates per symbol per account
|
||||||
|
|
||||||
|
# 'commissionReportEvent',
|
||||||
|
# XXX: ugh, it is a separate event from IB and it's
|
||||||
|
# emitted as follows:
|
||||||
|
# self.ib.commissionReportEvent.emit(trade, fill, report)
|
||||||
|
|
||||||
|
# XXX: not sure yet if we need these
|
||||||
|
# 'updatePortfolioEvent',
|
||||||
|
|
||||||
|
# XXX: these all seem to be weird ib_insync intrernal
|
||||||
|
# events that we probably don't care that much about
|
||||||
|
# given the internal design is wonky af..
|
||||||
|
# 'newOrderEvent',
|
||||||
|
# 'orderModifyEvent',
|
||||||
|
# 'cancelOrderEvent',
|
||||||
|
# 'openOrderEvent',
|
||||||
|
]:
|
||||||
|
eventkit_obj = getattr(client.ib, ev_name)
|
||||||
|
handler = partial(push_tradesies, eventkit_obj)
|
||||||
|
eventkit_obj.connect(handler)
|
||||||
|
|
||||||
|
# let the engine run and stream
|
||||||
|
await client.ib.disconnectedEvent
|
||||||
|
|
||||||
|
|
||||||
|
@tractor.context
|
||||||
|
async def trades_dialogue(
|
||||||
|
|
||||||
|
ctx: tractor.Context,
|
||||||
|
loglevel: str = None,
|
||||||
|
|
||||||
|
) -> AsyncIterator[dict[str, Any]]:
|
||||||
|
|
||||||
|
# XXX: required to propagate ``tractor`` loglevel to piker logging
|
||||||
|
get_console_log(loglevel or tractor.current_actor().loglevel)
|
||||||
|
|
||||||
|
accounts_def = config.load_accounts(['ib'])
|
||||||
|
|
||||||
|
global _client_cache
|
||||||
|
|
||||||
|
# deliver positions to subscriber before anything else
|
||||||
|
all_positions = []
|
||||||
|
accounts = set()
|
||||||
|
clients: list[tuple[Client, trio.MemoryReceiveChannel]] = []
|
||||||
|
|
||||||
|
async with (
|
||||||
|
trio.open_nursery() as nurse,
|
||||||
|
open_client_proxies() as (proxies, aioclients),
|
||||||
|
):
|
||||||
|
for account, proxy in proxies.items():
|
||||||
|
|
||||||
|
client = aioclients[account]
|
||||||
|
|
||||||
|
async def open_stream(
|
||||||
|
task_status: TaskStatus[
|
||||||
|
trio.abc.ReceiveChannel
|
||||||
|
] = trio.TASK_STATUS_IGNORED,
|
||||||
|
):
|
||||||
|
# each api client has a unique event stream
|
||||||
|
async with tractor.to_asyncio.open_channel_from(
|
||||||
|
recv_trade_updates,
|
||||||
|
client=client,
|
||||||
|
) as (first, trade_event_stream):
|
||||||
|
|
||||||
|
task_status.started(trade_event_stream)
|
||||||
|
await trio.sleep_forever()
|
||||||
|
|
||||||
|
trade_event_stream = await nurse.start(open_stream)
|
||||||
|
|
||||||
|
clients.append((client, trade_event_stream))
|
||||||
|
|
||||||
|
assert account in accounts_def
|
||||||
|
accounts.add(account)
|
||||||
|
|
||||||
|
for client in aioclients.values():
|
||||||
|
for pos in client.positions():
|
||||||
|
|
||||||
|
msg = pack_position(pos)
|
||||||
|
msg.account = accounts_def.inverse[msg.account]
|
||||||
|
|
||||||
|
assert msg.account in accounts, (
|
||||||
|
f'Position for unknown account: {msg.account}')
|
||||||
|
|
||||||
|
all_positions.append(msg.dict())
|
||||||
|
|
||||||
|
trades: list[dict] = []
|
||||||
|
for proxy in proxies.values():
|
||||||
|
trades.append(await proxy.trades())
|
||||||
|
|
||||||
|
log.info(f'Loaded {len(trades)} from this session')
|
||||||
|
# TODO: write trades to local ``trades.toml``
|
||||||
|
# - use above per-session trades data and write to local file
|
||||||
|
# - get the "flex reports" working and pull historical data and
|
||||||
|
# also save locally.
|
||||||
|
|
||||||
|
await ctx.started((
|
||||||
|
all_positions,
|
||||||
|
tuple(name for name in accounts_def if name in accounts),
|
||||||
|
))
|
||||||
|
|
||||||
|
async with (
|
||||||
|
ctx.open_stream() as ems_stream,
|
||||||
|
trio.open_nursery() as n,
|
||||||
|
):
|
||||||
|
# start order request handler **before** local trades event loop
|
||||||
|
n.start_soon(handle_order_requests, ems_stream, accounts_def)
|
||||||
|
|
||||||
|
# allocate event relay tasks for each client connection
|
||||||
|
for client, stream in clients:
|
||||||
|
n.start_soon(
|
||||||
|
deliver_trade_events,
|
||||||
|
stream,
|
||||||
|
ems_stream,
|
||||||
|
accounts_def
|
||||||
|
)
|
||||||
|
|
||||||
|
# block until cancelled
|
||||||
|
await trio.sleep_forever()
|
||||||
|
|
||||||
|
|
||||||
|
async def deliver_trade_events(
|
||||||
|
|
||||||
|
trade_event_stream: trio.MemoryReceiveChannel,
|
||||||
|
ems_stream: tractor.MsgStream,
|
||||||
|
accounts_def: dict[str, str],
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
'''Format and relay all trade events for a given client to the EMS.
|
||||||
|
|
||||||
|
'''
|
||||||
|
action_map = {'BOT': 'buy', 'SLD': 'sell'}
|
||||||
|
|
||||||
|
# TODO: for some reason we can receive a ``None`` here when the
|
||||||
|
# ib-gw goes down? Not sure exactly how that's happening looking
|
||||||
|
# at the eventkit code above but we should probably handle it...
|
||||||
|
async for event_name, item in trade_event_stream:
|
||||||
|
|
||||||
|
log.info(f'ib sending {event_name}:\n{pformat(item)}')
|
||||||
|
|
||||||
|
# TODO: templating the ib statuses in comparison with other
|
||||||
|
# brokers is likely the way to go:
|
||||||
|
# https://interactivebrokers.github.io/tws-api/interfaceIBApi_1_1EWrapper.html#a17f2a02d6449710b6394d0266a353313
|
||||||
|
# short list:
|
||||||
|
# - PendingSubmit
|
||||||
|
# - PendingCancel
|
||||||
|
# - PreSubmitted (simulated orders)
|
||||||
|
# - ApiCancelled (cancelled by client before submission
|
||||||
|
# to routing)
|
||||||
|
# - Cancelled
|
||||||
|
# - Filled
|
||||||
|
# - Inactive (reject or cancelled but not by trader)
|
||||||
|
|
||||||
|
# XXX: here's some other sucky cases from the api
|
||||||
|
# - short-sale but securities haven't been located, in this
|
||||||
|
# case we should probably keep the order in some kind of
|
||||||
|
# weird state or cancel it outright?
|
||||||
|
|
||||||
|
# status='PendingSubmit', message=''),
|
||||||
|
# status='Cancelled', message='Error 404,
|
||||||
|
# reqId 1550: Order held while securities are located.'),
|
||||||
|
# status='PreSubmitted', message='')],
|
||||||
|
|
||||||
|
if event_name == 'status':
|
||||||
|
|
||||||
|
# XXX: begin normalization of nonsense ib_insync internal
|
||||||
|
# object-state tracking representations...
|
||||||
|
|
||||||
|
# unwrap needed data from ib_insync internal types
|
||||||
|
trade: Trade = item
|
||||||
|
status: OrderStatus = trade.orderStatus
|
||||||
|
|
||||||
|
# skip duplicate filled updates - we get the deats
|
||||||
|
# from the execution details event
|
||||||
|
msg = BrokerdStatus(
|
||||||
|
|
||||||
|
reqid=trade.order.orderId,
|
||||||
|
time_ns=time.time_ns(), # cuz why not
|
||||||
|
account=accounts_def.inverse[trade.order.account],
|
||||||
|
|
||||||
|
# everyone doin camel case..
|
||||||
|
status=status.status.lower(), # force lower case
|
||||||
|
|
||||||
|
filled=status.filled,
|
||||||
|
reason=status.whyHeld,
|
||||||
|
|
||||||
|
# this seems to not be necessarily up to date in the
|
||||||
|
# execDetails event.. so we have to send it here I guess?
|
||||||
|
remaining=status.remaining,
|
||||||
|
|
||||||
|
broker_details={'name': 'ib'},
|
||||||
|
)
|
||||||
|
|
||||||
|
elif event_name == 'fill':
|
||||||
|
|
||||||
|
# for wtv reason this is a separate event type
|
||||||
|
# from IB, not sure why it's needed other then for extra
|
||||||
|
# complexity and over-engineering :eyeroll:.
|
||||||
|
# we may just end up dropping these events (or
|
||||||
|
# translating them to ``Status`` msgs) if we can
|
||||||
|
# show the equivalent status events are no more latent.
|
||||||
|
|
||||||
|
# unpack ib_insync types
|
||||||
|
# pep-0526 style:
|
||||||
|
# https://www.python.org/dev/peps/pep-0526/#global-and-local-variable-annotations
|
||||||
|
trade: Trade
|
||||||
|
fill: Fill
|
||||||
|
trade, fill = item
|
||||||
|
execu: Execution = fill.execution
|
||||||
|
|
||||||
|
# TODO: normalize out commissions details?
|
||||||
|
details = {
|
||||||
|
'contract': asdict(fill.contract),
|
||||||
|
'execution': asdict(fill.execution),
|
||||||
|
'commissions': asdict(fill.commissionReport),
|
||||||
|
'broker_time': execu.time, # supposedly server fill time
|
||||||
|
'name': 'ib',
|
||||||
|
}
|
||||||
|
|
||||||
|
msg = BrokerdFill(
|
||||||
|
# should match the value returned from `.submit_limit()`
|
||||||
|
reqid=execu.orderId,
|
||||||
|
time_ns=time.time_ns(), # cuz why not
|
||||||
|
|
||||||
|
action=action_map[execu.side],
|
||||||
|
size=execu.shares,
|
||||||
|
price=execu.price,
|
||||||
|
|
||||||
|
broker_details=details,
|
||||||
|
# XXX: required by order mode currently
|
||||||
|
broker_time=details['broker_time'],
|
||||||
|
|
||||||
|
)
|
||||||
|
|
||||||
|
elif event_name == 'error':
|
||||||
|
|
||||||
|
err: dict = item
|
||||||
|
|
||||||
|
# f$#$% gawd dammit insync..
|
||||||
|
con = err['contract']
|
||||||
|
if isinstance(con, Contract):
|
||||||
|
err['contract'] = asdict(con)
|
||||||
|
|
||||||
|
if err['reqid'] == -1:
|
||||||
|
log.error(f'TWS external order error:\n{pformat(err)}')
|
||||||
|
|
||||||
|
# TODO: what schema for this msg if we're going to make it
|
||||||
|
# portable across all backends?
|
||||||
|
# msg = BrokerdError(**err)
|
||||||
|
continue
|
||||||
|
|
||||||
|
elif event_name == 'position':
|
||||||
|
msg = pack_position(item)
|
||||||
|
msg.account = accounts_def.inverse[msg.account]
|
||||||
|
|
||||||
|
elif event_name == 'event':
|
||||||
|
|
||||||
|
# it's either a general system status event or an external
|
||||||
|
# trade event?
|
||||||
|
log.info(f"TWS system status: \n{pformat(item)}")
|
||||||
|
|
||||||
|
# TODO: support this again but needs parsing at the callback
|
||||||
|
# level...
|
||||||
|
# reqid = item.get('reqid', 0)
|
||||||
|
# if getattr(msg, 'reqid', 0) < -1:
|
||||||
|
# log.info(f"TWS triggered trade\n{pformat(msg.dict())}")
|
||||||
|
|
||||||
|
continue
|
||||||
|
|
||||||
|
# msg.reqid = 'tws-' + str(-1 * reqid)
|
||||||
|
|
||||||
|
# mark msg as from "external system"
|
||||||
|
# TODO: probably something better then this.. and start
|
||||||
|
# considering multiplayer/group trades tracking
|
||||||
|
# msg.broker_details['external_src'] = 'tws'
|
||||||
|
|
||||||
|
# XXX: we always serialize to a dict for msgpack
|
||||||
|
# translations, ideally we can move to an msgspec (or other)
|
||||||
|
# encoder # that can be enabled in ``tractor`` ahead of
|
||||||
|
# time so we can pass through the message types directly.
|
||||||
|
await ems_stream.send(msg.dict())
|
||||||
|
|
||||||
|
|
||||||
|
def load_flex_trades(
|
||||||
|
path: Optional[str] = None,
|
||||||
|
|
||||||
|
) -> dict[str, str]:
|
||||||
|
|
||||||
|
from pprint import pprint
|
||||||
|
from ib_insync import flexreport, util
|
||||||
|
|
||||||
|
conf = get_config()
|
||||||
|
|
||||||
|
if not path:
|
||||||
|
# load ``brokers.toml`` and try to get the flex
|
||||||
|
# token and query id that must be previously defined
|
||||||
|
# by the user.
|
||||||
|
token = conf.get('flex_token')
|
||||||
|
if not token:
|
||||||
|
raise ValueError(
|
||||||
|
'You must specify a ``flex_token`` field in your'
|
||||||
|
'`brokers.toml` in order load your trade log, see our'
|
||||||
|
'intructions for how to set this up here:\n'
|
||||||
|
'PUT LINK HERE!'
|
||||||
|
)
|
||||||
|
|
||||||
|
qid = conf['flex_trades_query_id']
|
||||||
|
|
||||||
|
# TODO: hack this into our logging
|
||||||
|
# system like we do with the API client..
|
||||||
|
util.logToConsole()
|
||||||
|
|
||||||
|
# TODO: rewrite the query part of this with async..httpx?
|
||||||
|
report = flexreport.FlexReport(
|
||||||
|
token=token,
|
||||||
|
queryId=qid,
|
||||||
|
)
|
||||||
|
|
||||||
|
else:
|
||||||
|
# XXX: another project we could potentially look at,
|
||||||
|
# https://pypi.org/project/ibflex/
|
||||||
|
report = flexreport.FlexReport(path=path)
|
||||||
|
|
||||||
|
trade_entries = report.extract('Trade')
|
||||||
|
trades = {
|
||||||
|
# XXX: LOL apparently ``toml`` has a bug
|
||||||
|
# where a section key error will show up in the write
|
||||||
|
# if you leave this as an ``int``?
|
||||||
|
str(t.__dict__['tradeID']): t.__dict__
|
||||||
|
for t in trade_entries
|
||||||
|
}
|
||||||
|
|
||||||
|
ln = len(trades)
|
||||||
|
log.info(f'Loaded {ln} trades from flex query')
|
||||||
|
|
||||||
|
trades_by_account = {}
|
||||||
|
for tid, trade in trades.items():
|
||||||
|
trades_by_account.setdefault(
|
||||||
|
# oddly for some so-called "BookTrade" entries
|
||||||
|
# this field seems to be blank, no cuckin clue.
|
||||||
|
# trade['ibExecID']
|
||||||
|
str(trade['accountId']), {}
|
||||||
|
)[tid] = trade
|
||||||
|
|
||||||
|
section = {'ib': trades_by_account}
|
||||||
|
pprint(section)
|
||||||
|
|
||||||
|
# TODO: load the config first and append in
|
||||||
|
# the new trades loaded here..
|
||||||
|
try:
|
||||||
|
config.write(section, 'trades')
|
||||||
|
except KeyError:
|
||||||
|
import pdbpp; pdbpp.set_trace() # noqa
|
||||||
|
|
||||||
|
|
||||||
|
if __name__ == '__main__':
|
||||||
|
load_flex_trades()
|
|
@ -0,0 +1,938 @@
|
||||||
|
# piker: trading gear for hackers
|
||||||
|
# Copyright (C) Tyler Goodlet (in stewardship for pikers)
|
||||||
|
|
||||||
|
# This program is free software: you can redistribute it and/or modify
|
||||||
|
# it under the terms of the GNU Affero General Public License as published by
|
||||||
|
# the Free Software Foundation, either version 3 of the License, or
|
||||||
|
# (at your option) any later version.
|
||||||
|
|
||||||
|
# This program is distributed in the hope that it will be useful,
|
||||||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
# GNU Affero General Public License for more details.
|
||||||
|
|
||||||
|
# You should have received a copy of the GNU Affero General Public License
|
||||||
|
# along with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
|
"""
|
||||||
|
Data feed endpoints pre-wrapped and ready for use with ``tractor``/``trio``.
|
||||||
|
|
||||||
|
"""
|
||||||
|
from __future__ import annotations
|
||||||
|
import asyncio
|
||||||
|
from contextlib import asynccontextmanager as acm
|
||||||
|
from dataclasses import asdict
|
||||||
|
from datetime import datetime
|
||||||
|
from math import isnan
|
||||||
|
import time
|
||||||
|
from typing import (
|
||||||
|
Callable,
|
||||||
|
Optional,
|
||||||
|
Awaitable,
|
||||||
|
)
|
||||||
|
|
||||||
|
from async_generator import aclosing
|
||||||
|
from fuzzywuzzy import process as fuzzy
|
||||||
|
import numpy as np
|
||||||
|
import pendulum
|
||||||
|
import tractor
|
||||||
|
import trio
|
||||||
|
from trio_typing import TaskStatus
|
||||||
|
|
||||||
|
from piker.data._sharedmem import ShmArray
|
||||||
|
from .._util import SymbolNotFound, NoData
|
||||||
|
from .api import (
|
||||||
|
_adhoc_futes_set,
|
||||||
|
log,
|
||||||
|
load_aio_clients,
|
||||||
|
ibis,
|
||||||
|
MethodProxy,
|
||||||
|
open_client_proxies,
|
||||||
|
get_preferred_data_client,
|
||||||
|
Ticker,
|
||||||
|
RequestError,
|
||||||
|
Contract,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
# https://interactivebrokers.github.io/tws-api/tick_types.html
|
||||||
|
tick_types = {
|
||||||
|
77: 'trade',
|
||||||
|
|
||||||
|
# a "utrade" aka an off exchange "unreportable" (dark) vlm:
|
||||||
|
# https://interactivebrokers.github.io/tws-api/tick_types.html#rt_volume
|
||||||
|
48: 'dark_trade',
|
||||||
|
|
||||||
|
# standard L1 ticks
|
||||||
|
0: 'bsize',
|
||||||
|
1: 'bid',
|
||||||
|
2: 'ask',
|
||||||
|
3: 'asize',
|
||||||
|
4: 'last',
|
||||||
|
5: 'size',
|
||||||
|
8: 'volume',
|
||||||
|
|
||||||
|
# ``ib_insync`` already packs these into
|
||||||
|
# quotes under the following fields.
|
||||||
|
# 55: 'trades_per_min', # `'tradeRate'`
|
||||||
|
# 56: 'vlm_per_min', # `'volumeRate'`
|
||||||
|
# 89: 'shortable', # `'shortableShares'`
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
@acm
|
||||||
|
async def open_data_client() -> MethodProxy:
|
||||||
|
'''
|
||||||
|
Open the first found preferred "data client" as defined in the
|
||||||
|
user's ``brokers.toml`` in the ``ib.prefer_data_account`` variable
|
||||||
|
and deliver that client wrapped in a ``MethodProxy``.
|
||||||
|
|
||||||
|
'''
|
||||||
|
async with (
|
||||||
|
open_client_proxies() as (proxies, clients),
|
||||||
|
):
|
||||||
|
account_name, client = get_preferred_data_client(clients)
|
||||||
|
proxy = proxies.get(f'ib.{account_name}')
|
||||||
|
if not proxy:
|
||||||
|
raise ValueError(
|
||||||
|
f'No preferred data client could be found for {account_name}!'
|
||||||
|
)
|
||||||
|
|
||||||
|
yield proxy
|
||||||
|
|
||||||
|
|
||||||
|
@acm
|
||||||
|
async def open_history_client(
|
||||||
|
symbol: str,
|
||||||
|
|
||||||
|
) -> tuple[Callable, int]:
|
||||||
|
'''
|
||||||
|
History retreival endpoint - delivers a historical frame callble
|
||||||
|
that takes in ``pendulum.datetime`` and returns ``numpy`` arrays.
|
||||||
|
|
||||||
|
'''
|
||||||
|
async with open_data_client() as proxy:
|
||||||
|
|
||||||
|
async def get_hist(
|
||||||
|
end_dt: Optional[datetime] = None,
|
||||||
|
start_dt: Optional[datetime] = None,
|
||||||
|
|
||||||
|
) -> tuple[np.ndarray, str]:
|
||||||
|
|
||||||
|
out, fails = await get_bars(proxy, symbol, end_dt=end_dt)
|
||||||
|
|
||||||
|
# TODO: add logic here to handle tradable hours and only grab
|
||||||
|
# valid bars in the range
|
||||||
|
if out is None:
|
||||||
|
# could be trying to retreive bars over weekend
|
||||||
|
log.error(f"Can't grab bars starting at {end_dt}!?!?")
|
||||||
|
raise NoData(
|
||||||
|
f'{end_dt}',
|
||||||
|
frame_size=2000,
|
||||||
|
)
|
||||||
|
|
||||||
|
bars, bars_array, first_dt, last_dt = out
|
||||||
|
|
||||||
|
# volume cleaning since there's -ve entries,
|
||||||
|
# wood luv to know what crookery that is..
|
||||||
|
vlm = bars_array['volume']
|
||||||
|
vlm[vlm < 0] = 0
|
||||||
|
|
||||||
|
return bars_array, first_dt, last_dt
|
||||||
|
|
||||||
|
# TODO: it seems like we can do async queries for ohlc
|
||||||
|
# but getting the order right still isn't working and I'm not
|
||||||
|
# quite sure why.. needs some tinkering and probably
|
||||||
|
# a lookthrough of the ``ib_insync`` machinery, for eg. maybe
|
||||||
|
# we have to do the batch queries on the `asyncio` side?
|
||||||
|
yield get_hist, {'erlangs': 1, 'rate': 6}
|
||||||
|
|
||||||
|
|
||||||
|
_pacing: str = (
|
||||||
|
'Historical Market Data Service error '
|
||||||
|
'message:Historical data request pacing violation'
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
async def get_bars(
|
||||||
|
|
||||||
|
proxy: MethodProxy,
|
||||||
|
fqsn: str,
|
||||||
|
|
||||||
|
# blank to start which tells ib to look up the latest datum
|
||||||
|
end_dt: str = '',
|
||||||
|
|
||||||
|
) -> (dict, np.ndarray):
|
||||||
|
'''
|
||||||
|
Retrieve historical data from a ``trio``-side task using
|
||||||
|
a ``MethoProxy``.
|
||||||
|
|
||||||
|
'''
|
||||||
|
fails = 0
|
||||||
|
bars: Optional[list] = None
|
||||||
|
first_dt: datetime = None
|
||||||
|
last_dt: datetime = None
|
||||||
|
|
||||||
|
if end_dt:
|
||||||
|
last_dt = pendulum.from_timestamp(end_dt.timestamp())
|
||||||
|
|
||||||
|
for _ in range(10):
|
||||||
|
try:
|
||||||
|
out = await proxy.bars(
|
||||||
|
fqsn=fqsn,
|
||||||
|
end_dt=end_dt,
|
||||||
|
)
|
||||||
|
if out:
|
||||||
|
bars, bars_array = out
|
||||||
|
|
||||||
|
else:
|
||||||
|
await tractor.breakpoint()
|
||||||
|
|
||||||
|
if bars_array is None:
|
||||||
|
raise SymbolNotFound(fqsn)
|
||||||
|
|
||||||
|
first_dt = pendulum.from_timestamp(
|
||||||
|
bars[0].date.timestamp())
|
||||||
|
|
||||||
|
last_dt = pendulum.from_timestamp(
|
||||||
|
bars[-1].date.timestamp())
|
||||||
|
|
||||||
|
time = bars_array['time']
|
||||||
|
assert time[-1] == last_dt.timestamp()
|
||||||
|
assert time[0] == first_dt.timestamp()
|
||||||
|
log.info(
|
||||||
|
f'{len(bars)} bars retreived for {first_dt} -> {last_dt}'
|
||||||
|
)
|
||||||
|
|
||||||
|
return (bars, bars_array, first_dt, last_dt), fails
|
||||||
|
|
||||||
|
except RequestError as err:
|
||||||
|
msg = err.message
|
||||||
|
# why do we always need to rebind this?
|
||||||
|
# _err = err
|
||||||
|
|
||||||
|
if 'No market data permissions for' in msg:
|
||||||
|
# TODO: signalling for no permissions searches
|
||||||
|
raise NoData(
|
||||||
|
f'Symbol: {fqsn}',
|
||||||
|
)
|
||||||
|
|
||||||
|
elif (
|
||||||
|
err.code == 162
|
||||||
|
and 'HMDS query returned no data' in err.message
|
||||||
|
):
|
||||||
|
# XXX: this is now done in the storage mgmt layer
|
||||||
|
# and we shouldn't implicitly decrement the frame dt
|
||||||
|
# index since the upper layer may be doing so
|
||||||
|
# concurrently and we don't want to be delivering frames
|
||||||
|
# that weren't asked for.
|
||||||
|
log.warning(
|
||||||
|
f'NO DATA found ending @ {end_dt}\n'
|
||||||
|
)
|
||||||
|
|
||||||
|
# try to decrement start point and look further back
|
||||||
|
# end_dt = last_dt = last_dt.subtract(seconds=2000)
|
||||||
|
|
||||||
|
raise NoData(
|
||||||
|
f'Symbol: {fqsn}',
|
||||||
|
frame_size=2000,
|
||||||
|
)
|
||||||
|
|
||||||
|
elif _pacing in msg:
|
||||||
|
|
||||||
|
log.warning(
|
||||||
|
'History throttle rate reached!\n'
|
||||||
|
'Resetting farms with `ctrl-alt-f` hack\n'
|
||||||
|
)
|
||||||
|
# TODO: we might have to put a task lock around this
|
||||||
|
# method..
|
||||||
|
hist_ev = proxy.status_event(
|
||||||
|
'HMDS data farm connection is OK:ushmds'
|
||||||
|
)
|
||||||
|
|
||||||
|
# XXX: other event messages we might want to try and
|
||||||
|
# wait for but i wasn't able to get any of this
|
||||||
|
# reliable..
|
||||||
|
# reconnect_start = proxy.status_event(
|
||||||
|
# 'Market data farm is connecting:usfuture'
|
||||||
|
# )
|
||||||
|
# live_ev = proxy.status_event(
|
||||||
|
# 'Market data farm connection is OK:usfuture'
|
||||||
|
# )
|
||||||
|
|
||||||
|
# try to wait on the reset event(s) to arrive, a timeout
|
||||||
|
# will trigger a retry up to 6 times (for now).
|
||||||
|
tries: int = 2
|
||||||
|
timeout: float = 10
|
||||||
|
|
||||||
|
# try 3 time with a data reset then fail over to
|
||||||
|
# a connection reset.
|
||||||
|
for i in range(1, tries):
|
||||||
|
|
||||||
|
log.warning('Sending DATA RESET request')
|
||||||
|
await data_reset_hack(reset_type='data')
|
||||||
|
|
||||||
|
with trio.move_on_after(timeout) as cs:
|
||||||
|
for name, ev in [
|
||||||
|
# TODO: not sure if waiting on other events
|
||||||
|
# is all that useful here or not. in theory
|
||||||
|
# you could wait on one of the ones above
|
||||||
|
# first to verify the reset request was
|
||||||
|
# sent?
|
||||||
|
('history', hist_ev),
|
||||||
|
]:
|
||||||
|
await ev.wait()
|
||||||
|
log.info(f"{name} DATA RESET")
|
||||||
|
break
|
||||||
|
|
||||||
|
if cs.cancelled_caught:
|
||||||
|
fails += 1
|
||||||
|
log.warning(
|
||||||
|
f'Data reset {name} timeout, retrying {i}.'
|
||||||
|
)
|
||||||
|
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
|
||||||
|
log.warning('Sending CONNECTION RESET')
|
||||||
|
await data_reset_hack(reset_type='connection')
|
||||||
|
|
||||||
|
with trio.move_on_after(timeout) as cs:
|
||||||
|
for name, ev in [
|
||||||
|
# TODO: not sure if waiting on other events
|
||||||
|
# is all that useful here or not. in theory
|
||||||
|
# you could wait on one of the ones above
|
||||||
|
# first to verify the reset request was
|
||||||
|
# sent?
|
||||||
|
('history', hist_ev),
|
||||||
|
]:
|
||||||
|
await ev.wait()
|
||||||
|
log.info(f"{name} DATA RESET")
|
||||||
|
|
||||||
|
if cs.cancelled_caught:
|
||||||
|
fails += 1
|
||||||
|
log.warning('Data CONNECTION RESET timeout!?')
|
||||||
|
|
||||||
|
else:
|
||||||
|
raise
|
||||||
|
|
||||||
|
return None, None
|
||||||
|
# else: # throttle wasn't fixed so error out immediately
|
||||||
|
# raise _err
|
||||||
|
|
||||||
|
|
||||||
|
async def backfill_bars(
|
||||||
|
|
||||||
|
fqsn: str,
|
||||||
|
shm: ShmArray, # type: ignore # noqa
|
||||||
|
|
||||||
|
# TODO: we want to avoid overrunning the underlying shm array buffer
|
||||||
|
# and we should probably calc the number of calls to make depending
|
||||||
|
# on that until we have the `marketstore` daemon in place in which
|
||||||
|
# case the shm size will be driven by user config and available sys
|
||||||
|
# memory.
|
||||||
|
count: int = 16,
|
||||||
|
|
||||||
|
task_status: TaskStatus[trio.CancelScope] = trio.TASK_STATUS_IGNORED,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
'''
|
||||||
|
Fill historical bars into shared mem / storage afap.
|
||||||
|
|
||||||
|
TODO: avoid pacing constraints:
|
||||||
|
https://github.com/pikers/piker/issues/128
|
||||||
|
|
||||||
|
'''
|
||||||
|
# last_dt1 = None
|
||||||
|
last_dt = None
|
||||||
|
|
||||||
|
with trio.CancelScope() as cs:
|
||||||
|
|
||||||
|
async with open_data_client() as proxy:
|
||||||
|
|
||||||
|
out, fails = await get_bars(proxy, fqsn)
|
||||||
|
|
||||||
|
if out is None:
|
||||||
|
raise RuntimeError("Could not pull currrent history?!")
|
||||||
|
|
||||||
|
(first_bars, bars_array, first_dt, last_dt) = out
|
||||||
|
vlm = bars_array['volume']
|
||||||
|
vlm[vlm < 0] = 0
|
||||||
|
last_dt = first_dt
|
||||||
|
|
||||||
|
# write historical data to buffer
|
||||||
|
shm.push(bars_array)
|
||||||
|
|
||||||
|
task_status.started(cs)
|
||||||
|
|
||||||
|
i = 0
|
||||||
|
while i < count:
|
||||||
|
|
||||||
|
out, fails = await get_bars(proxy, fqsn, end_dt=first_dt)
|
||||||
|
|
||||||
|
if out is None:
|
||||||
|
# could be trying to retreive bars over weekend
|
||||||
|
# TODO: add logic here to handle tradable hours and
|
||||||
|
# only grab valid bars in the range
|
||||||
|
log.error(f"Can't grab bars starting at {first_dt}!?!?")
|
||||||
|
|
||||||
|
# XXX: get_bars() should internally decrement dt by
|
||||||
|
# 2k seconds and try again.
|
||||||
|
continue
|
||||||
|
|
||||||
|
(first_bars, bars_array, first_dt, last_dt) = out
|
||||||
|
# last_dt1 = last_dt
|
||||||
|
# last_dt = first_dt
|
||||||
|
|
||||||
|
# volume cleaning since there's -ve entries,
|
||||||
|
# wood luv to know what crookery that is..
|
||||||
|
vlm = bars_array['volume']
|
||||||
|
vlm[vlm < 0] = 0
|
||||||
|
|
||||||
|
# TODO we should probably dig into forums to see what peeps
|
||||||
|
# think this data "means" and then use it as an indicator of
|
||||||
|
# sorts? dinkus has mentioned that $vlms for the day dont'
|
||||||
|
# match other platforms nor the summary stat tws shows in
|
||||||
|
# the monitor - it's probably worth investigating.
|
||||||
|
|
||||||
|
shm.push(bars_array, prepend=True)
|
||||||
|
i += 1
|
||||||
|
|
||||||
|
|
||||||
|
asset_type_map = {
|
||||||
|
'STK': 'stock',
|
||||||
|
'OPT': 'option',
|
||||||
|
'FUT': 'future',
|
||||||
|
'CONTFUT': 'continuous_future',
|
||||||
|
'CASH': 'forex',
|
||||||
|
'IND': 'index',
|
||||||
|
'CFD': 'cfd',
|
||||||
|
'BOND': 'bond',
|
||||||
|
'CMDTY': 'commodity',
|
||||||
|
'FOP': 'futures_option',
|
||||||
|
'FUND': 'mutual_fund',
|
||||||
|
'WAR': 'warrant',
|
||||||
|
'IOPT': 'warran',
|
||||||
|
'BAG': 'bag',
|
||||||
|
# 'NEWS': 'news',
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
_quote_streams: dict[str, trio.abc.ReceiveStream] = {}
|
||||||
|
|
||||||
|
|
||||||
|
async def _setup_quote_stream(
|
||||||
|
|
||||||
|
from_trio: asyncio.Queue,
|
||||||
|
to_trio: trio.abc.SendChannel,
|
||||||
|
|
||||||
|
symbol: str,
|
||||||
|
opts: tuple[int] = (
|
||||||
|
'375', # RT trade volume (excludes utrades)
|
||||||
|
'233', # RT trade volume (includes utrades)
|
||||||
|
'236', # Shortable shares
|
||||||
|
|
||||||
|
# these all appear to only be updated every 25s thus
|
||||||
|
# making them mostly useless and explains why the scanner
|
||||||
|
# is always slow XD
|
||||||
|
# '293', # Trade count for day
|
||||||
|
'294', # Trade rate / minute
|
||||||
|
'295', # Vlm rate / minute
|
||||||
|
),
|
||||||
|
contract: Optional[Contract] = None,
|
||||||
|
|
||||||
|
) -> trio.abc.ReceiveChannel:
|
||||||
|
'''
|
||||||
|
Stream a ticker using the std L1 api.
|
||||||
|
|
||||||
|
This task is ``asyncio``-side and must be called from
|
||||||
|
``tractor.to_asyncio.open_channel_from()``.
|
||||||
|
|
||||||
|
'''
|
||||||
|
global _quote_streams
|
||||||
|
|
||||||
|
to_trio.send_nowait(None)
|
||||||
|
|
||||||
|
async with load_aio_clients() as accts2clients:
|
||||||
|
caccount_name, client = get_preferred_data_client(accts2clients)
|
||||||
|
contract = contract or (await client.find_contract(symbol))
|
||||||
|
ticker: Ticker = client.ib.reqMktData(contract, ','.join(opts))
|
||||||
|
|
||||||
|
# NOTE: it's batch-wise and slow af but I guess could
|
||||||
|
# be good for backchecking? Seems to be every 5s maybe?
|
||||||
|
# ticker: Ticker = client.ib.reqTickByTickData(
|
||||||
|
# contract, 'Last',
|
||||||
|
# )
|
||||||
|
|
||||||
|
# # define a simple queue push routine that streams quote packets
|
||||||
|
# # to trio over the ``to_trio`` memory channel.
|
||||||
|
# to_trio, from_aio = trio.open_memory_channel(2**8) # type: ignore
|
||||||
|
def teardown():
|
||||||
|
ticker.updateEvent.disconnect(push)
|
||||||
|
log.error(f"Disconnected stream for `{symbol}`")
|
||||||
|
client.ib.cancelMktData(contract)
|
||||||
|
|
||||||
|
# decouple broadcast mem chan
|
||||||
|
_quote_streams.pop(symbol, None)
|
||||||
|
|
||||||
|
def push(t: Ticker) -> None:
|
||||||
|
"""
|
||||||
|
Push quotes to trio task.
|
||||||
|
|
||||||
|
"""
|
||||||
|
# log.debug(t)
|
||||||
|
try:
|
||||||
|
to_trio.send_nowait(t)
|
||||||
|
|
||||||
|
except (
|
||||||
|
trio.BrokenResourceError,
|
||||||
|
|
||||||
|
# XXX: HACK, not sure why this gets left stale (probably
|
||||||
|
# due to our terrible ``tractor.to_asyncio``
|
||||||
|
# implementation for streams.. but if the mem chan
|
||||||
|
# gets left here and starts blocking just kill the feed?
|
||||||
|
# trio.WouldBlock,
|
||||||
|
):
|
||||||
|
# XXX: eventkit's ``Event.emit()`` for whatever redic
|
||||||
|
# reason will catch and ignore regular exceptions
|
||||||
|
# resulting in tracebacks spammed to console..
|
||||||
|
# Manually do the dereg ourselves.
|
||||||
|
teardown()
|
||||||
|
except trio.WouldBlock:
|
||||||
|
log.warning(
|
||||||
|
f'channel is blocking symbol feed for {symbol}?'
|
||||||
|
f'\n{to_trio.statistics}'
|
||||||
|
)
|
||||||
|
|
||||||
|
# except trio.WouldBlock:
|
||||||
|
# # for slow debugging purposes to avoid clobbering prompt
|
||||||
|
# # with log msgs
|
||||||
|
# pass
|
||||||
|
|
||||||
|
ticker.updateEvent.connect(push)
|
||||||
|
try:
|
||||||
|
await asyncio.sleep(float('inf'))
|
||||||
|
finally:
|
||||||
|
teardown()
|
||||||
|
|
||||||
|
# return from_aio
|
||||||
|
|
||||||
|
|
||||||
|
@acm
|
||||||
|
async def open_aio_quote_stream(
|
||||||
|
|
||||||
|
symbol: str,
|
||||||
|
contract: Optional[Contract] = None,
|
||||||
|
|
||||||
|
) -> trio.abc.ReceiveStream:
|
||||||
|
|
||||||
|
from tractor.trionics import broadcast_receiver
|
||||||
|
global _quote_streams
|
||||||
|
|
||||||
|
from_aio = _quote_streams.get(symbol)
|
||||||
|
if from_aio:
|
||||||
|
|
||||||
|
# if we already have a cached feed deliver a rx side clone to consumer
|
||||||
|
async with broadcast_receiver(
|
||||||
|
from_aio,
|
||||||
|
2**6,
|
||||||
|
) as from_aio:
|
||||||
|
yield from_aio
|
||||||
|
return
|
||||||
|
|
||||||
|
async with tractor.to_asyncio.open_channel_from(
|
||||||
|
_setup_quote_stream,
|
||||||
|
symbol=symbol,
|
||||||
|
contract=contract,
|
||||||
|
|
||||||
|
) as (first, from_aio):
|
||||||
|
|
||||||
|
# cache feed for later consumers
|
||||||
|
_quote_streams[symbol] = from_aio
|
||||||
|
|
||||||
|
yield from_aio
|
||||||
|
|
||||||
|
|
||||||
|
# TODO: cython/mypyc/numba this!
|
||||||
|
def normalize(
|
||||||
|
ticker: Ticker,
|
||||||
|
calc_price: bool = False
|
||||||
|
|
||||||
|
) -> dict:
|
||||||
|
|
||||||
|
# should be real volume for this contract by default
|
||||||
|
calc_price = False
|
||||||
|
|
||||||
|
# check for special contract types
|
||||||
|
con = ticker.contract
|
||||||
|
if type(con) in (
|
||||||
|
ibis.Commodity,
|
||||||
|
ibis.Forex,
|
||||||
|
):
|
||||||
|
# commodities and forex don't have an exchange name and
|
||||||
|
# no real volume so we have to calculate the price
|
||||||
|
suffix = con.secType
|
||||||
|
# no real volume on this tract
|
||||||
|
calc_price = True
|
||||||
|
|
||||||
|
else:
|
||||||
|
suffix = con.primaryExchange
|
||||||
|
if not suffix:
|
||||||
|
suffix = con.exchange
|
||||||
|
|
||||||
|
# append a `.<suffix>` to the returned symbol
|
||||||
|
# key for derivatives that normally is the expiry
|
||||||
|
# date key.
|
||||||
|
expiry = con.lastTradeDateOrContractMonth
|
||||||
|
if expiry:
|
||||||
|
suffix += f'.{expiry}'
|
||||||
|
|
||||||
|
# convert named tuples to dicts so we send usable keys
|
||||||
|
new_ticks = []
|
||||||
|
for tick in ticker.ticks:
|
||||||
|
if tick and not isinstance(tick, dict):
|
||||||
|
td = tick._asdict()
|
||||||
|
td['type'] = tick_types.get(
|
||||||
|
td['tickType'],
|
||||||
|
'n/a',
|
||||||
|
)
|
||||||
|
|
||||||
|
new_ticks.append(td)
|
||||||
|
|
||||||
|
tbt = ticker.tickByTicks
|
||||||
|
if tbt:
|
||||||
|
print(f'tickbyticks:\n {ticker.tickByTicks}')
|
||||||
|
|
||||||
|
ticker.ticks = new_ticks
|
||||||
|
|
||||||
|
# some contracts don't have volume so we may want to calculate
|
||||||
|
# a midpoint price based on data we can acquire (such as bid / ask)
|
||||||
|
if calc_price:
|
||||||
|
ticker.ticks.append(
|
||||||
|
{'type': 'trade', 'price': ticker.marketPrice()}
|
||||||
|
)
|
||||||
|
|
||||||
|
# serialize for transport
|
||||||
|
data = asdict(ticker)
|
||||||
|
|
||||||
|
# generate fqsn with possible specialized suffix
|
||||||
|
# for derivatives, note the lowercase.
|
||||||
|
data['symbol'] = data['fqsn'] = '.'.join(
|
||||||
|
(con.symbol, suffix)
|
||||||
|
).lower()
|
||||||
|
|
||||||
|
# convert named tuples to dicts for transport
|
||||||
|
tbts = data.get('tickByTicks')
|
||||||
|
if tbts:
|
||||||
|
data['tickByTicks'] = [tbt._asdict() for tbt in tbts]
|
||||||
|
|
||||||
|
# add time stamps for downstream latency measurements
|
||||||
|
data['brokerd_ts'] = time.time()
|
||||||
|
|
||||||
|
# stupid stupid shit...don't even care any more..
|
||||||
|
# leave it until we do a proper latency study
|
||||||
|
# if ticker.rtTime is not None:
|
||||||
|
# data['broker_ts'] = data['rtTime_s'] = float(
|
||||||
|
# ticker.rtTime.timestamp) / 1000.
|
||||||
|
data.pop('rtTime')
|
||||||
|
|
||||||
|
return data
|
||||||
|
|
||||||
|
|
||||||
|
async def stream_quotes(
|
||||||
|
|
||||||
|
send_chan: trio.abc.SendChannel,
|
||||||
|
symbols: list[str],
|
||||||
|
feed_is_live: trio.Event,
|
||||||
|
loglevel: str = None,
|
||||||
|
|
||||||
|
# startup sync
|
||||||
|
task_status: TaskStatus[tuple[dict, dict]] = trio.TASK_STATUS_IGNORED,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
'''
|
||||||
|
Stream symbol quotes.
|
||||||
|
|
||||||
|
This is a ``trio`` callable routine meant to be invoked
|
||||||
|
once the brokerd is up.
|
||||||
|
|
||||||
|
'''
|
||||||
|
# TODO: support multiple subscriptions
|
||||||
|
sym = symbols[0]
|
||||||
|
log.info(f'request for real-time quotes: {sym}')
|
||||||
|
|
||||||
|
async with open_data_client() as proxy:
|
||||||
|
|
||||||
|
con, first_ticker, details = await proxy.get_sym_details(symbol=sym)
|
||||||
|
first_quote = normalize(first_ticker)
|
||||||
|
# print(f'first quote: {first_quote}')
|
||||||
|
|
||||||
|
def mk_init_msgs() -> dict[str, dict]:
|
||||||
|
'''
|
||||||
|
Collect a bunch of meta-data useful for feed startup and
|
||||||
|
pack in a `dict`-msg.
|
||||||
|
|
||||||
|
'''
|
||||||
|
# pass back some symbol info like min_tick, trading_hours, etc.
|
||||||
|
syminfo = asdict(details)
|
||||||
|
syminfo.update(syminfo['contract'])
|
||||||
|
|
||||||
|
# nested dataclass we probably don't need and that won't IPC
|
||||||
|
# serialize
|
||||||
|
syminfo.pop('secIdList')
|
||||||
|
|
||||||
|
# TODO: more consistent field translation
|
||||||
|
atype = syminfo['asset_type'] = asset_type_map[syminfo['secType']]
|
||||||
|
|
||||||
|
# for stocks it seems TWS reports too small a tick size
|
||||||
|
# such that you can't submit orders with that granularity?
|
||||||
|
min_tick = 0.01 if atype == 'stock' else 0
|
||||||
|
|
||||||
|
syminfo['price_tick_size'] = max(syminfo['minTick'], min_tick)
|
||||||
|
|
||||||
|
# for "traditional" assets, volume is normally discreet, not
|
||||||
|
# a float
|
||||||
|
syminfo['lot_tick_size'] = 0.0
|
||||||
|
|
||||||
|
ibclient = proxy._aio_ns.ib.client
|
||||||
|
host, port = ibclient.host, ibclient.port
|
||||||
|
|
||||||
|
# TODO: for loop through all symbols passed in
|
||||||
|
init_msgs = {
|
||||||
|
# pass back token, and bool, signalling if we're the writer
|
||||||
|
# and that history has been written
|
||||||
|
sym: {
|
||||||
|
'symbol_info': syminfo,
|
||||||
|
'fqsn': first_quote['fqsn'],
|
||||||
|
},
|
||||||
|
'status': {
|
||||||
|
'data_ep': f'{host}:{port}',
|
||||||
|
},
|
||||||
|
|
||||||
|
}
|
||||||
|
return init_msgs
|
||||||
|
|
||||||
|
init_msgs = mk_init_msgs()
|
||||||
|
|
||||||
|
# TODO: we should instead spawn a task that waits on a feed to start
|
||||||
|
# and let it wait indefinitely..instead of this hard coded stuff.
|
||||||
|
with trio.move_on_after(1):
|
||||||
|
contract, first_ticker, details = await proxy.get_quote(symbol=sym)
|
||||||
|
|
||||||
|
# it might be outside regular trading hours so see if we can at
|
||||||
|
# least grab history.
|
||||||
|
if isnan(first_ticker.last):
|
||||||
|
task_status.started((init_msgs, first_quote))
|
||||||
|
|
||||||
|
# it's not really live but this will unblock
|
||||||
|
# the brokerd feed task to tell the ui to update?
|
||||||
|
feed_is_live.set()
|
||||||
|
|
||||||
|
# block and let data history backfill code run.
|
||||||
|
await trio.sleep_forever()
|
||||||
|
return # we never expect feed to come up?
|
||||||
|
|
||||||
|
async with open_aio_quote_stream(
|
||||||
|
symbol=sym,
|
||||||
|
contract=con,
|
||||||
|
) as stream:
|
||||||
|
|
||||||
|
# ugh, clear ticks since we've consumed them
|
||||||
|
# (ahem, ib_insync is stateful trash)
|
||||||
|
first_ticker.ticks = []
|
||||||
|
|
||||||
|
task_status.started((init_msgs, first_quote))
|
||||||
|
|
||||||
|
async with aclosing(stream):
|
||||||
|
if type(first_ticker.contract) not in (
|
||||||
|
ibis.Commodity,
|
||||||
|
ibis.Forex
|
||||||
|
):
|
||||||
|
# wait for real volume on feed (trading might be closed)
|
||||||
|
while True:
|
||||||
|
ticker = await stream.receive()
|
||||||
|
|
||||||
|
# for a real volume contract we rait for the first
|
||||||
|
# "real" trade to take place
|
||||||
|
if (
|
||||||
|
# not calc_price
|
||||||
|
# and not ticker.rtTime
|
||||||
|
not ticker.rtTime
|
||||||
|
):
|
||||||
|
# spin consuming tickers until we get a real
|
||||||
|
# market datum
|
||||||
|
log.debug(f"New unsent ticker: {ticker}")
|
||||||
|
continue
|
||||||
|
else:
|
||||||
|
log.debug("Received first real volume tick")
|
||||||
|
# ugh, clear ticks since we've consumed them
|
||||||
|
# (ahem, ib_insync is truly stateful trash)
|
||||||
|
ticker.ticks = []
|
||||||
|
|
||||||
|
# XXX: this works because we don't use
|
||||||
|
# ``aclosing()`` above?
|
||||||
|
break
|
||||||
|
|
||||||
|
quote = normalize(ticker)
|
||||||
|
log.debug(f"First ticker received {quote}")
|
||||||
|
|
||||||
|
# tell caller quotes are now coming in live
|
||||||
|
feed_is_live.set()
|
||||||
|
|
||||||
|
# last = time.time()
|
||||||
|
async for ticker in stream:
|
||||||
|
quote = normalize(ticker)
|
||||||
|
await send_chan.send({quote['fqsn']: quote})
|
||||||
|
|
||||||
|
# ugh, clear ticks since we've consumed them
|
||||||
|
ticker.ticks = []
|
||||||
|
# last = time.time()
|
||||||
|
|
||||||
|
|
||||||
|
async def data_reset_hack(
|
||||||
|
reset_type: str = 'data',
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
'''
|
||||||
|
Run key combos for resetting data feeds and yield back to caller
|
||||||
|
when complete.
|
||||||
|
|
||||||
|
This is a linux-only hack around:
|
||||||
|
|
||||||
|
https://interactivebrokers.github.io/tws-api/historical_limitations.html#pacing_violations
|
||||||
|
|
||||||
|
TODOs:
|
||||||
|
- a return type that hopefully determines if the hack was
|
||||||
|
successful.
|
||||||
|
- other OS support?
|
||||||
|
- integration with ``ib-gw`` run in docker + Xorg?
|
||||||
|
|
||||||
|
'''
|
||||||
|
|
||||||
|
async def vnc_click_hack(
|
||||||
|
reset_type: str = 'data'
|
||||||
|
) -> None:
|
||||||
|
'''
|
||||||
|
Reset the data or netowork connection for the VNC attached
|
||||||
|
ib gateway using magic combos.
|
||||||
|
|
||||||
|
'''
|
||||||
|
key = {'data': 'f', 'connection': 'r'}[reset_type]
|
||||||
|
|
||||||
|
import asyncvnc
|
||||||
|
|
||||||
|
async with asyncvnc.connect(
|
||||||
|
'localhost',
|
||||||
|
port=3003,
|
||||||
|
# password='ibcansmbz',
|
||||||
|
) as client:
|
||||||
|
|
||||||
|
# move to middle of screen
|
||||||
|
# 640x1800
|
||||||
|
client.mouse.move(
|
||||||
|
x=500,
|
||||||
|
y=500,
|
||||||
|
)
|
||||||
|
client.mouse.click()
|
||||||
|
client.keyboard.press('Ctrl', 'Alt', key) # keys are stacked
|
||||||
|
|
||||||
|
await tractor.to_asyncio.run_task(vnc_click_hack)
|
||||||
|
|
||||||
|
# we don't really need the ``xdotool`` approach any more B)
|
||||||
|
return True
|
||||||
|
|
||||||
|
|
||||||
|
@tractor.context
|
||||||
|
async def open_symbol_search(
|
||||||
|
ctx: tractor.Context,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
|
||||||
|
# TODO: load user defined symbol set locally for fast search?
|
||||||
|
await ctx.started({})
|
||||||
|
|
||||||
|
async with open_data_client() as proxy:
|
||||||
|
async with ctx.open_stream() as stream:
|
||||||
|
|
||||||
|
last = time.time()
|
||||||
|
|
||||||
|
async for pattern in stream:
|
||||||
|
log.debug(f'received {pattern}')
|
||||||
|
now = time.time()
|
||||||
|
|
||||||
|
assert pattern, 'IB can not accept blank search pattern'
|
||||||
|
|
||||||
|
# throttle search requests to no faster then 1Hz
|
||||||
|
diff = now - last
|
||||||
|
if diff < 1.0:
|
||||||
|
log.debug('throttle sleeping')
|
||||||
|
await trio.sleep(diff)
|
||||||
|
try:
|
||||||
|
pattern = stream.receive_nowait()
|
||||||
|
except trio.WouldBlock:
|
||||||
|
pass
|
||||||
|
|
||||||
|
if not pattern or pattern.isspace():
|
||||||
|
log.warning('empty pattern received, skipping..')
|
||||||
|
|
||||||
|
# TODO: *BUG* if nothing is returned here the client
|
||||||
|
# side will cache a null set result and not showing
|
||||||
|
# anything to the use on re-searches when this query
|
||||||
|
# timed out. We probably need a special "timeout" msg
|
||||||
|
# or something...
|
||||||
|
|
||||||
|
# XXX: this unblocks the far end search task which may
|
||||||
|
# hold up a multi-search nursery block
|
||||||
|
await stream.send({})
|
||||||
|
|
||||||
|
continue
|
||||||
|
|
||||||
|
log.debug(f'searching for {pattern}')
|
||||||
|
|
||||||
|
last = time.time()
|
||||||
|
|
||||||
|
# async batch search using api stocks endpoint and module
|
||||||
|
# defined adhoc symbol set.
|
||||||
|
stock_results = []
|
||||||
|
|
||||||
|
async def stash_results(target: Awaitable[list]):
|
||||||
|
stock_results.extend(await target)
|
||||||
|
|
||||||
|
async with trio.open_nursery() as sn:
|
||||||
|
sn.start_soon(
|
||||||
|
stash_results,
|
||||||
|
proxy.search_symbols(
|
||||||
|
pattern=pattern,
|
||||||
|
upto=5,
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
# trigger async request
|
||||||
|
await trio.sleep(0)
|
||||||
|
|
||||||
|
# match against our ad-hoc set immediately
|
||||||
|
adhoc_matches = fuzzy.extractBests(
|
||||||
|
pattern,
|
||||||
|
list(_adhoc_futes_set),
|
||||||
|
score_cutoff=90,
|
||||||
|
)
|
||||||
|
log.info(f'fuzzy matched adhocs: {adhoc_matches}')
|
||||||
|
adhoc_match_results = {}
|
||||||
|
if adhoc_matches:
|
||||||
|
# TODO: do we need to pull contract details?
|
||||||
|
adhoc_match_results = {i[0]: {} for i in adhoc_matches}
|
||||||
|
|
||||||
|
log.debug(f'fuzzy matching stocks {stock_results}')
|
||||||
|
stock_matches = fuzzy.extractBests(
|
||||||
|
pattern,
|
||||||
|
stock_results,
|
||||||
|
score_cutoff=50,
|
||||||
|
)
|
||||||
|
|
||||||
|
matches = adhoc_match_results | {
|
||||||
|
item[0]: {} for item in stock_matches
|
||||||
|
}
|
||||||
|
# TODO: we used to deliver contract details
|
||||||
|
# {item[2]: item[0] for item in stock_matches}
|
||||||
|
|
||||||
|
log.debug(f"sending matches: {matches.keys()}")
|
||||||
|
await stream.send(matches)
|
|
@ -80,7 +80,9 @@ def mk_check(
|
||||||
|
|
||||||
return check_lt
|
return check_lt
|
||||||
|
|
||||||
raise ValueError('trigger: {trigger_price}, last: {known_last}')
|
raise ValueError(
|
||||||
|
f'trigger: {trigger_price}, last: {known_last}'
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
@dataclass
|
@dataclass
|
||||||
|
|
|
@ -1,5 +1,5 @@
|
||||||
# piker: trading gear for hackers
|
# piker: trading gear for hackers
|
||||||
# Copyright (C) 2018-present Tyler Goodlet (in stewardship of piker0)
|
# Copyright (C) 2018-present Tyler Goodlet (in stewardship for pikers)
|
||||||
|
|
||||||
# This program is free software: you can redistribute it and/or modify
|
# This program is free software: you can redistribute it and/or modify
|
||||||
# it under the terms of the GNU Affero General Public License as published by
|
# it under the terms of the GNU Affero General Public License as published by
|
||||||
|
@ -16,6 +16,7 @@
|
||||||
|
|
||||||
"""
|
"""
|
||||||
Broker configuration mgmt.
|
Broker configuration mgmt.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
import platform
|
import platform
|
||||||
import sys
|
import sys
|
||||||
|
@ -50,7 +51,7 @@ def get_app_dir(app_name, roaming=True, force_posix=False):
|
||||||
Unix (POSIX):
|
Unix (POSIX):
|
||||||
``~/.foo-bar``
|
``~/.foo-bar``
|
||||||
Win XP (roaming):
|
Win XP (roaming):
|
||||||
``C:\Documents and Settings\<user>\Local Settings\Application Data\Foo Bar``
|
``C:\Documents and Settings\<user>\Local Settings\Application Data\Foo``
|
||||||
Win XP (not roaming):
|
Win XP (not roaming):
|
||||||
``C:\Documents and Settings\<user>\Application Data\Foo Bar``
|
``C:\Documents and Settings\<user>\Application Data\Foo Bar``
|
||||||
Win 7 (roaming):
|
Win 7 (roaming):
|
||||||
|
@ -81,7 +82,8 @@ def get_app_dir(app_name, roaming=True, force_posix=False):
|
||||||
folder = os.path.expanduser("~")
|
folder = os.path.expanduser("~")
|
||||||
return os.path.join(folder, app_name)
|
return os.path.join(folder, app_name)
|
||||||
if force_posix:
|
if force_posix:
|
||||||
return os.path.join(os.path.expanduser("~/.{}".format(_posixify(app_name))))
|
return os.path.join(
|
||||||
|
os.path.expanduser("~/.{}".format(_posixify(app_name))))
|
||||||
if sys.platform == "darwin":
|
if sys.platform == "darwin":
|
||||||
return os.path.join(
|
return os.path.join(
|
||||||
os.path.expanduser("~/Library/Application Support"), app_name
|
os.path.expanduser("~/Library/Application Support"), app_name
|
||||||
|
@ -107,7 +109,12 @@ if _parent_user:
|
||||||
]
|
]
|
||||||
)
|
)
|
||||||
|
|
||||||
_file_name = 'brokers.toml'
|
_conf_names: set[str] = {
|
||||||
|
'brokers',
|
||||||
|
'trades',
|
||||||
|
'watchlists',
|
||||||
|
}
|
||||||
|
|
||||||
_watchlists_data_path = os.path.join(_config_dir, 'watchlists.json')
|
_watchlists_data_path = os.path.join(_config_dir, 'watchlists.json')
|
||||||
_context_defaults = dict(
|
_context_defaults = dict(
|
||||||
default_map={
|
default_map={
|
||||||
|
@ -129,23 +136,43 @@ def _override_config_dir(
|
||||||
_config_dir = path
|
_config_dir = path
|
||||||
|
|
||||||
|
|
||||||
def get_broker_conf_path():
|
def _conf_fn_w_ext(
|
||||||
|
name: str,
|
||||||
|
) -> str:
|
||||||
|
# change this if we ever change the config file format.
|
||||||
|
return f'{name}.toml'
|
||||||
|
|
||||||
|
|
||||||
|
def get_conf_path(
|
||||||
|
conf_name: str = 'brokers',
|
||||||
|
|
||||||
|
) -> str:
|
||||||
"""Return the default config path normally under
|
"""Return the default config path normally under
|
||||||
``~/.config/piker`` on linux.
|
``~/.config/piker`` on linux.
|
||||||
|
|
||||||
Contains files such as:
|
Contains files such as:
|
||||||
- brokers.toml
|
- brokers.toml
|
||||||
- watchlists.toml
|
- watchlists.toml
|
||||||
|
- trades.toml
|
||||||
|
|
||||||
|
# maybe coming soon ;)
|
||||||
- signals.toml
|
- signals.toml
|
||||||
- strats.toml
|
- strats.toml
|
||||||
|
|
||||||
"""
|
"""
|
||||||
return os.path.join(_config_dir, _file_name)
|
assert conf_name in _conf_names
|
||||||
|
fn = _conf_fn_w_ext(conf_name)
|
||||||
|
return os.path.join(
|
||||||
|
_config_dir,
|
||||||
|
fn,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
def repodir():
|
def repodir():
|
||||||
"""Return the abspath to the repo directory.
|
'''
|
||||||
"""
|
Return the abspath to the repo directory.
|
||||||
|
|
||||||
|
'''
|
||||||
dirpath = os.path.abspath(
|
dirpath = os.path.abspath(
|
||||||
# we're 3 levels down in **this** module file
|
# we're 3 levels down in **this** module file
|
||||||
dirname(dirname(os.path.realpath(__file__)))
|
dirname(dirname(os.path.realpath(__file__)))
|
||||||
|
@ -154,16 +181,27 @@ def repodir():
|
||||||
|
|
||||||
|
|
||||||
def load(
|
def load(
|
||||||
|
conf_name: str = 'brokers',
|
||||||
path: str = None
|
path: str = None
|
||||||
|
|
||||||
) -> (dict, str):
|
) -> (dict, str):
|
||||||
"""Load broker config.
|
'''
|
||||||
"""
|
Load config file by name.
|
||||||
path = path or get_broker_conf_path()
|
|
||||||
|
'''
|
||||||
|
path = path or get_conf_path(conf_name)
|
||||||
if not os.path.isfile(path):
|
if not os.path.isfile(path):
|
||||||
shutil.copyfile(
|
fn = _conf_fn_w_ext(conf_name)
|
||||||
os.path.join(repodir(), 'config', 'brokers.toml'),
|
|
||||||
path,
|
template = os.path.join(
|
||||||
|
repodir(),
|
||||||
|
'config',
|
||||||
|
fn
|
||||||
)
|
)
|
||||||
|
# try to copy in a template config to the user's directory
|
||||||
|
# if one exists.
|
||||||
|
if os.path.isfile(template):
|
||||||
|
shutil.copyfile(template, path)
|
||||||
|
|
||||||
config = toml.load(path)
|
config = toml.load(path)
|
||||||
log.debug(f"Read config file {path}")
|
log.debug(f"Read config file {path}")
|
||||||
|
@ -172,13 +210,17 @@ def load(
|
||||||
|
|
||||||
def write(
|
def write(
|
||||||
config: dict, # toml config as dict
|
config: dict, # toml config as dict
|
||||||
|
name: str = 'brokers',
|
||||||
path: str = None,
|
path: str = None,
|
||||||
|
|
||||||
) -> None:
|
) -> None:
|
||||||
"""Write broker config to disk.
|
''''
|
||||||
|
Write broker config to disk.
|
||||||
|
|
||||||
Create a ``brokers.ini`` file if one does not exist.
|
Create a ``brokers.ini`` file if one does not exist.
|
||||||
"""
|
|
||||||
path = path or get_broker_conf_path()
|
'''
|
||||||
|
path = path or get_conf_path(name)
|
||||||
dirname = os.path.dirname(path)
|
dirname = os.path.dirname(path)
|
||||||
if not os.path.isdir(dirname):
|
if not os.path.isdir(dirname):
|
||||||
log.debug(f"Creating config dir {_config_dir}")
|
log.debug(f"Creating config dir {_config_dir}")
|
||||||
|
@ -188,7 +230,10 @@ def write(
|
||||||
raise ValueError(
|
raise ValueError(
|
||||||
"Watch out you're trying to write a blank config!")
|
"Watch out you're trying to write a blank config!")
|
||||||
|
|
||||||
log.debug(f"Writing config file {path}")
|
log.debug(
|
||||||
|
f"Writing config `{name}` file to:\n"
|
||||||
|
f"{path}"
|
||||||
|
)
|
||||||
with open(path, 'w') as cf:
|
with open(path, 'w') as cf:
|
||||||
return toml.dump(config, cf)
|
return toml.dump(config, cf)
|
||||||
|
|
||||||
|
@ -218,4 +263,5 @@ def load_accounts(
|
||||||
|
|
||||||
# our default paper engine entry
|
# our default paper engine entry
|
||||||
accounts['paper'] = None
|
accounts['paper'] = None
|
||||||
|
|
||||||
return accounts
|
return accounts
|
||||||
|
|
|
@ -19,6 +19,7 @@ Supervisor for docker with included specific-image service helpers.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
import os
|
import os
|
||||||
|
import time
|
||||||
from typing import (
|
from typing import (
|
||||||
Optional,
|
Optional,
|
||||||
Callable,
|
Callable,
|
||||||
|
@ -98,8 +99,6 @@ async def open_docker(
|
||||||
finally:
|
finally:
|
||||||
if client:
|
if client:
|
||||||
client.close()
|
client.close()
|
||||||
for c in client.containers.list():
|
|
||||||
c.kill()
|
|
||||||
|
|
||||||
|
|
||||||
class Container:
|
class Container:
|
||||||
|
@ -188,45 +187,65 @@ class Container:
|
||||||
|
|
||||||
async def cancel(
|
async def cancel(
|
||||||
self,
|
self,
|
||||||
|
stop_msg: str,
|
||||||
) -> None:
|
) -> None:
|
||||||
|
|
||||||
cid = self.cntr.id
|
cid = self.cntr.id
|
||||||
|
# first try a graceful cancel
|
||||||
|
log.cancel(
|
||||||
|
f'SIGINT cancelling container: {cid}\n'
|
||||||
|
f'waiting on stop msg: "{stop_msg}"'
|
||||||
|
)
|
||||||
self.try_signal('SIGINT')
|
self.try_signal('SIGINT')
|
||||||
|
|
||||||
with trio.move_on_after(0.5) as cs:
|
start = time.time()
|
||||||
cs.shield = True
|
for _ in range(30):
|
||||||
await self.process_logs_until('initiating graceful shutdown')
|
|
||||||
await self.process_logs_until('exiting...',)
|
|
||||||
|
|
||||||
for _ in range(10):
|
|
||||||
with trio.move_on_after(0.5) as cs:
|
with trio.move_on_after(0.5) as cs:
|
||||||
cs.shield = True
|
cs.shield = True
|
||||||
await self.process_logs_until('exiting...',)
|
await self.process_logs_until(stop_msg)
|
||||||
|
|
||||||
|
# if we aren't cancelled on above checkpoint then we
|
||||||
|
# assume we read the expected stop msg and terminated.
|
||||||
break
|
break
|
||||||
|
|
||||||
if cs.cancelled_caught:
|
|
||||||
# get out the big guns, bc apparently marketstore
|
|
||||||
# doesn't actually know how to terminate gracefully
|
|
||||||
# :eyeroll:...
|
|
||||||
self.try_signal('SIGKILL')
|
|
||||||
|
|
||||||
try:
|
try:
|
||||||
log.info('Waiting on container shutdown: {cid}')
|
log.info(f'Polling for container shutdown:\n{cid}')
|
||||||
|
|
||||||
|
if self.cntr.status not in {'exited', 'not-running'}:
|
||||||
self.cntr.wait(
|
self.cntr.wait(
|
||||||
timeout=0.1,
|
timeout=0.1,
|
||||||
condition='not-running',
|
condition='not-running',
|
||||||
)
|
)
|
||||||
|
|
||||||
break
|
break
|
||||||
|
|
||||||
except (
|
except (
|
||||||
ReadTimeout,
|
ReadTimeout,
|
||||||
|
):
|
||||||
|
log.info(f'Still waiting on container:\n{cid}')
|
||||||
|
continue
|
||||||
|
|
||||||
|
except (
|
||||||
|
docker.errors.APIError,
|
||||||
ConnectionError,
|
ConnectionError,
|
||||||
):
|
):
|
||||||
log.error(f'failed to wait on container {cid}')
|
log.exception('Docker connection failure')
|
||||||
raise
|
break
|
||||||
|
|
||||||
else:
|
else:
|
||||||
raise RuntimeError('Failed to cancel container {cid}')
|
delay = time.time() - start
|
||||||
|
log.error(
|
||||||
|
f'Failed to kill container {cid} after {delay}s\n'
|
||||||
|
'sending SIGKILL..'
|
||||||
|
)
|
||||||
|
# get out the big guns, bc apparently marketstore
|
||||||
|
# doesn't actually know how to terminate gracefully
|
||||||
|
# :eyeroll:...
|
||||||
|
self.try_signal('SIGKILL')
|
||||||
|
self.cntr.wait(
|
||||||
|
timeout=3,
|
||||||
|
condition='not-running',
|
||||||
|
)
|
||||||
|
|
||||||
log.cancel(f'Container stopped: {cid}')
|
log.cancel(f'Container stopped: {cid}')
|
||||||
|
|
||||||
|
@ -247,13 +266,16 @@ async def open_ahabd(
|
||||||
# params, etc. passing to ``Containter.run()``?
|
# params, etc. passing to ``Containter.run()``?
|
||||||
# call into endpoint for container config/init
|
# call into endpoint for container config/init
|
||||||
ep_func = NamespacePath(endpoint).load_ref()
|
ep_func = NamespacePath(endpoint).load_ref()
|
||||||
dcntr, cntr_config = ep_func(client)
|
(
|
||||||
|
dcntr,
|
||||||
|
cntr_config,
|
||||||
|
start_msg,
|
||||||
|
stop_msg,
|
||||||
|
) = ep_func(client)
|
||||||
cntr = Container(dcntr)
|
cntr = Container(dcntr)
|
||||||
|
|
||||||
with trio.move_on_after(1):
|
with trio.move_on_after(1):
|
||||||
found = await cntr.process_logs_until(
|
found = await cntr.process_logs_until(start_msg)
|
||||||
"launching tcp listener for all services...",
|
|
||||||
)
|
|
||||||
|
|
||||||
if not found and cntr not in client.containers.list():
|
if not found and cntr not in client.containers.list():
|
||||||
raise RuntimeError(
|
raise RuntimeError(
|
||||||
|
@ -273,16 +295,9 @@ async def open_ahabd(
|
||||||
# callers to have root perms?
|
# callers to have root perms?
|
||||||
await trio.sleep_forever()
|
await trio.sleep_forever()
|
||||||
|
|
||||||
except (
|
finally:
|
||||||
BaseException,
|
|
||||||
# trio.Cancelled,
|
|
||||||
# KeyboardInterrupt,
|
|
||||||
):
|
|
||||||
|
|
||||||
with trio.CancelScope(shield=True):
|
with trio.CancelScope(shield=True):
|
||||||
await cntr.cancel()
|
await cntr.cancel(stop_msg)
|
||||||
|
|
||||||
raise
|
|
||||||
|
|
||||||
|
|
||||||
async def start_ahab(
|
async def start_ahab(
|
||||||
|
|
|
@ -22,7 +22,7 @@ financial data flows.
|
||||||
from __future__ import annotations
|
from __future__ import annotations
|
||||||
from collections import Counter
|
from collections import Counter
|
||||||
import time
|
import time
|
||||||
from typing import TYPE_CHECKING, Optional
|
from typing import TYPE_CHECKING, Optional, Union
|
||||||
|
|
||||||
import tractor
|
import tractor
|
||||||
import trio
|
import trio
|
||||||
|
@ -32,6 +32,7 @@ from ..log import get_logger
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
from ._sharedmem import ShmArray
|
from ._sharedmem import ShmArray
|
||||||
|
from .feed import _FeedsBus
|
||||||
|
|
||||||
log = get_logger(__name__)
|
log = get_logger(__name__)
|
||||||
|
|
||||||
|
@ -219,7 +220,7 @@ async def iter_ohlc_periods(
|
||||||
|
|
||||||
async def sample_and_broadcast(
|
async def sample_and_broadcast(
|
||||||
|
|
||||||
bus: '_FeedsBus', # noqa
|
bus: _FeedsBus, # noqa
|
||||||
shm: ShmArray,
|
shm: ShmArray,
|
||||||
quote_stream: trio.abc.ReceiveChannel,
|
quote_stream: trio.abc.ReceiveChannel,
|
||||||
brokername: str,
|
brokername: str,
|
||||||
|
@ -298,7 +299,13 @@ async def sample_and_broadcast(
|
||||||
# end up triggering backpressure which which will
|
# end up triggering backpressure which which will
|
||||||
# eventually block this producer end of the feed and
|
# eventually block this producer end of the feed and
|
||||||
# thus other consumers still attached.
|
# thus other consumers still attached.
|
||||||
subs = bus._subscribers[broker_symbol.lower()]
|
subs: list[
|
||||||
|
tuple[
|
||||||
|
Union[tractor.MsgStream, trio.MemorySendChannel],
|
||||||
|
tractor.Context,
|
||||||
|
Optional[float], # tick throttle in Hz
|
||||||
|
]
|
||||||
|
] = bus._subscribers[broker_symbol.lower()]
|
||||||
|
|
||||||
# NOTE: by default the broker backend doesn't append
|
# NOTE: by default the broker backend doesn't append
|
||||||
# it's own "name" into the fqsn schema (but maybe it
|
# it's own "name" into the fqsn schema (but maybe it
|
||||||
|
@ -307,7 +314,7 @@ async def sample_and_broadcast(
|
||||||
bsym = f'{broker_symbol}.{brokername}'
|
bsym = f'{broker_symbol}.{brokername}'
|
||||||
lags: int = 0
|
lags: int = 0
|
||||||
|
|
||||||
for (stream, tick_throttle) in subs:
|
for (stream, ctx, tick_throttle) in subs:
|
||||||
|
|
||||||
try:
|
try:
|
||||||
with trio.move_on_after(0.2) as cs:
|
with trio.move_on_after(0.2) as cs:
|
||||||
|
@ -319,25 +326,41 @@ async def sample_and_broadcast(
|
||||||
(bsym, quote)
|
(bsym, quote)
|
||||||
)
|
)
|
||||||
except trio.WouldBlock:
|
except trio.WouldBlock:
|
||||||
ctx = getattr(stream, '_ctx', None)
|
chan = ctx.chan
|
||||||
if ctx:
|
if ctx:
|
||||||
log.warning(
|
log.warning(
|
||||||
f'Feed overrun {bus.brokername} ->'
|
f'Feed overrun {bus.brokername} ->'
|
||||||
f'{ctx.channel.uid} !!!'
|
f'{chan.uid} !!!'
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
key = id(stream)
|
key = id(stream)
|
||||||
overruns[key] += 1
|
overruns[key] += 1
|
||||||
log.warning(
|
log.warning(
|
||||||
f'Feed overrun {bus.brokername} -> '
|
f'Feed overrun {broker_symbol}'
|
||||||
|
'@{bus.brokername} -> '
|
||||||
f'feed @ {tick_throttle} Hz'
|
f'feed @ {tick_throttle} Hz'
|
||||||
)
|
)
|
||||||
if overruns[key] > 6:
|
if overruns[key] > 6:
|
||||||
|
# TODO: should we check for the
|
||||||
|
# context being cancelled? this
|
||||||
|
# could happen but the
|
||||||
|
# channel-ipc-pipe is still up.
|
||||||
|
if not chan.connected():
|
||||||
log.warning(
|
log.warning(
|
||||||
f'Dropping consumer {stream}'
|
'Dropping broken consumer:\n'
|
||||||
|
f'{broker_symbol}:'
|
||||||
|
f'{ctx.cid}@{chan.uid}'
|
||||||
)
|
)
|
||||||
await stream.aclose()
|
await stream.aclose()
|
||||||
raise trio.BrokenResourceError
|
raise trio.BrokenResourceError
|
||||||
|
else:
|
||||||
|
log.warning(
|
||||||
|
'Feed getting overrun bro!\n'
|
||||||
|
f'{broker_symbol}:'
|
||||||
|
f'{ctx.cid}@{chan.uid}'
|
||||||
|
)
|
||||||
|
continue
|
||||||
|
|
||||||
else:
|
else:
|
||||||
await stream.send(
|
await stream.send(
|
||||||
{bsym: quote}
|
{bsym: quote}
|
||||||
|
@ -353,11 +376,12 @@ async def sample_and_broadcast(
|
||||||
trio.ClosedResourceError,
|
trio.ClosedResourceError,
|
||||||
trio.EndOfChannel,
|
trio.EndOfChannel,
|
||||||
):
|
):
|
||||||
ctx = getattr(stream, '_ctx', None)
|
chan = ctx.chan
|
||||||
if ctx:
|
if ctx:
|
||||||
log.warning(
|
log.warning(
|
||||||
f'{ctx.chan.uid} dropped '
|
'Dropped `brokerd`-quotes-feed connection:\n'
|
||||||
'`brokerd`-quotes-feed connection'
|
f'{broker_symbol}:'
|
||||||
|
f'{ctx.cid}@{chan.uid}'
|
||||||
)
|
)
|
||||||
if tick_throttle:
|
if tick_throttle:
|
||||||
assert stream._closed
|
assert stream._closed
|
||||||
|
@ -370,7 +394,11 @@ async def sample_and_broadcast(
|
||||||
try:
|
try:
|
||||||
subs.remove((stream, tick_throttle))
|
subs.remove((stream, tick_throttle))
|
||||||
except ValueError:
|
except ValueError:
|
||||||
log.error(f'{stream} was already removed from subs!?')
|
log.error(
|
||||||
|
f'Stream was already removed from subs!?\n'
|
||||||
|
f'{broker_symbol}:'
|
||||||
|
f'{ctx.cid}@{chan.uid}'
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
# TODO: a less naive throttler, here's some snippets:
|
# TODO: a less naive throttler, here's some snippets:
|
||||||
|
@ -482,6 +510,7 @@ async def uniform_rate_send(
|
||||||
# if the feed consumer goes down then drop
|
# if the feed consumer goes down then drop
|
||||||
# out of this rate limiter
|
# out of this rate limiter
|
||||||
log.warning(f'{stream} closed')
|
log.warning(f'{stream} closed')
|
||||||
|
await stream.aclose()
|
||||||
return
|
return
|
||||||
|
|
||||||
# reset send cycle state
|
# reset send cycle state
|
||||||
|
|
|
@ -20,6 +20,7 @@ NumPy compatible shared memory buffers for real-time IPC streaming.
|
||||||
"""
|
"""
|
||||||
from __future__ import annotations
|
from __future__ import annotations
|
||||||
from sys import byteorder
|
from sys import byteorder
|
||||||
|
import time
|
||||||
from typing import Optional
|
from typing import Optional
|
||||||
from multiprocessing.shared_memory import SharedMemory, _USE_POSIX
|
from multiprocessing.shared_memory import SharedMemory, _USE_POSIX
|
||||||
|
|
||||||
|
@ -98,7 +99,12 @@ class SharedInt:
|
||||||
if _USE_POSIX:
|
if _USE_POSIX:
|
||||||
# We manually unlink to bypass all the "resource tracker"
|
# We manually unlink to bypass all the "resource tracker"
|
||||||
# nonsense meant for non-SC systems.
|
# nonsense meant for non-SC systems.
|
||||||
shm_unlink(self._shm.name)
|
name = self._shm.name
|
||||||
|
try:
|
||||||
|
shm_unlink(name)
|
||||||
|
except FileNotFoundError:
|
||||||
|
# might be a teardown race here?
|
||||||
|
log.warning(f'Shm for {name} already unlinked?')
|
||||||
|
|
||||||
|
|
||||||
class _Token(BaseModel):
|
class _Token(BaseModel):
|
||||||
|
@ -536,8 +542,26 @@ def attach_shm_array(
|
||||||
if key in _known_tokens:
|
if key in _known_tokens:
|
||||||
assert _Token.from_msg(_known_tokens[key]) == token, "WTF"
|
assert _Token.from_msg(_known_tokens[key]) == token, "WTF"
|
||||||
|
|
||||||
|
# XXX: ugh, looks like due to the ``shm_open()`` C api we can't
|
||||||
|
# actually place files in a subdir, see discussion here:
|
||||||
|
# https://stackoverflow.com/a/11103289
|
||||||
|
|
||||||
# attach to array buffer and view as per dtype
|
# attach to array buffer and view as per dtype
|
||||||
shm = SharedMemory(name=key)
|
_err: Optional[Exception] = None
|
||||||
|
for _ in range(3):
|
||||||
|
try:
|
||||||
|
shm = SharedMemory(
|
||||||
|
name=key,
|
||||||
|
create=False,
|
||||||
|
)
|
||||||
|
break
|
||||||
|
except OSError as oserr:
|
||||||
|
_err = oserr
|
||||||
|
time.sleep(0.1)
|
||||||
|
else:
|
||||||
|
if _err:
|
||||||
|
raise _err
|
||||||
|
|
||||||
shmarr = np.ndarray(
|
shmarr = np.ndarray(
|
||||||
(size,),
|
(size,),
|
||||||
dtype=token.dtype,
|
dtype=token.dtype,
|
||||||
|
|
|
@ -33,6 +33,7 @@ from typing import (
|
||||||
Generator,
|
Generator,
|
||||||
Awaitable,
|
Awaitable,
|
||||||
TYPE_CHECKING,
|
TYPE_CHECKING,
|
||||||
|
Union,
|
||||||
)
|
)
|
||||||
|
|
||||||
import trio
|
import trio
|
||||||
|
@ -40,12 +41,13 @@ from trio.abc import ReceiveChannel
|
||||||
from trio_typing import TaskStatus
|
from trio_typing import TaskStatus
|
||||||
import trimeter
|
import trimeter
|
||||||
import tractor
|
import tractor
|
||||||
|
from tractor.trionics import maybe_open_context
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
import pendulum
|
import pendulum
|
||||||
import numpy as np
|
import numpy as np
|
||||||
|
|
||||||
from ..brokers import get_brokermod
|
from ..brokers import get_brokermod
|
||||||
from .._cacheables import maybe_open_context
|
from ..calc import humanize
|
||||||
from ..log import get_logger, get_console_log
|
from ..log import get_logger, get_console_log
|
||||||
from .._daemon import (
|
from .._daemon import (
|
||||||
maybe_spawn_brokerd,
|
maybe_spawn_brokerd,
|
||||||
|
@ -115,7 +117,13 @@ class _FeedsBus(BaseModel):
|
||||||
# https://github.com/samuelcolvin/pydantic/issues/2816
|
# https://github.com/samuelcolvin/pydantic/issues/2816
|
||||||
_subscribers: dict[
|
_subscribers: dict[
|
||||||
str,
|
str,
|
||||||
list[tuple[tractor.MsgStream, Optional[float]]]
|
list[
|
||||||
|
tuple[
|
||||||
|
Union[tractor.MsgStream, trio.MemorySendChannel],
|
||||||
|
tractor.Context,
|
||||||
|
Optional[float], # tick throttle in Hz
|
||||||
|
]
|
||||||
|
]
|
||||||
] = {}
|
] = {}
|
||||||
|
|
||||||
async def start_task(
|
async def start_task(
|
||||||
|
@ -336,7 +344,7 @@ async def start_backfill(
|
||||||
log.debug(f'New datetime index:\n{pformat(dtrange)}')
|
log.debug(f'New datetime index:\n{pformat(dtrange)}')
|
||||||
|
|
||||||
for end_dt in dtrange:
|
for end_dt in dtrange:
|
||||||
log.warning(f'Yielding next frame start {end_dt}')
|
log.info(f'Yielding next frame start {end_dt}')
|
||||||
start = yield end_dt
|
start = yield end_dt
|
||||||
|
|
||||||
# if caller sends a new start date, reset to that
|
# if caller sends a new start date, reset to that
|
||||||
|
@ -692,6 +700,7 @@ async def manage_history(
|
||||||
|
|
||||||
bfqsn = fqsn.replace('.' + mod.name, '')
|
bfqsn = fqsn.replace('.' + mod.name, '')
|
||||||
open_history_client = getattr(mod, 'open_history_client', None)
|
open_history_client = getattr(mod, 'open_history_client', None)
|
||||||
|
assert open_history_client
|
||||||
|
|
||||||
if is_up and opened and open_history_client:
|
if is_up and opened and open_history_client:
|
||||||
|
|
||||||
|
@ -720,6 +729,7 @@ async def manage_history(
|
||||||
bfqsn,
|
bfqsn,
|
||||||
shm,
|
shm,
|
||||||
last_tsdb_dt=last_tsdb_dt,
|
last_tsdb_dt=last_tsdb_dt,
|
||||||
|
tsdb_is_up=True,
|
||||||
storage=storage,
|
storage=storage,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
|
@ -1115,10 +1125,10 @@ async def open_feed_bus(
|
||||||
recv,
|
recv,
|
||||||
stream,
|
stream,
|
||||||
)
|
)
|
||||||
sub = (send, tick_throttle)
|
sub = (send, ctx, tick_throttle)
|
||||||
|
|
||||||
else:
|
else:
|
||||||
sub = (stream, tick_throttle)
|
sub = (stream, ctx, tick_throttle)
|
||||||
|
|
||||||
subs = bus._subscribers[bfqsn]
|
subs = bus._subscribers[bfqsn]
|
||||||
subs.append(sub)
|
subs.append(sub)
|
||||||
|
@ -1201,10 +1211,10 @@ class Feed:
|
||||||
shm: ShmArray
|
shm: ShmArray
|
||||||
mod: ModuleType
|
mod: ModuleType
|
||||||
first_quotes: dict # symbol names to first quote dicts
|
first_quotes: dict # symbol names to first quote dicts
|
||||||
|
|
||||||
_portal: tractor.Portal
|
_portal: tractor.Portal
|
||||||
|
|
||||||
stream: trio.abc.ReceiveChannel[dict[str, Any]]
|
stream: trio.abc.ReceiveChannel[dict[str, Any]]
|
||||||
|
status: dict[str, Any]
|
||||||
|
|
||||||
throttle_rate: Optional[int] = None
|
throttle_rate: Optional[int] = None
|
||||||
|
|
||||||
_trade_stream: Optional[AsyncIterator[dict[str, Any]]] = None
|
_trade_stream: Optional[AsyncIterator[dict[str, Any]]] = None
|
||||||
|
@ -1345,9 +1355,24 @@ async def open_feed(
|
||||||
first_quotes=first_quotes,
|
first_quotes=first_quotes,
|
||||||
stream=stream,
|
stream=stream,
|
||||||
_portal=portal,
|
_portal=portal,
|
||||||
|
status={},
|
||||||
throttle_rate=tick_throttle,
|
throttle_rate=tick_throttle,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# fill out "status info" that the UI can show
|
||||||
|
host, port = feed.portal.channel.raddr
|
||||||
|
if host == '127.0.0.1':
|
||||||
|
host = 'localhost'
|
||||||
|
|
||||||
|
feed.status.update({
|
||||||
|
'actor_name': feed.portal.channel.uid[0],
|
||||||
|
'host': host,
|
||||||
|
'port': port,
|
||||||
|
'shm': f'{humanize(feed.shm._shm.size)}',
|
||||||
|
'throttle_rate': feed.throttle_rate,
|
||||||
|
})
|
||||||
|
feed.status.update(init_msg.pop('status', {}))
|
||||||
|
|
||||||
for sym, data in init_msg.items():
|
for sym, data in init_msg.items():
|
||||||
si = data['symbol_info']
|
si = data['symbol_info']
|
||||||
fqsn = data['fqsn'] + f'.{brokername}'
|
fqsn = data['fqsn'] + f'.{brokername}'
|
||||||
|
|
|
@ -127,10 +127,15 @@ def start_marketstore(
|
||||||
import os
|
import os
|
||||||
import docker
|
import docker
|
||||||
from .. import config
|
from .. import config
|
||||||
|
|
||||||
get_console_log('info', name=__name__)
|
get_console_log('info', name=__name__)
|
||||||
|
|
||||||
yml_file = os.path.join(config._config_dir, 'mkts.yml')
|
mktsdir = os.path.join(config._config_dir, 'marketstore')
|
||||||
|
|
||||||
|
# create when dne
|
||||||
|
if not os.path.isdir(mktsdir):
|
||||||
|
os.mkdir(mktsdir)
|
||||||
|
|
||||||
|
yml_file = os.path.join(mktsdir, 'mkts.yml')
|
||||||
if not os.path.isfile(yml_file):
|
if not os.path.isfile(yml_file):
|
||||||
log.warning(
|
log.warning(
|
||||||
f'No `marketstore` config exists?: {yml_file}\n'
|
f'No `marketstore` config exists?: {yml_file}\n'
|
||||||
|
@ -143,14 +148,14 @@ def start_marketstore(
|
||||||
# create a mount from user's local piker config dir into container
|
# create a mount from user's local piker config dir into container
|
||||||
config_dir_mnt = docker.types.Mount(
|
config_dir_mnt = docker.types.Mount(
|
||||||
target='/etc',
|
target='/etc',
|
||||||
source=config._config_dir,
|
source=mktsdir,
|
||||||
type='bind',
|
type='bind',
|
||||||
)
|
)
|
||||||
|
|
||||||
# create a user config subdir where the marketstore
|
# create a user config subdir where the marketstore
|
||||||
# backing filesystem database can be persisted.
|
# backing filesystem database can be persisted.
|
||||||
persistent_data_dir = os.path.join(
|
persistent_data_dir = os.path.join(
|
||||||
config._config_dir, 'data',
|
mktsdir, 'data',
|
||||||
)
|
)
|
||||||
if not os.path.isdir(persistent_data_dir):
|
if not os.path.isdir(persistent_data_dir):
|
||||||
os.mkdir(persistent_data_dir)
|
os.mkdir(persistent_data_dir)
|
||||||
|
@ -180,7 +185,14 @@ def start_marketstore(
|
||||||
init=True,
|
init=True,
|
||||||
# remove=True,
|
# remove=True,
|
||||||
)
|
)
|
||||||
return dcntr, _config
|
return (
|
||||||
|
dcntr,
|
||||||
|
_config,
|
||||||
|
|
||||||
|
# expected startup and stop msgs
|
||||||
|
"launching tcp listener for all services...",
|
||||||
|
"exiting...",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
_tick_tbk_ids: tuple[str, str] = ('1Sec', 'TICK')
|
_tick_tbk_ids: tuple[str, str] = ('1Sec', 'TICK')
|
||||||
|
@ -383,7 +395,12 @@ class Storage:
|
||||||
]:
|
]:
|
||||||
|
|
||||||
first_tsdb_dt, last_tsdb_dt = None, None
|
first_tsdb_dt, last_tsdb_dt = None, None
|
||||||
tsdb_arrays = await self.read_ohlcv(fqsn)
|
tsdb_arrays = await self.read_ohlcv(
|
||||||
|
fqsn,
|
||||||
|
# on first load we don't need to pull the max
|
||||||
|
# history per request size worth.
|
||||||
|
limit=3000,
|
||||||
|
)
|
||||||
log.info(f'Loaded tsdb history {tsdb_arrays}')
|
log.info(f'Loaded tsdb history {tsdb_arrays}')
|
||||||
|
|
||||||
if tsdb_arrays:
|
if tsdb_arrays:
|
||||||
|
@ -401,6 +418,7 @@ class Storage:
|
||||||
fqsn: str,
|
fqsn: str,
|
||||||
timeframe: Optional[Union[int, str]] = None,
|
timeframe: Optional[Union[int, str]] = None,
|
||||||
end: Optional[int] = None,
|
end: Optional[int] = None,
|
||||||
|
limit: int = int(800e3),
|
||||||
|
|
||||||
) -> tuple[
|
) -> tuple[
|
||||||
MarketstoreClient,
|
MarketstoreClient,
|
||||||
|
@ -423,7 +441,7 @@ class Storage:
|
||||||
|
|
||||||
# TODO: figure the max limit here given the
|
# TODO: figure the max limit here given the
|
||||||
# ``purepc`` msg size limit of purerpc: 33554432
|
# ``purepc`` msg size limit of purerpc: 33554432
|
||||||
limit=int(800e3),
|
limit=limit,
|
||||||
)
|
)
|
||||||
|
|
||||||
if timeframe is None:
|
if timeframe is None:
|
||||||
|
@ -639,12 +657,13 @@ async def tsdb_history_update(
|
||||||
tsdb_arrays = await storage.read_ohlcv(fqsn)
|
tsdb_arrays = await storage.read_ohlcv(fqsn)
|
||||||
# hist diffing
|
# hist diffing
|
||||||
if tsdb_arrays:
|
if tsdb_arrays:
|
||||||
onesec = tsdb_arrays[1]
|
for secs in (1, 60):
|
||||||
|
ts = tsdb_arrays.get(secs)
|
||||||
|
if ts is not None and len(ts):
|
||||||
# these aren't currently used but can be referenced from
|
# these aren't currently used but can be referenced from
|
||||||
# within the embedded ipython shell below.
|
# within the embedded ipython shell below.
|
||||||
to_append = ohlcv[ohlcv['time'] > onesec['Epoch'][-1]]
|
to_append = ohlcv[ohlcv['time'] > ts['Epoch'][-1]]
|
||||||
to_prepend = ohlcv[ohlcv['time'] < onesec['Epoch'][0]]
|
to_prepend = ohlcv[ohlcv['time'] < ts['Epoch'][0]]
|
||||||
|
|
||||||
profiler('Finished db arrays diffs')
|
profiler('Finished db arrays diffs')
|
||||||
|
|
||||||
|
|
|
@ -361,7 +361,7 @@ async def cascade(
|
||||||
) -> tuple[TaskTracker, int]:
|
) -> tuple[TaskTracker, int]:
|
||||||
# TODO: adopt an incremental update engine/approach
|
# TODO: adopt an incremental update engine/approach
|
||||||
# where possible here eventually!
|
# where possible here eventually!
|
||||||
log.warning(f're-syncing fsp {func_name} to source')
|
log.debug(f're-syncing fsp {func_name} to source')
|
||||||
tracker.cs.cancel()
|
tracker.cs.cancel()
|
||||||
await tracker.complete.wait()
|
await tracker.complete.wait()
|
||||||
tracker, index = await n.start(fsp_target)
|
tracker, index = await n.start(fsp_target)
|
||||||
|
|
|
@ -25,10 +25,13 @@ from pygments import highlight, lexers, formatters
|
||||||
|
|
||||||
# Makes it so we only see the full module name when using ``__name__``
|
# Makes it so we only see the full module name when using ``__name__``
|
||||||
# without the extra "piker." prefix.
|
# without the extra "piker." prefix.
|
||||||
_proj_name = 'piker'
|
_proj_name: str = 'piker'
|
||||||
|
|
||||||
|
|
||||||
def get_logger(name: str = None) -> logging.Logger:
|
def get_logger(
|
||||||
|
name: str = None,
|
||||||
|
|
||||||
|
) -> logging.Logger:
|
||||||
'''Return the package log or a sub-log for `name` if provided.
|
'''Return the package log or a sub-log for `name` if provided.
|
||||||
'''
|
'''
|
||||||
return tractor.log.get_logger(name=name, _root_name=_proj_name)
|
return tractor.log.get_logger(name=name, _root_name=_proj_name)
|
||||||
|
|
|
@ -50,7 +50,10 @@ from ._cursor import (
|
||||||
from ..data._sharedmem import ShmArray
|
from ..data._sharedmem import ShmArray
|
||||||
from ._l1 import L1Labels
|
from ._l1 import L1Labels
|
||||||
from ._ohlc import BarItems
|
from ._ohlc import BarItems
|
||||||
from ._curve import FastAppendCurve
|
from ._curve import (
|
||||||
|
Curve,
|
||||||
|
StepCurve,
|
||||||
|
)
|
||||||
from ._style import (
|
from ._style import (
|
||||||
hcolor,
|
hcolor,
|
||||||
CHART_MARGINS,
|
CHART_MARGINS,
|
||||||
|
@ -979,11 +982,6 @@ class ChartPlotWidget(pg.PlotWidget):
|
||||||
graphics=graphics,
|
graphics=graphics,
|
||||||
)
|
)
|
||||||
|
|
||||||
# TODO: i think we can eventually remove this if
|
|
||||||
# we write the ``Flow.update_graphics()`` method right?
|
|
||||||
# draw after to allow self.scene() to work...
|
|
||||||
graphics.draw_from_data(shm.array)
|
|
||||||
|
|
||||||
self._add_sticky(name, bg_color='davies')
|
self._add_sticky(name, bg_color='davies')
|
||||||
|
|
||||||
return graphics, data_key
|
return graphics, data_key
|
||||||
|
@ -1056,6 +1054,7 @@ class ChartPlotWidget(pg.PlotWidget):
|
||||||
color: Optional[str] = None,
|
color: Optional[str] = None,
|
||||||
add_label: bool = True,
|
add_label: bool = True,
|
||||||
pi: Optional[pg.PlotItem] = None,
|
pi: Optional[pg.PlotItem] = None,
|
||||||
|
step_mode: bool = False,
|
||||||
|
|
||||||
**pdi_kwargs,
|
**pdi_kwargs,
|
||||||
|
|
||||||
|
@ -1072,31 +1071,18 @@ class ChartPlotWidget(pg.PlotWidget):
|
||||||
|
|
||||||
data_key = array_key or name
|
data_key = array_key or name
|
||||||
|
|
||||||
# yah, we wrote our own B)
|
curve_type = {
|
||||||
data = shm.array
|
None: Curve,
|
||||||
curve = FastAppendCurve(
|
'step': StepCurve,
|
||||||
y=data[data_key],
|
# TODO:
|
||||||
x=data['index'],
|
# 'bars': BarsItems
|
||||||
# antialias=True,
|
}['step' if step_mode else None]
|
||||||
|
|
||||||
|
curve = curve_type(
|
||||||
name=name,
|
name=name,
|
||||||
|
|
||||||
# XXX: pretty sure this is just more overhead
|
|
||||||
# on data reads and makes graphics rendering no faster
|
|
||||||
# clipToView=True,
|
|
||||||
|
|
||||||
**pdi_kwargs,
|
**pdi_kwargs,
|
||||||
)
|
)
|
||||||
|
|
||||||
# XXX: see explanation for different caching modes:
|
|
||||||
# https://stackoverflow.com/a/39410081
|
|
||||||
# seems to only be useful if we don't re-generate the entire
|
|
||||||
# QPainterPath every time
|
|
||||||
# curve.setCacheMode(QtWidgets.QGraphicsItem.DeviceCoordinateCache)
|
|
||||||
|
|
||||||
# don't ever use this - it's a colossal nightmare of artefacts
|
|
||||||
# and is disastrous for performance.
|
|
||||||
# curve.setCacheMode(QtWidgets.QGraphicsItem.ItemCoordinateCache)
|
|
||||||
|
|
||||||
pi = pi or self.plotItem
|
pi = pi or self.plotItem
|
||||||
|
|
||||||
self._flows[data_key] = Flow(
|
self._flows[data_key] = Flow(
|
||||||
|
@ -1258,7 +1244,9 @@ class ChartPlotWidget(pg.PlotWidget):
|
||||||
def maxmin(
|
def maxmin(
|
||||||
self,
|
self,
|
||||||
name: Optional[str] = None,
|
name: Optional[str] = None,
|
||||||
bars_range: Optional[tuple[int, int, int, int]] = None,
|
bars_range: Optional[tuple[
|
||||||
|
int, int, int, int, int, int
|
||||||
|
]] = None,
|
||||||
|
|
||||||
) -> tuple[float, float]:
|
) -> tuple[float, float]:
|
||||||
'''
|
'''
|
||||||
|
@ -1267,6 +1255,7 @@ class ChartPlotWidget(pg.PlotWidget):
|
||||||
If ``bars_range`` is provided use that range.
|
If ``bars_range`` is provided use that range.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
|
# print(f'Chart[{self.name}].maxmin()')
|
||||||
profiler = pg.debug.Profiler(
|
profiler = pg.debug.Profiler(
|
||||||
msg=f'`{str(self)}.maxmin(name={name})`: `{self.name}`',
|
msg=f'`{str(self)}.maxmin(name={name})`: `{self.name}`',
|
||||||
disabled=not pg_profile_enabled(),
|
disabled=not pg_profile_enabled(),
|
||||||
|
@ -1286,7 +1275,14 @@ class ChartPlotWidget(pg.PlotWidget):
|
||||||
key = res = 0, 0
|
key = res = 0, 0
|
||||||
|
|
||||||
else:
|
else:
|
||||||
first, l, lbar, rbar, r, last = bars_range or flow.datums_range()
|
(
|
||||||
|
first,
|
||||||
|
l,
|
||||||
|
lbar,
|
||||||
|
rbar,
|
||||||
|
r,
|
||||||
|
last,
|
||||||
|
) = bars_range or flow.datums_range()
|
||||||
profiler(f'{self.name} got bars range')
|
profiler(f'{self.name} got bars range')
|
||||||
|
|
||||||
key = round(lbar), round(rbar)
|
key = round(lbar), round(rbar)
|
||||||
|
|
|
@ -138,51 +138,20 @@ def ohlc_flatten(
|
||||||
return x, flat
|
return x, flat
|
||||||
|
|
||||||
|
|
||||||
def ohlc_to_m4_line(
|
|
||||||
ohlc: np.ndarray,
|
|
||||||
px_width: int,
|
|
||||||
|
|
||||||
downsample: bool = False,
|
|
||||||
uppx: Optional[float] = None,
|
|
||||||
pretrace: bool = False,
|
|
||||||
|
|
||||||
) -> tuple[np.ndarray, np.ndarray]:
|
|
||||||
'''
|
|
||||||
Convert an OHLC struct-array to a m4 downsampled 1-d array.
|
|
||||||
|
|
||||||
'''
|
|
||||||
xpts, flat = ohlc_flatten(
|
|
||||||
ohlc,
|
|
||||||
use_mxmn=pretrace,
|
|
||||||
)
|
|
||||||
|
|
||||||
if downsample:
|
|
||||||
bins, x, y = ds_m4(
|
|
||||||
xpts,
|
|
||||||
flat,
|
|
||||||
px_width=px_width,
|
|
||||||
uppx=uppx,
|
|
||||||
# log_scale=bool(uppx)
|
|
||||||
)
|
|
||||||
x = np.broadcast_to(x[:, None], y.shape)
|
|
||||||
x = (x + np.array([-0.43, 0, 0, 0.43])).flatten()
|
|
||||||
y = y.flatten()
|
|
||||||
|
|
||||||
return x, y
|
|
||||||
else:
|
|
||||||
return xpts, flat
|
|
||||||
|
|
||||||
|
|
||||||
def ds_m4(
|
def ds_m4(
|
||||||
x: np.ndarray,
|
x: np.ndarray,
|
||||||
y: np.ndarray,
|
y: np.ndarray,
|
||||||
|
# units-per-pixel-x(dimension)
|
||||||
|
uppx: float,
|
||||||
|
|
||||||
|
# XXX: troll zone / easter egg..
|
||||||
|
# want to mess with ur pal, pass in the actual
|
||||||
|
# pixel width here instead of uppx-proper (i.e. pass
|
||||||
|
# in our ``pg.GraphicsObject`` derivative's ``.px_width()``
|
||||||
|
# gto mega-trip-out ur bud). Hint, it used to be implemented
|
||||||
|
# (wrongly) using "pixel width", so check the git history ;)
|
||||||
|
|
||||||
# this is the width of the data in view
|
|
||||||
# in display-device-local pixel units.
|
|
||||||
px_width: int,
|
|
||||||
uppx: Optional[float] = None,
|
|
||||||
xrange: Optional[float] = None,
|
xrange: Optional[float] = None,
|
||||||
# log_scale: bool = True,
|
|
||||||
|
|
||||||
) -> tuple[int, np.ndarray, np.ndarray]:
|
) -> tuple[int, np.ndarray, np.ndarray]:
|
||||||
'''
|
'''
|
||||||
|
@ -209,29 +178,8 @@ def ds_m4(
|
||||||
# "i didn't show it in the sample code, but it's accounted for
|
# "i didn't show it in the sample code, but it's accounted for
|
||||||
# in the start and end indices and number of bins"
|
# in the start and end indices and number of bins"
|
||||||
|
|
||||||
# optionally log-scale down the "supposed pxs on screen"
|
|
||||||
# as the units-per-px (uppx) get's large.
|
|
||||||
# if log_scale:
|
|
||||||
# assert uppx, 'You must provide a `uppx` value to use log scaling!'
|
|
||||||
# # uppx = uppx * math.log(uppx, 2)
|
|
||||||
|
|
||||||
# # scaler = 2**7 / (1 + math.log(uppx, 2))
|
|
||||||
# scaler = round(
|
|
||||||
# max(
|
|
||||||
# # NOTE: found that a 16x px width brought greater
|
|
||||||
# # detail, likely due to dpi scaling?
|
|
||||||
# # px_width=px_width * 16,
|
|
||||||
# 2**7 / (1 + math.log(uppx, 2)),
|
|
||||||
# 1
|
|
||||||
# )
|
|
||||||
# )
|
|
||||||
# px_width *= scaler
|
|
||||||
|
|
||||||
# else:
|
|
||||||
# px_width *= 16
|
|
||||||
|
|
||||||
# should never get called unless actually needed
|
# should never get called unless actually needed
|
||||||
assert px_width > 1 and uppx > 0
|
assert uppx > 1
|
||||||
|
|
||||||
# NOTE: if we didn't pre-slice the data to downsample
|
# NOTE: if we didn't pre-slice the data to downsample
|
||||||
# you could in theory pass these as the slicing params,
|
# you could in theory pass these as the slicing params,
|
||||||
|
@ -248,16 +196,9 @@ def ds_m4(
|
||||||
# uppx *= max(4 / (1 + math.log(uppx, 2)), 1)
|
# uppx *= max(4 / (1 + math.log(uppx, 2)), 1)
|
||||||
|
|
||||||
pxw = math.ceil(xrange / uppx)
|
pxw = math.ceil(xrange / uppx)
|
||||||
# px_width = math.ceil(px_width)
|
|
||||||
|
|
||||||
# ratio of indexed x-value to width of raster in pixels.
|
# scale up the frame "width" directly with uppx
|
||||||
# this is more or less, uppx: units-per-pixel.
|
w = uppx
|
||||||
# w = xrange / float(px_width)
|
|
||||||
# uppx = uppx * math.log(uppx, 2)
|
|
||||||
# w2 = px_width / uppx
|
|
||||||
|
|
||||||
# scale up the width as the uppx get's large
|
|
||||||
w = uppx # * math.log(uppx, 666)
|
|
||||||
|
|
||||||
# ensure we make more then enough
|
# ensure we make more then enough
|
||||||
# frames (windows) for the output pixel
|
# frames (windows) for the output pixel
|
||||||
|
@ -276,9 +217,7 @@ def ds_m4(
|
||||||
# print(
|
# print(
|
||||||
# f'uppx: {uppx}\n'
|
# f'uppx: {uppx}\n'
|
||||||
# f'xrange: {xrange}\n'
|
# f'xrange: {xrange}\n'
|
||||||
# f'px_width: {px_width}\n'
|
|
||||||
# f'pxw: {pxw}\n'
|
# f'pxw: {pxw}\n'
|
||||||
# f'WTF w:{w}, w2:{w2}\n'
|
|
||||||
# f'frames: {frames}\n'
|
# f'frames: {frames}\n'
|
||||||
# )
|
# )
|
||||||
assert frames >= (xrange / uppx)
|
assert frames >= (xrange / uppx)
|
||||||
|
|
|
@ -98,25 +98,30 @@ class LineDot(pg.CurvePoint):
|
||||||
ev: QtCore.QEvent,
|
ev: QtCore.QEvent,
|
||||||
|
|
||||||
) -> bool:
|
) -> bool:
|
||||||
if not isinstance(
|
|
||||||
ev, QtCore.QDynamicPropertyChangeEvent
|
if (
|
||||||
) or self.curve() is None:
|
not isinstance(ev, QtCore.QDynamicPropertyChangeEvent)
|
||||||
|
or self.curve() is None
|
||||||
|
):
|
||||||
return False
|
return False
|
||||||
|
|
||||||
# TODO: get rid of this ``.getData()`` and
|
# TODO: get rid of this ``.getData()`` and
|
||||||
# make a more pythonic api to retreive backing
|
# make a more pythonic api to retreive backing
|
||||||
# numpy arrays...
|
# numpy arrays...
|
||||||
(x, y) = self.curve().getData()
|
# (x, y) = self.curve().getData()
|
||||||
index = self.property('index')
|
# index = self.property('index')
|
||||||
# first = self._plot._arrays['ohlc'][0]['index']
|
# # first = self._plot._arrays['ohlc'][0]['index']
|
||||||
# first = x[0]
|
# # first = x[0]
|
||||||
# i = index - first
|
# # i = index - first
|
||||||
if index:
|
# if index:
|
||||||
i = round(index - x[0])
|
# i = round(index - x[0])
|
||||||
if i > 0 and i < len(y):
|
# if i > 0 and i < len(y):
|
||||||
newPos = (index, y[i])
|
# newPos = (index, y[i])
|
||||||
QtWidgets.QGraphicsItem.setPos(self, *newPos)
|
# QtWidgets.QGraphicsItem.setPos(
|
||||||
return True
|
# self,
|
||||||
|
# *newPos,
|
||||||
|
# )
|
||||||
|
# return True
|
||||||
|
|
||||||
return False
|
return False
|
||||||
|
|
||||||
|
|
|
@ -18,27 +18,31 @@
|
||||||
Fast, smooth, sexy curves.
|
Fast, smooth, sexy curves.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
from typing import Optional
|
from contextlib import contextmanager as cm
|
||||||
|
from typing import Optional, Callable
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
import pyqtgraph as pg
|
import pyqtgraph as pg
|
||||||
from PyQt5 import QtGui, QtWidgets
|
from PyQt5 import QtWidgets
|
||||||
from PyQt5.QtWidgets import QGraphicsItem
|
from PyQt5.QtWidgets import QGraphicsItem
|
||||||
from PyQt5.QtCore import (
|
from PyQt5.QtCore import (
|
||||||
Qt,
|
Qt,
|
||||||
QLineF,
|
QLineF,
|
||||||
QSizeF,
|
QSizeF,
|
||||||
QRectF,
|
QRectF,
|
||||||
|
# QRect,
|
||||||
QPointF,
|
QPointF,
|
||||||
)
|
)
|
||||||
|
from PyQt5.QtGui import (
|
||||||
|
QPainter,
|
||||||
|
QPainterPath,
|
||||||
|
)
|
||||||
from .._profile import pg_profile_enabled, ms_slower_then
|
from .._profile import pg_profile_enabled, ms_slower_then
|
||||||
from ._style import hcolor
|
from ._style import hcolor
|
||||||
# from ._compression import (
|
# from ._compression import (
|
||||||
# # ohlc_to_m4_line,
|
# # ohlc_to_m4_line,
|
||||||
# ds_m4,
|
# ds_m4,
|
||||||
# )
|
# )
|
||||||
from ._pathops import xy_downsample
|
|
||||||
from ..log import get_logger
|
from ..log import get_logger
|
||||||
|
|
||||||
|
|
||||||
|
@ -53,24 +57,43 @@ _line_styles: dict[str, int] = {
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
||||||
class FastAppendCurve(pg.GraphicsObject):
|
class Curve(pg.GraphicsObject):
|
||||||
'''
|
'''
|
||||||
A faster, append friendly version of ``pyqtgraph.PlotCurveItem``
|
A faster, simpler, append friendly version of
|
||||||
built for real-time data updates.
|
``pyqtgraph.PlotCurveItem`` built for highly customizable real-time
|
||||||
|
updates.
|
||||||
|
|
||||||
The main difference is avoiding regeneration of the entire
|
This type is a much stripped down version of a ``pyqtgraph`` style
|
||||||
historical path where possible and instead only updating the "new"
|
"graphics object" in the sense that the internal lower level
|
||||||
segment(s) via a ``numpy`` array diff calc. Further the "last"
|
graphics which are drawn in the ``.paint()`` method are actually
|
||||||
graphic segment is drawn independently such that near-term (high
|
rendered outside of this class entirely and instead are assigned as
|
||||||
frequency) discrete-time-sampled style updates don't trigger a full
|
state (instance vars) here and then drawn during a Qt graphics
|
||||||
path redraw.
|
cycle.
|
||||||
|
|
||||||
|
The main motivation for this more modular, composed design is that
|
||||||
|
lower level graphics data can be rendered in different threads and
|
||||||
|
then read and drawn in this main thread without having to worry
|
||||||
|
about dealing with Qt's concurrency primitives. See
|
||||||
|
``piker.ui._flows.Renderer`` for details and logic related to lower
|
||||||
|
level path generation and incremental update. The main differences in
|
||||||
|
the path generation code include:
|
||||||
|
|
||||||
|
- avoiding regeneration of the entire historical path where possible
|
||||||
|
and instead only updating the "new" segment(s) via a ``numpy``
|
||||||
|
array diff calc.
|
||||||
|
- here, the "last" graphics datum-segment is drawn independently
|
||||||
|
such that near-term (high frequency) discrete-time-sampled style
|
||||||
|
updates don't trigger a full path redraw.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
|
|
||||||
|
# sub-type customization methods
|
||||||
|
sub_br: Optional[Callable] = None
|
||||||
|
sub_paint: Optional[Callable] = None
|
||||||
|
declare_paintables: Optional[Callable] = None
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
|
|
||||||
x: np.ndarray = None,
|
|
||||||
y: np.ndarray = None,
|
|
||||||
*args,
|
*args,
|
||||||
|
|
||||||
step_mode: bool = False,
|
step_mode: bool = False,
|
||||||
|
@ -84,31 +107,25 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
|
|
||||||
) -> None:
|
) -> None:
|
||||||
|
|
||||||
# brutaaalll, see comments within..
|
|
||||||
self._y = self.yData = y
|
|
||||||
self._x = self.xData = x
|
|
||||||
self._vr: Optional[tuple] = None
|
|
||||||
self._avr: Optional[tuple] = None
|
|
||||||
self._br = None
|
|
||||||
|
|
||||||
self._name = name
|
self._name = name
|
||||||
self.path: Optional[QtGui.QPainterPath] = None
|
|
||||||
|
|
||||||
|
# brutaaalll, see comments within..
|
||||||
|
self.yData = None
|
||||||
|
self.xData = None
|
||||||
|
|
||||||
|
# self._last_cap: int = 0
|
||||||
|
self.path: Optional[QPainterPath] = None
|
||||||
|
|
||||||
|
# additional path used for appends which tries to avoid
|
||||||
|
# triggering an update/redraw of the presumably larger
|
||||||
|
# historical ``.path`` above.
|
||||||
self.use_fpath = use_fpath
|
self.use_fpath = use_fpath
|
||||||
self.fast_path: Optional[QtGui.QPainterPath] = None
|
self.fast_path: Optional[QPainterPath] = None
|
||||||
|
|
||||||
# TODO: we can probably just dispense with the parent since
|
# TODO: we can probably just dispense with the parent since
|
||||||
# we're basically only using the pen setting now...
|
# we're basically only using the pen setting now...
|
||||||
super().__init__(*args, **kwargs)
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
# self._xrange: tuple[int, int] = self.dataBounds(ax=0)
|
|
||||||
self._xrange: Optional[tuple[int, int]] = None
|
|
||||||
# self._x_iv_range = None
|
|
||||||
|
|
||||||
# self._last_draw = time.time()
|
|
||||||
self._in_ds: bool = False
|
|
||||||
self._last_uppx: float = 0
|
|
||||||
|
|
||||||
# all history of curve is drawn in single px thickness
|
# all history of curve is drawn in single px thickness
|
||||||
pen = pg.mkPen(hcolor(color))
|
pen = pg.mkPen(hcolor(color))
|
||||||
pen.setStyle(_line_styles[style])
|
pen.setStyle(_line_styles[style])
|
||||||
|
@ -122,26 +139,44 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
# self.last_step_pen = pg.mkPen(hcolor(color), width=2)
|
# self.last_step_pen = pg.mkPen(hcolor(color), width=2)
|
||||||
self.last_step_pen = pg.mkPen(pen, width=2)
|
self.last_step_pen = pg.mkPen(pen, width=2)
|
||||||
|
|
||||||
self._last_line: Optional[QLineF] = None
|
# self._last_line: Optional[QLineF] = None
|
||||||
self._last_step_rect: Optional[QRectF] = None
|
self._last_line = QLineF()
|
||||||
|
self._last_w: float = 1
|
||||||
|
|
||||||
# flat-top style histogram-like discrete curve
|
# flat-top style histogram-like discrete curve
|
||||||
self._step_mode: bool = step_mode
|
# self._step_mode: bool = step_mode
|
||||||
|
|
||||||
# self._fill = True
|
# self._fill = True
|
||||||
self._brush = pg.functions.mkBrush(hcolor(fill_color or color))
|
self._brush = pg.functions.mkBrush(hcolor(fill_color or color))
|
||||||
|
|
||||||
|
# NOTE: this setting seems to mostly prevent redraws on mouse
|
||||||
|
# interaction which is a huge boon for avg interaction latency.
|
||||||
|
|
||||||
# TODO: one question still remaining is if this makes trasform
|
# TODO: one question still remaining is if this makes trasform
|
||||||
# interactions slower (such as zooming) and if so maybe if/when
|
# interactions slower (such as zooming) and if so maybe if/when
|
||||||
# we implement a "history" mode for the view we disable this in
|
# we implement a "history" mode for the view we disable this in
|
||||||
# that mode?
|
# that mode?
|
||||||
# if step_mode:
|
|
||||||
# don't enable caching by default for the case where the
|
# don't enable caching by default for the case where the
|
||||||
# only thing drawn is the "last" line segment which can
|
# only thing drawn is the "last" line segment which can
|
||||||
# have a weird artifact where it won't be fully drawn to its
|
# have a weird artifact where it won't be fully drawn to its
|
||||||
# endpoint (something we saw on trade rate curves)
|
# endpoint (something we saw on trade rate curves)
|
||||||
self.setCacheMode(QGraphicsItem.DeviceCoordinateCache)
|
self.setCacheMode(QGraphicsItem.DeviceCoordinateCache)
|
||||||
|
|
||||||
|
# XXX: see explanation for different caching modes:
|
||||||
|
# https://stackoverflow.com/a/39410081
|
||||||
|
# seems to only be useful if we don't re-generate the entire
|
||||||
|
# QPainterPath every time
|
||||||
|
# curve.setCacheMode(QtWidgets.QGraphicsItem.DeviceCoordinateCache)
|
||||||
|
|
||||||
|
# don't ever use this - it's a colossal nightmare of artefacts
|
||||||
|
# and is disastrous for performance.
|
||||||
|
# curve.setCacheMode(QtWidgets.QGraphicsItem.ItemCoordinateCache)
|
||||||
|
|
||||||
|
# allow sub-type customization
|
||||||
|
declare = self.declare_paintables
|
||||||
|
if declare:
|
||||||
|
declare()
|
||||||
|
|
||||||
# TODO: probably stick this in a new parent
|
# TODO: probably stick this in a new parent
|
||||||
# type which will contain our own version of
|
# type which will contain our own version of
|
||||||
# what ``PlotCurveItem`` had in terms of base
|
# what ``PlotCurveItem`` had in terms of base
|
||||||
|
@ -164,9 +199,6 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
vr = self.viewRect()
|
vr = self.viewRect()
|
||||||
l, r = int(vr.left()), int(vr.right())
|
l, r = int(vr.left()), int(vr.right())
|
||||||
|
|
||||||
if not self._xrange:
|
|
||||||
return 0
|
|
||||||
|
|
||||||
start, stop = self._xrange
|
start, stop = self._xrange
|
||||||
lbar = max(l, start)
|
lbar = max(l, start)
|
||||||
rbar = min(r, stop)
|
rbar = min(r, stop)
|
||||||
|
@ -175,381 +207,10 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
QLineF(lbar, 0, rbar, 0)
|
QLineF(lbar, 0, rbar, 0)
|
||||||
).length()
|
).length()
|
||||||
|
|
||||||
def update_from_array(
|
|
||||||
self,
|
|
||||||
|
|
||||||
# full array input history
|
|
||||||
x: np.ndarray,
|
|
||||||
y: np.ndarray,
|
|
||||||
|
|
||||||
# pre-sliced array data that's "in view"
|
|
||||||
x_iv: np.ndarray,
|
|
||||||
y_iv: np.ndarray,
|
|
||||||
|
|
||||||
view_range: Optional[tuple[int, int]] = None,
|
|
||||||
profiler: Optional[pg.debug.Profiler] = None,
|
|
||||||
draw_last: bool = True,
|
|
||||||
slice_to_head: int = -1,
|
|
||||||
do_append: bool = True,
|
|
||||||
should_redraw: bool = False,
|
|
||||||
|
|
||||||
) -> QtGui.QPainterPath:
|
|
||||||
'''
|
|
||||||
Update curve from input 2-d data.
|
|
||||||
|
|
||||||
Compare with a cached "x-range" state and (pre/a)ppend based on
|
|
||||||
a length diff.
|
|
||||||
|
|
||||||
'''
|
|
||||||
profiler = profiler or pg.debug.Profiler(
|
|
||||||
msg=f'FastAppendCurve.update_from_array(): `{self._name}`',
|
|
||||||
disabled=not pg_profile_enabled(),
|
|
||||||
ms_threshold=ms_slower_then,
|
|
||||||
)
|
|
||||||
flip_cache = False
|
|
||||||
|
|
||||||
if self._xrange:
|
|
||||||
istart, istop = self._xrange
|
|
||||||
else:
|
|
||||||
self._xrange = istart, istop = x[0], x[-1]
|
|
||||||
|
|
||||||
# compute the length diffs between the first/last index entry in
|
|
||||||
# the input data and the last indexes we have on record from the
|
|
||||||
# last time we updated the curve index.
|
|
||||||
prepend_length = int(istart - x[0])
|
|
||||||
append_length = int(x[-1] - istop)
|
|
||||||
|
|
||||||
# this is the diff-mode, "data"-rendered index
|
|
||||||
# tracking var..
|
|
||||||
self._xrange = x[0], x[-1]
|
|
||||||
|
|
||||||
# print(f"xrange: {self._xrange}")
|
|
||||||
|
|
||||||
# XXX: lol brutal, the internals of `CurvePoint` (inherited by
|
|
||||||
# our `LineDot`) required ``.getData()`` to work..
|
|
||||||
# self.xData = x
|
|
||||||
# self.yData = y
|
|
||||||
# self._x, self._y = x, y
|
|
||||||
|
|
||||||
# downsampling incremental state checking
|
|
||||||
uppx = self.x_uppx()
|
|
||||||
px_width = self.px_width()
|
|
||||||
uppx_diff = (uppx - self._last_uppx)
|
|
||||||
|
|
||||||
new_sample_rate = False
|
|
||||||
should_ds = self._in_ds
|
|
||||||
showing_src_data = self._in_ds
|
|
||||||
# should_redraw = False
|
|
||||||
|
|
||||||
# by default we only pull data up to the last (current) index
|
|
||||||
x_out_full = x_out = x[:slice_to_head]
|
|
||||||
y_out_full = y_out = y[:slice_to_head]
|
|
||||||
|
|
||||||
# if a view range is passed, plan to draw the
|
|
||||||
# source ouput that's "in view" of the chart.
|
|
||||||
if (
|
|
||||||
view_range
|
|
||||||
# and not self._in_ds
|
|
||||||
# and not prepend_length > 0
|
|
||||||
):
|
|
||||||
# print(f'{self._name} vr: {view_range}')
|
|
||||||
|
|
||||||
# by default we only pull data up to the last (current) index
|
|
||||||
x_out, y_out = x_iv[:slice_to_head], y_iv[:slice_to_head]
|
|
||||||
profiler(f'view range slice {view_range}')
|
|
||||||
|
|
||||||
vl, vr = view_range
|
|
||||||
|
|
||||||
# last_ivr = self._x_iv_range
|
|
||||||
# ix_iv, iy_iv = self._x_iv_range = (x_iv[0], x_iv[-1])
|
|
||||||
|
|
||||||
zoom_or_append = False
|
|
||||||
last_vr = self._vr
|
|
||||||
last_ivr = self._avr
|
|
||||||
|
|
||||||
if last_vr:
|
|
||||||
# relative slice indices
|
|
||||||
lvl, lvr = last_vr
|
|
||||||
# abs slice indices
|
|
||||||
al, ar = last_ivr
|
|
||||||
|
|
||||||
# append_length = int(x[-1] - istop)
|
|
||||||
# append_length = int(x_iv[-1] - ar)
|
|
||||||
|
|
||||||
# left_change = abs(x_iv[0] - al) >= 1
|
|
||||||
# right_change = abs(x_iv[-1] - ar) >= 1
|
|
||||||
|
|
||||||
if (
|
|
||||||
# likely a zoom view change
|
|
||||||
(vr - lvr) > 2 or vl < lvl
|
|
||||||
# append / prepend update
|
|
||||||
# we had an append update where the view range
|
|
||||||
# didn't change but the data-viewed (shifted)
|
|
||||||
# underneath, so we need to redraw.
|
|
||||||
# or left_change and right_change and last_vr == view_range
|
|
||||||
|
|
||||||
# not (left_change and right_change) and ivr
|
|
||||||
# (
|
|
||||||
# or abs(x_iv[ivr] - livr) > 1
|
|
||||||
):
|
|
||||||
zoom_or_append = True
|
|
||||||
|
|
||||||
# if last_ivr:
|
|
||||||
# liivl, liivr = last_ivr
|
|
||||||
|
|
||||||
if (
|
|
||||||
view_range != last_vr
|
|
||||||
and (
|
|
||||||
append_length > 1
|
|
||||||
or zoom_or_append
|
|
||||||
)
|
|
||||||
):
|
|
||||||
should_redraw = True
|
|
||||||
# print("REDRAWING BRUH")
|
|
||||||
|
|
||||||
self._vr = view_range
|
|
||||||
self._avr = x_iv[0], x_iv[slice_to_head]
|
|
||||||
|
|
||||||
# x_last = x_iv[-1]
|
|
||||||
# y_last = y_iv[-1]
|
|
||||||
# self._last_vr = view_range
|
|
||||||
|
|
||||||
# self.disable_cache()
|
|
||||||
# flip_cache = True
|
|
||||||
|
|
||||||
if prepend_length > 0:
|
|
||||||
should_redraw = True
|
|
||||||
|
|
||||||
# check for downsampling conditions
|
|
||||||
if (
|
|
||||||
# std m4 downsample conditions
|
|
||||||
px_width
|
|
||||||
and abs(uppx_diff) >= 1
|
|
||||||
):
|
|
||||||
log.info(
|
|
||||||
f'{self._name} sampler change: {self._last_uppx} -> {uppx}'
|
|
||||||
)
|
|
||||||
self._last_uppx = uppx
|
|
||||||
new_sample_rate = True
|
|
||||||
showing_src_data = False
|
|
||||||
should_redraw = True
|
|
||||||
should_ds = True
|
|
||||||
|
|
||||||
elif (
|
|
||||||
uppx <= 2
|
|
||||||
and self._in_ds
|
|
||||||
):
|
|
||||||
# we should de-downsample back to our original
|
|
||||||
# source data so we clear our path data in prep
|
|
||||||
# to generate a new one from original source data.
|
|
||||||
should_redraw = True
|
|
||||||
new_sample_rate = True
|
|
||||||
should_ds = False
|
|
||||||
showing_src_data = True
|
|
||||||
|
|
||||||
# no_path_yet = self.path is None
|
|
||||||
if (
|
|
||||||
self.path is None
|
|
||||||
or should_redraw
|
|
||||||
or new_sample_rate
|
|
||||||
or prepend_length > 0
|
|
||||||
):
|
|
||||||
if should_redraw:
|
|
||||||
if self.path:
|
|
||||||
self.path.clear()
|
|
||||||
profiler('cleared paths due to `should_redraw=True`')
|
|
||||||
|
|
||||||
if self.fast_path:
|
|
||||||
self.fast_path.clear()
|
|
||||||
|
|
||||||
profiler('cleared paths due to `should_redraw` set')
|
|
||||||
|
|
||||||
if new_sample_rate and showing_src_data:
|
|
||||||
# if self._in_ds:
|
|
||||||
log.info(f'DEDOWN -> {self._name}')
|
|
||||||
|
|
||||||
self._in_ds = False
|
|
||||||
|
|
||||||
elif should_ds and uppx and px_width > 1:
|
|
||||||
|
|
||||||
x_out, y_out = xy_downsample(
|
|
||||||
x_out,
|
|
||||||
y_out,
|
|
||||||
px_width,
|
|
||||||
uppx,
|
|
||||||
)
|
|
||||||
profiler(f'FULL PATH downsample redraw={should_ds}')
|
|
||||||
self._in_ds = True
|
|
||||||
|
|
||||||
self.path = pg.functions.arrayToQPath(
|
|
||||||
x_out,
|
|
||||||
y_out,
|
|
||||||
connect='all',
|
|
||||||
finiteCheck=False,
|
|
||||||
path=self.path,
|
|
||||||
)
|
|
||||||
self.prepareGeometryChange()
|
|
||||||
profiler(
|
|
||||||
f'generated fresh path. (should_redraw: {should_redraw} should_ds: {should_ds} new_sample_rate: {new_sample_rate})'
|
|
||||||
)
|
|
||||||
# profiler(f'DRAW PATH IN VIEW -> {self._name}')
|
|
||||||
|
|
||||||
# reserve mem allocs see:
|
|
||||||
# - https://doc.qt.io/qt-5/qpainterpath.html#reserve
|
|
||||||
# - https://doc.qt.io/qt-5/qpainterpath.html#capacity
|
|
||||||
# - https://doc.qt.io/qt-5/qpainterpath.html#clear
|
|
||||||
# XXX: right now this is based on had hoc checks on a
|
|
||||||
# hidpi 3840x2160 4k monitor but we should optimize for
|
|
||||||
# the target display(s) on the sys.
|
|
||||||
# if no_path_yet:
|
|
||||||
# self.path.reserve(int(500e3))
|
|
||||||
|
|
||||||
# TODO: get this piecewise prepend working - right now it's
|
|
||||||
# giving heck on vwap...
|
|
||||||
# elif prepend_length:
|
|
||||||
# breakpoint()
|
|
||||||
|
|
||||||
# prepend_path = pg.functions.arrayToQPath(
|
|
||||||
# x[0:prepend_length],
|
|
||||||
# y[0:prepend_length],
|
|
||||||
# connect='all'
|
|
||||||
# )
|
|
||||||
|
|
||||||
# # swap prepend path in "front"
|
|
||||||
# old_path = self.path
|
|
||||||
# self.path = prepend_path
|
|
||||||
# # self.path.moveTo(new_x[0], new_y[0])
|
|
||||||
# self.path.connectPath(old_path)
|
|
||||||
|
|
||||||
elif (
|
|
||||||
append_length > 0
|
|
||||||
and do_append
|
|
||||||
and not should_redraw
|
|
||||||
# and not view_range
|
|
||||||
):
|
|
||||||
print(f'{self._name} append len: {append_length}')
|
|
||||||
new_x = x[-append_length - 2:slice_to_head]
|
|
||||||
new_y = y[-append_length - 2:slice_to_head]
|
|
||||||
profiler('sliced append path')
|
|
||||||
|
|
||||||
profiler(
|
|
||||||
f'diffed array input, append_length={append_length}'
|
|
||||||
)
|
|
||||||
|
|
||||||
# if should_ds:
|
|
||||||
# new_x, new_y = xy_downsample(
|
|
||||||
# new_x,
|
|
||||||
# new_y,
|
|
||||||
# px_width,
|
|
||||||
# uppx,
|
|
||||||
# )
|
|
||||||
# profiler(f'fast path downsample redraw={should_ds}')
|
|
||||||
|
|
||||||
append_path = pg.functions.arrayToQPath(
|
|
||||||
new_x,
|
|
||||||
new_y,
|
|
||||||
connect='all',
|
|
||||||
finiteCheck=False,
|
|
||||||
path=self.fast_path,
|
|
||||||
)
|
|
||||||
profiler('generated append qpath')
|
|
||||||
|
|
||||||
if self.use_fpath:
|
|
||||||
# an attempt at trying to make append-updates faster..
|
|
||||||
if self.fast_path is None:
|
|
||||||
self.fast_path = append_path
|
|
||||||
# self.fast_path.reserve(int(6e3))
|
|
||||||
else:
|
|
||||||
self.fast_path.connectPath(append_path)
|
|
||||||
size = self.fast_path.capacity()
|
|
||||||
profiler(f'connected fast path w size: {size}')
|
|
||||||
|
|
||||||
# print(f"append_path br: {append_path.boundingRect()}")
|
|
||||||
# self.path.moveTo(new_x[0], new_y[0])
|
|
||||||
# path.connectPath(append_path)
|
|
||||||
|
|
||||||
# XXX: lol this causes a hang..
|
|
||||||
# self.path = self.path.simplified()
|
|
||||||
else:
|
|
||||||
size = self.path.capacity()
|
|
||||||
profiler(f'connected history path w size: {size}')
|
|
||||||
self.path.connectPath(append_path)
|
|
||||||
|
|
||||||
# other merging ideas:
|
|
||||||
# https://stackoverflow.com/questions/8936225/how-to-merge-qpainterpaths
|
|
||||||
# path.addPath(append_path)
|
|
||||||
# path.closeSubpath()
|
|
||||||
|
|
||||||
# TODO: try out new work from `pyqtgraph` main which
|
|
||||||
# should repair horrid perf:
|
|
||||||
# https://github.com/pyqtgraph/pyqtgraph/pull/2032
|
|
||||||
# ok, nope still horrible XD
|
|
||||||
# if self._fill:
|
|
||||||
# # XXX: super slow set "union" op
|
|
||||||
# self.path = self.path.united(append_path).simplified()
|
|
||||||
|
|
||||||
# self.disable_cache()
|
|
||||||
# flip_cache = True
|
|
||||||
|
|
||||||
if draw_last:
|
|
||||||
self.draw_last(x, y)
|
|
||||||
profiler('draw last segment')
|
|
||||||
|
|
||||||
|
|
||||||
# if flip_cache:
|
|
||||||
# # # XXX: seems to be needed to avoid artifacts (see above).
|
|
||||||
# self.setCacheMode(QGraphicsItem.DeviceCoordinateCache)
|
|
||||||
|
|
||||||
# trigger redraw of path
|
|
||||||
# do update before reverting to cache mode
|
|
||||||
self.update()
|
|
||||||
profiler('.update()')
|
|
||||||
|
|
||||||
def draw_last(
|
|
||||||
self,
|
|
||||||
x: np.ndarray,
|
|
||||||
y: np.ndarray,
|
|
||||||
|
|
||||||
) -> None:
|
|
||||||
x_last = x[-1]
|
|
||||||
y_last = y[-1]
|
|
||||||
|
|
||||||
# draw the "current" step graphic segment so it lines up with
|
|
||||||
# the "middle" of the current (OHLC) sample.
|
|
||||||
if self._step_mode:
|
|
||||||
self._last_line = QLineF(
|
|
||||||
x_last - 0.5, 0,
|
|
||||||
x_last + 0.5, 0,
|
|
||||||
# x_last, 0,
|
|
||||||
# x_last, 0,
|
|
||||||
)
|
|
||||||
self._last_step_rect = QRectF(
|
|
||||||
x_last - 0.5, 0,
|
|
||||||
x_last + 0.5, y_last
|
|
||||||
# x_last, 0,
|
|
||||||
# x_last, y_last
|
|
||||||
)
|
|
||||||
# print(
|
|
||||||
# f"path br: {self.path.boundingRect()}",
|
|
||||||
# f"fast path br: {self.fast_path.boundingRect()}",
|
|
||||||
# f"last rect br: {self._last_step_rect}",
|
|
||||||
# )
|
|
||||||
else:
|
|
||||||
self._last_line = QLineF(
|
|
||||||
x[-2], y[-2],
|
|
||||||
x_last, y_last
|
|
||||||
)
|
|
||||||
|
|
||||||
self.update()
|
|
||||||
|
|
||||||
# XXX: lol brutal, the internals of `CurvePoint` (inherited by
|
# XXX: lol brutal, the internals of `CurvePoint` (inherited by
|
||||||
# our `LineDot`) required ``.getData()`` to work..
|
# our `LineDot`) required ``.getData()`` to work..
|
||||||
def getData(self):
|
def getData(self):
|
||||||
return self._x, self._y
|
return self.xData, self.yData
|
||||||
|
|
||||||
# TODO: drop the above after ``Cursor`` re-work
|
|
||||||
def get_arrays(self) -> tuple[np.ndarray, np.ndarray]:
|
|
||||||
return self._x, self._y
|
|
||||||
|
|
||||||
def clear(self):
|
def clear(self):
|
||||||
'''
|
'''
|
||||||
|
@ -572,29 +233,18 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
# self.fast_path.clear()
|
# self.fast_path.clear()
|
||||||
self.fast_path = None
|
self.fast_path = None
|
||||||
|
|
||||||
# self.disable_cache()
|
@cm
|
||||||
# self.setCacheMode(QGraphicsItem.DeviceCoordinateCache)
|
|
||||||
|
|
||||||
def reset_cache(self) -> None:
|
def reset_cache(self) -> None:
|
||||||
self.disable_cache()
|
|
||||||
self.setCacheMode(QGraphicsItem.DeviceCoordinateCache)
|
|
||||||
|
|
||||||
def disable_cache(self) -> None:
|
|
||||||
'''
|
|
||||||
Disable the use of the pixel coordinate cache and trigger a geo event.
|
|
||||||
|
|
||||||
'''
|
|
||||||
# XXX: pretty annoying but, without this there's little
|
|
||||||
# artefacts on the append updates to the curve...
|
|
||||||
self.setCacheMode(QtWidgets.QGraphicsItem.NoCache)
|
self.setCacheMode(QtWidgets.QGraphicsItem.NoCache)
|
||||||
# self.prepareGeometryChange()
|
yield
|
||||||
|
self.setCacheMode(QGraphicsItem.DeviceCoordinateCache)
|
||||||
|
|
||||||
def boundingRect(self):
|
def boundingRect(self):
|
||||||
'''
|
'''
|
||||||
Compute and then cache our rect.
|
Compute and then cache our rect.
|
||||||
'''
|
'''
|
||||||
if self.path is None:
|
if self.path is None:
|
||||||
return QtGui.QPainterPath().boundingRect()
|
return QPainterPath().boundingRect()
|
||||||
else:
|
else:
|
||||||
# dynamically override this method after initial
|
# dynamically override this method after initial
|
||||||
# path is created to avoid requiring the above None check
|
# path is created to avoid requiring the above None check
|
||||||
|
@ -606,14 +256,15 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
Post init ``.boundingRect()```.
|
Post init ``.boundingRect()```.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
hb = self.path.controlPointRect()
|
|
||||||
# hb = self.path.boundingRect()
|
# hb = self.path.boundingRect()
|
||||||
|
hb = self.path.controlPointRect()
|
||||||
hb_size = hb.size()
|
hb_size = hb.size()
|
||||||
|
|
||||||
fp = self.fast_path
|
fp = self.fast_path
|
||||||
if fp:
|
if fp:
|
||||||
fhb = fp.controlPointRect()
|
fhb = fp.controlPointRect()
|
||||||
hb_size = fhb.size() + hb_size
|
hb_size = fhb.size() + hb_size
|
||||||
|
|
||||||
# print(f'hb_size: {hb_size}')
|
# print(f'hb_size: {hb_size}')
|
||||||
|
|
||||||
# if self._last_step_rect:
|
# if self._last_step_rect:
|
||||||
|
@ -629,13 +280,26 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
# # hb_size.height() + 1
|
# # hb_size.height() + 1
|
||||||
# )
|
# )
|
||||||
|
|
||||||
# if self._last_step_rect:
|
|
||||||
# br = self._last_step_rect.bottomRight()
|
# br = self._last_step_rect.bottomRight()
|
||||||
|
|
||||||
# else:
|
w = hb_size.width()
|
||||||
# hb_size += QSizeF(1, 1)
|
h = hb_size.height()
|
||||||
w = hb_size.width() + 1
|
|
||||||
h = hb_size.height() + 1
|
sbr = self.sub_br
|
||||||
|
if sbr:
|
||||||
|
w, h = self.sub_br(w, h)
|
||||||
|
else:
|
||||||
|
# assume plain line graphic and use
|
||||||
|
# default unit step in each direction.
|
||||||
|
|
||||||
|
# only on a plane line do we include
|
||||||
|
# and extra index step's worth of width
|
||||||
|
# since in the step case the end of the curve
|
||||||
|
# actually terminates earlier so we don't need
|
||||||
|
# this for the last step.
|
||||||
|
w += self._last_w
|
||||||
|
# ll = self._last_line
|
||||||
|
h += 1 # ll.y2() - ll.y1()
|
||||||
|
|
||||||
# br = QPointF(
|
# br = QPointF(
|
||||||
# self._vr[-1],
|
# self._vr[-1],
|
||||||
|
@ -656,43 +320,37 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
# hb_size,
|
# hb_size,
|
||||||
QSizeF(w, h)
|
QSizeF(w, h)
|
||||||
)
|
)
|
||||||
self._br = br
|
|
||||||
# print(f'bounding rect: {br}')
|
# print(f'bounding rect: {br}')
|
||||||
return br
|
return br
|
||||||
|
|
||||||
def paint(
|
def paint(
|
||||||
self,
|
self,
|
||||||
p: QtGui.QPainter,
|
p: QPainter,
|
||||||
opt: QtWidgets.QStyleOptionGraphicsItem,
|
opt: QtWidgets.QStyleOptionGraphicsItem,
|
||||||
w: QtWidgets.QWidget
|
w: QtWidgets.QWidget
|
||||||
|
|
||||||
) -> None:
|
) -> None:
|
||||||
|
|
||||||
profiler = pg.debug.Profiler(
|
profiler = pg.debug.Profiler(
|
||||||
msg=f'FastAppendCurve.paint(): `{self._name}`',
|
msg=f'Curve.paint(): `{self._name}`',
|
||||||
disabled=not pg_profile_enabled(),
|
disabled=not pg_profile_enabled(),
|
||||||
ms_threshold=ms_slower_then,
|
ms_threshold=ms_slower_then,
|
||||||
)
|
)
|
||||||
self.prepareGeometryChange()
|
|
||||||
|
|
||||||
if (
|
sub_paint = self.sub_paint
|
||||||
self._step_mode
|
if sub_paint:
|
||||||
and self._last_step_rect
|
sub_paint(p, profiler)
|
||||||
):
|
|
||||||
brush = self._brush
|
|
||||||
|
|
||||||
# p.drawLines(*tuple(filter(bool, self._last_step_lines)))
|
|
||||||
# p.drawRect(self._last_step_rect)
|
|
||||||
p.fillRect(self._last_step_rect, brush)
|
|
||||||
profiler('.fillRect()')
|
|
||||||
|
|
||||||
if self._last_line:
|
|
||||||
p.setPen(self.last_step_pen)
|
p.setPen(self.last_step_pen)
|
||||||
p.drawLine(self._last_line)
|
p.drawLine(self._last_line)
|
||||||
profiler('.drawLine()')
|
profiler('.drawLine()')
|
||||||
p.setPen(self._pen)
|
p.setPen(self._pen)
|
||||||
|
|
||||||
path = self.path
|
path = self.path
|
||||||
|
# cap = path.capacity()
|
||||||
|
# if cap != self._last_cap:
|
||||||
|
# print(f'NEW CAPACITY: {self._last_cap} -> {cap}')
|
||||||
|
# self._last_cap = cap
|
||||||
|
|
||||||
if path:
|
if path:
|
||||||
p.drawPath(path)
|
p.drawPath(path)
|
||||||
|
@ -710,3 +368,117 @@ class FastAppendCurve(pg.GraphicsObject):
|
||||||
# if self._fill:
|
# if self._fill:
|
||||||
# brush = self.opts['brush']
|
# brush = self.opts['brush']
|
||||||
# p.fillPath(self.path, brush)
|
# p.fillPath(self.path, brush)
|
||||||
|
|
||||||
|
def draw_last_datum(
|
||||||
|
self,
|
||||||
|
path: QPainterPath,
|
||||||
|
src_data: np.ndarray,
|
||||||
|
render_data: np.ndarray,
|
||||||
|
reset: bool,
|
||||||
|
array_key: str,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
# default line draw last call
|
||||||
|
# with self.reset_cache():
|
||||||
|
x = render_data['index']
|
||||||
|
y = render_data[array_key]
|
||||||
|
|
||||||
|
# draw the "current" step graphic segment so it
|
||||||
|
# lines up with the "middle" of the current
|
||||||
|
# (OHLC) sample.
|
||||||
|
self._last_line = QLineF(
|
||||||
|
x[-2], y[-2],
|
||||||
|
x[-1], y[-1],
|
||||||
|
)
|
||||||
|
|
||||||
|
return x, y
|
||||||
|
|
||||||
|
|
||||||
|
# TODO: this should probably be a "downsampled" curve type
|
||||||
|
# that draws a bar-style (but for the px column) last graphics
|
||||||
|
# element such that the current datum in view can be shown
|
||||||
|
# (via it's max / min) even when highly zoomed out.
|
||||||
|
class FlattenedOHLC(Curve):
|
||||||
|
|
||||||
|
def draw_last_datum(
|
||||||
|
self,
|
||||||
|
path: QPainterPath,
|
||||||
|
src_data: np.ndarray,
|
||||||
|
render_data: np.ndarray,
|
||||||
|
reset: bool,
|
||||||
|
array_key: str,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
lasts = src_data[-2:]
|
||||||
|
x = lasts['index']
|
||||||
|
y = lasts['close']
|
||||||
|
|
||||||
|
# draw the "current" step graphic segment so it
|
||||||
|
# lines up with the "middle" of the current
|
||||||
|
# (OHLC) sample.
|
||||||
|
self._last_line = QLineF(
|
||||||
|
x[-2], y[-2],
|
||||||
|
x[-1], y[-1]
|
||||||
|
)
|
||||||
|
return x, y
|
||||||
|
|
||||||
|
|
||||||
|
class StepCurve(Curve):
|
||||||
|
|
||||||
|
def declare_paintables(
|
||||||
|
self,
|
||||||
|
) -> None:
|
||||||
|
self._last_step_rect = QRectF()
|
||||||
|
|
||||||
|
def draw_last_datum(
|
||||||
|
self,
|
||||||
|
path: QPainterPath,
|
||||||
|
src_data: np.ndarray,
|
||||||
|
render_data: np.ndarray,
|
||||||
|
reset: bool,
|
||||||
|
array_key: str,
|
||||||
|
|
||||||
|
w: float = 0.5,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
|
||||||
|
# TODO: remove this and instead place all step curve
|
||||||
|
# updating into pre-path data render callbacks.
|
||||||
|
# full input data
|
||||||
|
x = src_data['index']
|
||||||
|
y = src_data[array_key]
|
||||||
|
|
||||||
|
x_last = x[-1]
|
||||||
|
y_last = y[-1]
|
||||||
|
|
||||||
|
# lol, commenting this makes step curves
|
||||||
|
# all "black" for me :eyeroll:..
|
||||||
|
self._last_line = QLineF(
|
||||||
|
x_last - w, 0,
|
||||||
|
x_last + w, 0,
|
||||||
|
)
|
||||||
|
self._last_step_rect = QRectF(
|
||||||
|
x_last - w, 0,
|
||||||
|
x_last + w, y_last,
|
||||||
|
)
|
||||||
|
return x, y
|
||||||
|
|
||||||
|
def sub_paint(
|
||||||
|
self,
|
||||||
|
p: QPainter,
|
||||||
|
profiler: pg.debug.Profiler,
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
# p.drawLines(*tuple(filter(bool, self._last_step_lines)))
|
||||||
|
# p.drawRect(self._last_step_rect)
|
||||||
|
p.fillRect(self._last_step_rect, self._brush)
|
||||||
|
profiler('.fillRect()')
|
||||||
|
|
||||||
|
def sub_br(
|
||||||
|
self,
|
||||||
|
path_w: float,
|
||||||
|
path_h: float,
|
||||||
|
|
||||||
|
) -> (float, float):
|
||||||
|
# passthrough
|
||||||
|
return path_w, path_h
|
||||||
|
|
|
@ -32,7 +32,7 @@ import trio
|
||||||
import pendulum
|
import pendulum
|
||||||
import pyqtgraph as pg
|
import pyqtgraph as pg
|
||||||
|
|
||||||
from .. import brokers
|
# from .. import brokers
|
||||||
from ..data.feed import open_feed
|
from ..data.feed import open_feed
|
||||||
from ._axes import YAxisLabel
|
from ._axes import YAxisLabel
|
||||||
from ._chart import (
|
from ._chart import (
|
||||||
|
@ -263,6 +263,7 @@ async def graphics_update_loop(
|
||||||
'vars': {
|
'vars': {
|
||||||
'tick_margin': tick_margin,
|
'tick_margin': tick_margin,
|
||||||
'i_last': i_last,
|
'i_last': i_last,
|
||||||
|
'i_last_append': i_last,
|
||||||
'last_mx_vlm': last_mx_vlm,
|
'last_mx_vlm': last_mx_vlm,
|
||||||
'last_mx': last_mx,
|
'last_mx': last_mx,
|
||||||
'last_mn': last_mn,
|
'last_mn': last_mn,
|
||||||
|
@ -320,8 +321,8 @@ def graphics_update_cycle(
|
||||||
profiler = pg.debug.Profiler(
|
profiler = pg.debug.Profiler(
|
||||||
msg=f'Graphics loop cycle for: `{chart.name}`',
|
msg=f'Graphics loop cycle for: `{chart.name}`',
|
||||||
delayed=True,
|
delayed=True,
|
||||||
# disabled=not pg_profile_enabled(),
|
disabled=not pg_profile_enabled(),
|
||||||
disabled=True,
|
# disabled=True,
|
||||||
ms_threshold=ms_slower_then,
|
ms_threshold=ms_slower_then,
|
||||||
|
|
||||||
# ms_threshold=1/12 * 1e3,
|
# ms_threshold=1/12 * 1e3,
|
||||||
|
@ -340,7 +341,7 @@ def graphics_update_cycle(
|
||||||
for sym, quote in ds.quotes.items():
|
for sym, quote in ds.quotes.items():
|
||||||
|
|
||||||
# compute the first available graphic's x-units-per-pixel
|
# compute the first available graphic's x-units-per-pixel
|
||||||
xpx = vlm_chart.view.x_uppx()
|
uppx = vlm_chart.view.x_uppx()
|
||||||
|
|
||||||
# NOTE: vlm may be written by the ``brokerd`` backend
|
# NOTE: vlm may be written by the ``brokerd`` backend
|
||||||
# event though a tick sample is not emitted.
|
# event though a tick sample is not emitted.
|
||||||
|
@ -359,13 +360,32 @@ def graphics_update_cycle(
|
||||||
i_diff = i_step - vars['i_last']
|
i_diff = i_step - vars['i_last']
|
||||||
vars['i_last'] = i_step
|
vars['i_last'] = i_step
|
||||||
|
|
||||||
|
append_diff = i_step - vars['i_last_append']
|
||||||
|
|
||||||
|
# update the "last datum" (aka extending the flow graphic with
|
||||||
|
# new data) only if the number of unit steps is >= the number of
|
||||||
|
# such unit steps per pixel (aka uppx). Iow, if the zoom level
|
||||||
|
# is such that a datum(s) update to graphics wouldn't span
|
||||||
|
# to a new pixel, we don't update yet.
|
||||||
|
do_append = (append_diff >= uppx)
|
||||||
|
if do_append:
|
||||||
|
vars['i_last_append'] = i_step
|
||||||
|
|
||||||
|
do_rt_update = uppx < update_uppx
|
||||||
|
# print(
|
||||||
|
# f'append_diff:{append_diff}\n'
|
||||||
|
# f'uppx:{uppx}\n'
|
||||||
|
# f'do_append: {do_append}'
|
||||||
|
# )
|
||||||
|
|
||||||
|
# TODO: we should only run mxmn when we know
|
||||||
|
# an update is due via ``do_append`` above.
|
||||||
(
|
(
|
||||||
brange,
|
brange,
|
||||||
mx_in_view,
|
mx_in_view,
|
||||||
mn_in_view,
|
mn_in_view,
|
||||||
mx_vlm_in_view,
|
mx_vlm_in_view,
|
||||||
) = ds.maxmin()
|
) = ds.maxmin()
|
||||||
|
|
||||||
l, lbar, rbar, r = brange
|
l, lbar, rbar, r = brange
|
||||||
mx = mx_in_view + tick_margin
|
mx = mx_in_view + tick_margin
|
||||||
mn = mn_in_view - tick_margin
|
mn = mn_in_view - tick_margin
|
||||||
|
@ -389,8 +409,9 @@ def graphics_update_cycle(
|
||||||
# left unless we get one of the following:
|
# left unless we get one of the following:
|
||||||
if (
|
if (
|
||||||
(
|
(
|
||||||
i_diff > 0 # no new sample step
|
# i_diff > 0 # no new sample step
|
||||||
and xpx < 4 # chart is zoomed out very far
|
do_append
|
||||||
|
# and uppx < 4 # chart is zoomed out very far
|
||||||
and liv
|
and liv
|
||||||
)
|
)
|
||||||
or trigger_all
|
or trigger_all
|
||||||
|
@ -399,71 +420,11 @@ def graphics_update_cycle(
|
||||||
# pixel in a curve should show new data based on uppx
|
# pixel in a curve should show new data based on uppx
|
||||||
# and then iff update curves and shift?
|
# and then iff update curves and shift?
|
||||||
chart.increment_view(steps=i_diff)
|
chart.increment_view(steps=i_diff)
|
||||||
profiler('view incremented')
|
|
||||||
|
|
||||||
if vlm_chart:
|
if vlm_chart:
|
||||||
# always update y-label
|
vlm_chart.increment_view(steps=i_diff)
|
||||||
ds.vlm_sticky.update_from_data(
|
|
||||||
*array[-1][['index', 'volume']]
|
|
||||||
)
|
|
||||||
|
|
||||||
if (
|
profiler('view incremented')
|
||||||
(
|
|
||||||
xpx < update_uppx
|
|
||||||
or i_diff > 0
|
|
||||||
and liv
|
|
||||||
)
|
|
||||||
or trigger_all
|
|
||||||
):
|
|
||||||
# TODO: make it so this doesn't have to be called
|
|
||||||
# once the $vlm is up?
|
|
||||||
vlm_chart.update_graphics_from_flow(
|
|
||||||
'volume',
|
|
||||||
# UGGGh, see ``maxmin()`` impl in `._fsp` for
|
|
||||||
# the overlayed plotitems... we need a better
|
|
||||||
# bay to invoke a maxmin per overlay..
|
|
||||||
render=False,
|
|
||||||
# XXX: ^^^^ THIS IS SUPER IMPORTANT! ^^^^
|
|
||||||
# without this, since we disable the
|
|
||||||
# 'volume' (units) chart after the $vlm starts
|
|
||||||
# up we need to be sure to enable this
|
|
||||||
# auto-ranging otherwise there will be no handler
|
|
||||||
# connected to update accompanying overlay
|
|
||||||
# graphics..
|
|
||||||
)
|
|
||||||
profiler('`vlm_chart.update_graphics_from_flow()`')
|
|
||||||
|
|
||||||
if (
|
|
||||||
mx_vlm_in_view != vars['last_mx_vlm']
|
|
||||||
):
|
|
||||||
yrange = (0, mx_vlm_in_view * 1.375)
|
|
||||||
vlm_chart.view._set_yrange(
|
|
||||||
yrange=yrange,
|
|
||||||
)
|
|
||||||
profiler('`vlm_chart.view._set_yrange()`')
|
|
||||||
# print(f'mx vlm: {last_mx_vlm} -> {mx_vlm_in_view}')
|
|
||||||
vars['last_mx_vlm'] = mx_vlm_in_view
|
|
||||||
|
|
||||||
for curve_name, flow in vlm_chart._flows.items():
|
|
||||||
|
|
||||||
if not flow.render:
|
|
||||||
continue
|
|
||||||
|
|
||||||
update_fsp_chart(
|
|
||||||
vlm_chart,
|
|
||||||
flow,
|
|
||||||
curve_name,
|
|
||||||
array_key=curve_name,
|
|
||||||
do_append=xpx < update_uppx,
|
|
||||||
)
|
|
||||||
# is this even doing anything?
|
|
||||||
# (pretty sure it's the real-time
|
|
||||||
# resizing from last quote?)
|
|
||||||
fvb = flow.plot.vb
|
|
||||||
fvb._set_yrange(
|
|
||||||
autoscale_linked_plots=False,
|
|
||||||
name=curve_name,
|
|
||||||
)
|
|
||||||
|
|
||||||
ticks_frame = quote.get('ticks', ())
|
ticks_frame = quote.get('ticks', ())
|
||||||
|
|
||||||
|
@ -510,15 +471,20 @@ def graphics_update_cycle(
|
||||||
|
|
||||||
# update ohlc sampled price bars
|
# update ohlc sampled price bars
|
||||||
if (
|
if (
|
||||||
xpx < update_uppx
|
do_rt_update
|
||||||
or i_diff > 0
|
or do_append
|
||||||
or trigger_all
|
or trigger_all
|
||||||
):
|
):
|
||||||
chart.update_graphics_from_flow(
|
chart.update_graphics_from_flow(
|
||||||
chart.name,
|
chart.name,
|
||||||
do_append=xpx < update_uppx,
|
# do_append=uppx < update_uppx,
|
||||||
|
do_append=do_append,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# NOTE: we always update the "last" datum
|
||||||
|
# since the current range should at least be updated
|
||||||
|
# to it's max/min on the last pixel.
|
||||||
|
|
||||||
# iterate in FIFO order per tick-frame
|
# iterate in FIFO order per tick-frame
|
||||||
for typ, tick in lasts.items():
|
for typ, tick in lasts.items():
|
||||||
|
|
||||||
|
@ -623,11 +589,10 @@ def graphics_update_cycle(
|
||||||
vars['last_mx'], vars['last_mn'] = mx, mn
|
vars['last_mx'], vars['last_mn'] = mx, mn
|
||||||
|
|
||||||
# run synchronous update on all linked flows
|
# run synchronous update on all linked flows
|
||||||
for curve_name, flow in chart._flows.items():
|
|
||||||
# TODO: should the "main" (aka source) flow be special?
|
# TODO: should the "main" (aka source) flow be special?
|
||||||
if curve_name == chart.data_key:
|
for curve_name, flow in chart._flows.items():
|
||||||
continue
|
# update any overlayed fsp flows
|
||||||
|
if curve_name != chart.data_key:
|
||||||
update_fsp_chart(
|
update_fsp_chart(
|
||||||
chart,
|
chart,
|
||||||
flow,
|
flow,
|
||||||
|
@ -635,6 +600,105 @@ def graphics_update_cycle(
|
||||||
array_key=curve_name,
|
array_key=curve_name,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# even if we're downsampled bigly
|
||||||
|
# draw the last datum in the final
|
||||||
|
# px column to give the user the mx/mn
|
||||||
|
# range of that set.
|
||||||
|
if (
|
||||||
|
not do_append
|
||||||
|
# and not do_rt_update
|
||||||
|
and liv
|
||||||
|
):
|
||||||
|
flow.draw_last(
|
||||||
|
array_key=curve_name,
|
||||||
|
only_last_uppx=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
# volume chart logic..
|
||||||
|
# TODO: can we unify this with the above loop?
|
||||||
|
if vlm_chart:
|
||||||
|
# always update y-label
|
||||||
|
ds.vlm_sticky.update_from_data(
|
||||||
|
*array[-1][['index', 'volume']]
|
||||||
|
)
|
||||||
|
|
||||||
|
if (
|
||||||
|
(
|
||||||
|
do_rt_update
|
||||||
|
or do_append
|
||||||
|
and liv
|
||||||
|
)
|
||||||
|
or trigger_all
|
||||||
|
):
|
||||||
|
# TODO: make it so this doesn't have to be called
|
||||||
|
# once the $vlm is up?
|
||||||
|
vlm_chart.update_graphics_from_flow(
|
||||||
|
'volume',
|
||||||
|
# UGGGh, see ``maxmin()`` impl in `._fsp` for
|
||||||
|
# the overlayed plotitems... we need a better
|
||||||
|
# bay to invoke a maxmin per overlay..
|
||||||
|
render=False,
|
||||||
|
# XXX: ^^^^ THIS IS SUPER IMPORTANT! ^^^^
|
||||||
|
# without this, since we disable the
|
||||||
|
# 'volume' (units) chart after the $vlm starts
|
||||||
|
# up we need to be sure to enable this
|
||||||
|
# auto-ranging otherwise there will be no handler
|
||||||
|
# connected to update accompanying overlay
|
||||||
|
# graphics..
|
||||||
|
)
|
||||||
|
profiler('`vlm_chart.update_graphics_from_flow()`')
|
||||||
|
|
||||||
|
if (
|
||||||
|
mx_vlm_in_view != vars['last_mx_vlm']
|
||||||
|
):
|
||||||
|
yrange = (0, mx_vlm_in_view * 1.375)
|
||||||
|
vlm_chart.view._set_yrange(
|
||||||
|
yrange=yrange,
|
||||||
|
)
|
||||||
|
profiler('`vlm_chart.view._set_yrange()`')
|
||||||
|
# print(f'mx vlm: {last_mx_vlm} -> {mx_vlm_in_view}')
|
||||||
|
vars['last_mx_vlm'] = mx_vlm_in_view
|
||||||
|
|
||||||
|
for curve_name, flow in vlm_chart._flows.items():
|
||||||
|
|
||||||
|
if (
|
||||||
|
curve_name != 'volume' and
|
||||||
|
flow.render and (
|
||||||
|
liv and
|
||||||
|
do_rt_update or do_append
|
||||||
|
)
|
||||||
|
):
|
||||||
|
update_fsp_chart(
|
||||||
|
vlm_chart,
|
||||||
|
flow,
|
||||||
|
curve_name,
|
||||||
|
array_key=curve_name,
|
||||||
|
# do_append=uppx < update_uppx,
|
||||||
|
do_append=do_append,
|
||||||
|
)
|
||||||
|
# is this even doing anything?
|
||||||
|
# (pretty sure it's the real-time
|
||||||
|
# resizing from last quote?)
|
||||||
|
fvb = flow.plot.vb
|
||||||
|
fvb._set_yrange(
|
||||||
|
name=curve_name,
|
||||||
|
)
|
||||||
|
|
||||||
|
elif (
|
||||||
|
curve_name != 'volume'
|
||||||
|
and not do_append
|
||||||
|
and liv
|
||||||
|
and uppx >= 1
|
||||||
|
# even if we're downsampled bigly
|
||||||
|
# draw the last datum in the final
|
||||||
|
# px column to give the user the mx/mn
|
||||||
|
# range of that set.
|
||||||
|
):
|
||||||
|
# always update the last datum-element
|
||||||
|
# graphic for all flows
|
||||||
|
# print(f'drawing last {flow.name}')
|
||||||
|
flow.draw_last(array_key=curve_name)
|
||||||
|
|
||||||
|
|
||||||
async def display_symbol_data(
|
async def display_symbol_data(
|
||||||
godwidget: GodWidget,
|
godwidget: GodWidget,
|
||||||
|
|
|
@ -49,10 +49,6 @@ from . import _style
|
||||||
log = get_logger(__name__)
|
log = get_logger(__name__)
|
||||||
|
|
||||||
# pyqtgraph global config
|
# pyqtgraph global config
|
||||||
# might as well enable this for now?
|
|
||||||
pg.useOpenGL = True
|
|
||||||
pg.enableExperimental = True
|
|
||||||
|
|
||||||
# engage core tweaks that give us better response
|
# engage core tweaks that give us better response
|
||||||
# latency then the average pg user
|
# latency then the average pg user
|
||||||
_do_overrides()
|
_do_overrides()
|
||||||
|
|
|
@ -0,0 +1,83 @@
|
||||||
|
# piker: trading gear for hackers
|
||||||
|
# Copyright (C) Tyler Goodlet (in stewardship for piker0)
|
||||||
|
|
||||||
|
# This program is free software: you can redistribute it and/or modify
|
||||||
|
# it under the terms of the GNU Affero General Public License as published by
|
||||||
|
# the Free Software Foundation, either version 3 of the License, or
|
||||||
|
# (at your option) any later version.
|
||||||
|
|
||||||
|
# This program is distributed in the hope that it will be useful,
|
||||||
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
|
# GNU Affero General Public License for more details.
|
||||||
|
|
||||||
|
# You should have received a copy of the GNU Affero General Public License
|
||||||
|
# along with this program. If not, see <https://www.gnu.org/licenses/>.
|
||||||
|
|
||||||
|
"""
|
||||||
|
Feed status and controls widget(s) for embedding in a UI-pane.
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from __future__ import annotations
|
||||||
|
from textwrap import dedent
|
||||||
|
from typing import TYPE_CHECKING
|
||||||
|
|
||||||
|
# from PyQt5.QtCore import Qt
|
||||||
|
|
||||||
|
from ._style import _font, _font_small
|
||||||
|
# from ..calc import humanize
|
||||||
|
from ._label import FormatLabel
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from ._chart import ChartPlotWidget
|
||||||
|
from ..data.feed import Feed
|
||||||
|
from ._forms import FieldsForm
|
||||||
|
|
||||||
|
|
||||||
|
def mk_feed_label(
|
||||||
|
form: FieldsForm,
|
||||||
|
feed: Feed,
|
||||||
|
chart: ChartPlotWidget,
|
||||||
|
|
||||||
|
) -> FormatLabel:
|
||||||
|
'''
|
||||||
|
Generate a label from feed meta-data to be displayed
|
||||||
|
in a UI sidepane.
|
||||||
|
|
||||||
|
TODO: eventually buttons for changing settings over
|
||||||
|
a feed control protocol.
|
||||||
|
|
||||||
|
'''
|
||||||
|
status = feed.status
|
||||||
|
assert status
|
||||||
|
|
||||||
|
msg = dedent("""
|
||||||
|
actor: **{actor_name}**\n
|
||||||
|
|_ @**{host}:{port}**\n
|
||||||
|
""")
|
||||||
|
|
||||||
|
for key, val in status.items():
|
||||||
|
if key in ('host', 'port', 'actor_name'):
|
||||||
|
continue
|
||||||
|
msg += f'\n|_ {key}: **{{{key}}}**\n'
|
||||||
|
|
||||||
|
feed_label = FormatLabel(
|
||||||
|
fmt_str=msg,
|
||||||
|
# |_ streams: **{symbols}**\n
|
||||||
|
font=_font.font,
|
||||||
|
font_size=_font_small.px_size,
|
||||||
|
font_color='default_lightest',
|
||||||
|
)
|
||||||
|
|
||||||
|
# form.vbox.setAlignment(feed_label, Qt.AlignBottom)
|
||||||
|
# form.vbox.setAlignment(Qt.AlignBottom)
|
||||||
|
_ = chart.height() - (
|
||||||
|
form.height() +
|
||||||
|
form.fill_bar.height()
|
||||||
|
# feed_label.height()
|
||||||
|
)
|
||||||
|
|
||||||
|
feed_label.format(**feed.status)
|
||||||
|
|
||||||
|
return feed_label
|
1267
piker/ui/_flows.py
1267
piker/ui/_flows.py
File diff suppressed because it is too large
Load Diff
|
@ -750,12 +750,12 @@ def mk_order_pane_layout(
|
||||||
parent=parent,
|
parent=parent,
|
||||||
fields_schema={
|
fields_schema={
|
||||||
'account': {
|
'account': {
|
||||||
'label': '**account**:',
|
'label': '**accnt**:',
|
||||||
'type': 'select',
|
'type': 'select',
|
||||||
'default_value': ['paper'],
|
'default_value': ['paper'],
|
||||||
},
|
},
|
||||||
'size_unit': {
|
'size_unit': {
|
||||||
'label': '**allocate**:',
|
'label': '**alloc**:',
|
||||||
'type': 'select',
|
'type': 'select',
|
||||||
'default_value': [
|
'default_value': [
|
||||||
'$ size',
|
'$ size',
|
||||||
|
|
|
@ -440,7 +440,7 @@ class FspAdmin:
|
||||||
# if the chart isn't hidden try to update
|
# if the chart isn't hidden try to update
|
||||||
# the data on screen.
|
# the data on screen.
|
||||||
if not self.linked.isHidden():
|
if not self.linked.isHidden():
|
||||||
log.info(f'Re-syncing graphics for fsp: {ns_path}')
|
log.debug(f'Re-syncing graphics for fsp: {ns_path}')
|
||||||
self.linked.graphics_cycle(
|
self.linked.graphics_cycle(
|
||||||
trigger_all=True,
|
trigger_all=True,
|
||||||
prepend_update_index=info['first'],
|
prepend_update_index=info['first'],
|
||||||
|
@ -635,7 +635,7 @@ async def open_vlm_displays(
|
||||||
)
|
)
|
||||||
|
|
||||||
# force 0 to always be in view
|
# force 0 to always be in view
|
||||||
def maxmin(
|
def multi_maxmin(
|
||||||
names: list[str],
|
names: list[str],
|
||||||
|
|
||||||
) -> tuple[float, float]:
|
) -> tuple[float, float]:
|
||||||
|
@ -651,7 +651,7 @@ async def open_vlm_displays(
|
||||||
|
|
||||||
return 0, mx
|
return 0, mx
|
||||||
|
|
||||||
chart.view.maxmin = partial(maxmin, names=['volume'])
|
chart.view.maxmin = partial(multi_maxmin, names=['volume'])
|
||||||
|
|
||||||
# TODO: fix the x-axis label issue where if you put
|
# TODO: fix the x-axis label issue where if you put
|
||||||
# the axis on the left it's totally not lined up...
|
# the axis on the left it's totally not lined up...
|
||||||
|
@ -741,19 +741,20 @@ async def open_vlm_displays(
|
||||||
'dolla_vlm',
|
'dolla_vlm',
|
||||||
'dark_vlm',
|
'dark_vlm',
|
||||||
]
|
]
|
||||||
dvlm_rate_fields = [
|
# dvlm_rate_fields = [
|
||||||
'dvlm_rate',
|
# 'dvlm_rate',
|
||||||
'dark_dvlm_rate',
|
# 'dark_dvlm_rate',
|
||||||
]
|
# ]
|
||||||
trade_rate_fields = [
|
trade_rate_fields = [
|
||||||
'trade_rate',
|
'trade_rate',
|
||||||
'dark_trade_rate',
|
'dark_trade_rate',
|
||||||
]
|
]
|
||||||
|
|
||||||
group_mxmn = partial(
|
group_mxmn = partial(
|
||||||
maxmin,
|
multi_maxmin,
|
||||||
# keep both regular and dark vlm in view
|
# keep both regular and dark vlm in view
|
||||||
names=fields + dvlm_rate_fields,
|
names=fields,
|
||||||
|
# names=fields + dvlm_rate_fields,
|
||||||
)
|
)
|
||||||
|
|
||||||
# add custom auto range handler
|
# add custom auto range handler
|
||||||
|
@ -820,11 +821,11 @@ async def open_vlm_displays(
|
||||||
)
|
)
|
||||||
await started.wait()
|
await started.wait()
|
||||||
|
|
||||||
chart_curves(
|
# chart_curves(
|
||||||
dvlm_rate_fields,
|
# dvlm_rate_fields,
|
||||||
dvlm_pi,
|
# dvlm_pi,
|
||||||
fr_shm,
|
# fr_shm,
|
||||||
)
|
# )
|
||||||
|
|
||||||
# TODO: is there a way to "sync" the dual axes such that only
|
# TODO: is there a way to "sync" the dual axes such that only
|
||||||
# one curve is needed?
|
# one curve is needed?
|
||||||
|
@ -862,7 +863,7 @@ async def open_vlm_displays(
|
||||||
)
|
)
|
||||||
# add custom auto range handler
|
# add custom auto range handler
|
||||||
tr_pi.vb.maxmin = partial(
|
tr_pi.vb.maxmin = partial(
|
||||||
maxmin,
|
multi_maxmin,
|
||||||
# keep both regular and dark vlm in view
|
# keep both regular and dark vlm in view
|
||||||
names=trade_rate_fields,
|
names=trade_rate_fields,
|
||||||
)
|
)
|
||||||
|
|
|
@ -570,6 +570,13 @@ class ChartView(ViewBox):
|
||||||
self._resetTarget()
|
self._resetTarget()
|
||||||
self.scaleBy(s, focal)
|
self.scaleBy(s, focal)
|
||||||
|
|
||||||
|
# XXX: the order of the next 2 lines i'm pretty sure
|
||||||
|
# matters, we want the resize to trigger before the graphics
|
||||||
|
# update, but i gotta feelin that because this one is signal
|
||||||
|
# based (and thus not necessarily sync invoked right away)
|
||||||
|
# that calling the resize method manually might work better.
|
||||||
|
self.sigRangeChangedManually.emit(mask)
|
||||||
|
|
||||||
# XXX: without this is seems as though sometimes
|
# XXX: without this is seems as though sometimes
|
||||||
# when zooming in from far out (and maybe vice versa?)
|
# when zooming in from far out (and maybe vice versa?)
|
||||||
# the signal isn't being fired enough since if you pan
|
# the signal isn't being fired enough since if you pan
|
||||||
|
@ -580,12 +587,6 @@ class ChartView(ViewBox):
|
||||||
# fires don't happen?
|
# fires don't happen?
|
||||||
self.maybe_downsample_graphics()
|
self.maybe_downsample_graphics()
|
||||||
|
|
||||||
self.sigRangeChangedManually.emit(mask)
|
|
||||||
|
|
||||||
# self._ic.set()
|
|
||||||
# self._ic = None
|
|
||||||
# self.chart.resume_all_feeds()
|
|
||||||
|
|
||||||
ev.accept()
|
ev.accept()
|
||||||
|
|
||||||
def mouseDragEvent(
|
def mouseDragEvent(
|
||||||
|
@ -748,7 +749,6 @@ class ChartView(ViewBox):
|
||||||
# set from recursion errors.
|
# set from recursion errors.
|
||||||
autoscale_linked_plots: bool = False,
|
autoscale_linked_plots: bool = False,
|
||||||
name: Optional[str] = None,
|
name: Optional[str] = None,
|
||||||
# autoscale_overlays: bool = False,
|
|
||||||
|
|
||||||
) -> None:
|
) -> None:
|
||||||
'''
|
'''
|
||||||
|
@ -759,8 +759,10 @@ class ChartView(ViewBox):
|
||||||
data set.
|
data set.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
|
name = self.name
|
||||||
|
# print(f'YRANGE ON {name}')
|
||||||
profiler = pg.debug.Profiler(
|
profiler = pg.debug.Profiler(
|
||||||
msg=f'`ChartView._set_yrange()`: `{self.name}`',
|
msg=f'`ChartView._set_yrange()`: `{name}`',
|
||||||
disabled=not pg_profile_enabled(),
|
disabled=not pg_profile_enabled(),
|
||||||
ms_threshold=ms_slower_then,
|
ms_threshold=ms_slower_then,
|
||||||
delayed=True,
|
delayed=True,
|
||||||
|
@ -788,47 +790,17 @@ class ChartView(ViewBox):
|
||||||
elif yrange is not None:
|
elif yrange is not None:
|
||||||
ylow, yhigh = yrange
|
ylow, yhigh = yrange
|
||||||
|
|
||||||
# calculate max, min y values in viewable x-range from data.
|
|
||||||
# Make sure min bars/datums on screen is adhered.
|
|
||||||
# else:
|
|
||||||
# TODO: eventually we should point to the
|
|
||||||
# ``FlowsTable`` (or wtv) which should perform
|
|
||||||
# the group operations?
|
|
||||||
|
|
||||||
# flow = chart._flows[name or chart.name]
|
|
||||||
# br = bars_range or chart.bars_range()
|
|
||||||
# br = bars_range or chart.bars_range()
|
|
||||||
# profiler(f'got bars range: {br}')
|
|
||||||
|
|
||||||
# TODO: maybe should be a method on the
|
|
||||||
# chart widget/item?
|
|
||||||
# if False:
|
|
||||||
# if autoscale_linked_plots:
|
|
||||||
# # avoid recursion by sibling plots
|
|
||||||
# linked = self.linkedsplits
|
|
||||||
# plots = list(linked.subplots.copy().values())
|
|
||||||
# main = linked.chart
|
|
||||||
# if main:
|
|
||||||
# plots.append(main)
|
|
||||||
|
|
||||||
# for chart in plots:
|
|
||||||
# if chart and not chart._static_yrange:
|
|
||||||
# chart.cv._set_yrange(
|
|
||||||
# # bars_range=br,
|
|
||||||
# autoscale_linked_plots=False,
|
|
||||||
# )
|
|
||||||
# profiler('autoscaled linked plots')
|
|
||||||
|
|
||||||
if set_range:
|
if set_range:
|
||||||
|
|
||||||
if not yrange:
|
|
||||||
# XXX: only compute the mxmn range
|
# XXX: only compute the mxmn range
|
||||||
# if none is provided as input!
|
# if none is provided as input!
|
||||||
|
if not yrange:
|
||||||
|
# flow = chart._flows[name]
|
||||||
yrange = self._maxmin()
|
yrange = self._maxmin()
|
||||||
|
|
||||||
if yrange is None:
|
if yrange is None:
|
||||||
log.warning(f'No yrange provided for {self.name}!?')
|
log.warning(f'No yrange provided for {name}!?')
|
||||||
print(f"WTF NO YRANGE {self.name}")
|
print(f"WTF NO YRANGE {name}")
|
||||||
return
|
return
|
||||||
|
|
||||||
ylow, yhigh = yrange
|
ylow, yhigh = yrange
|
||||||
|
@ -921,7 +893,10 @@ class ChartView(ViewBox):
|
||||||
else:
|
else:
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
def maybe_downsample_graphics(self):
|
def maybe_downsample_graphics(
|
||||||
|
self,
|
||||||
|
autoscale_overlays: bool = True,
|
||||||
|
):
|
||||||
|
|
||||||
profiler = pg.debug.Profiler(
|
profiler = pg.debug.Profiler(
|
||||||
msg=f'ChartView.maybe_downsample_graphics() for {self.name}',
|
msg=f'ChartView.maybe_downsample_graphics() for {self.name}',
|
||||||
|
@ -955,9 +930,18 @@ class ChartView(ViewBox):
|
||||||
chart.update_graphics_from_flow(
|
chart.update_graphics_from_flow(
|
||||||
name,
|
name,
|
||||||
use_vr=True,
|
use_vr=True,
|
||||||
|
|
||||||
# gets passed down into graphics obj
|
|
||||||
# profiler=profiler,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# for each overlay on this chart auto-scale the
|
||||||
|
# y-range to max-min values.
|
||||||
|
if autoscale_overlays:
|
||||||
|
overlay = chart.pi_overlay
|
||||||
|
if overlay:
|
||||||
|
for pi in overlay.overlays:
|
||||||
|
pi.vb._set_yrange(
|
||||||
|
# TODO: get the range once up front...
|
||||||
|
# bars_range=br,
|
||||||
|
)
|
||||||
|
profiler('autoscaled linked plots')
|
||||||
|
|
||||||
profiler(f'<{chart_name}>.update_graphics_from_flow({name})')
|
profiler(f'<{chart_name}>.update_graphics_from_flow({name})')
|
||||||
|
|
|
@ -27,13 +27,11 @@ import numpy as np
|
||||||
import pyqtgraph as pg
|
import pyqtgraph as pg
|
||||||
from PyQt5 import QtCore, QtGui, QtWidgets
|
from PyQt5 import QtCore, QtGui, QtWidgets
|
||||||
from PyQt5.QtCore import QLineF, QPointF
|
from PyQt5.QtCore import QLineF, QPointF
|
||||||
|
from PyQt5.QtGui import QPainterPath
|
||||||
|
|
||||||
from .._profile import pg_profile_enabled, ms_slower_then
|
from .._profile import pg_profile_enabled, ms_slower_then
|
||||||
from ._style import hcolor
|
from ._style import hcolor
|
||||||
from ..log import get_logger
|
from ..log import get_logger
|
||||||
from ._curve import FastAppendCurve
|
|
||||||
from ._compression import ohlc_flatten
|
|
||||||
from ._pathops import gen_ohlc_qpath
|
|
||||||
|
|
||||||
if TYPE_CHECKING:
|
if TYPE_CHECKING:
|
||||||
from ._chart import LinkedSplits
|
from ._chart import LinkedSplits
|
||||||
|
@ -44,6 +42,7 @@ log = get_logger(__name__)
|
||||||
|
|
||||||
def bar_from_ohlc_row(
|
def bar_from_ohlc_row(
|
||||||
row: np.ndarray,
|
row: np.ndarray,
|
||||||
|
# 0.5 is no overlap between arms, 1.0 is full overlap
|
||||||
w: float = 0.43
|
w: float = 0.43
|
||||||
|
|
||||||
) -> tuple[QLineF]:
|
) -> tuple[QLineF]:
|
||||||
|
@ -87,11 +86,6 @@ class BarItems(pg.GraphicsObject):
|
||||||
"Price range" bars graphics rendered from a OHLC sampled sequence.
|
"Price range" bars graphics rendered from a OHLC sampled sequence.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
sigPlotChanged = QtCore.pyqtSignal(object)
|
|
||||||
|
|
||||||
# 0.5 is no overlap between arms, 1.0 is full overlap
|
|
||||||
w: float = 0.43
|
|
||||||
|
|
||||||
def __init__(
|
def __init__(
|
||||||
self,
|
self,
|
||||||
linked: LinkedSplits,
|
linked: LinkedSplits,
|
||||||
|
@ -111,143 +105,14 @@ class BarItems(pg.GraphicsObject):
|
||||||
self.last_bar_pen = pg.mkPen(hcolor(last_bar_color), width=2)
|
self.last_bar_pen = pg.mkPen(hcolor(last_bar_color), width=2)
|
||||||
self._name = name
|
self._name = name
|
||||||
|
|
||||||
self._ds_line_xy: Optional[
|
|
||||||
tuple[np.ndarray, np.ndarray]
|
|
||||||
] = None
|
|
||||||
|
|
||||||
# NOTE: this prevents redraws on mouse interaction which is
|
|
||||||
# a huge boon for avg interaction latency.
|
|
||||||
|
|
||||||
# TODO: one question still remaining is if this makes trasform
|
|
||||||
# interactions slower (such as zooming) and if so maybe if/when
|
|
||||||
# we implement a "history" mode for the view we disable this in
|
|
||||||
# that mode?
|
|
||||||
self.setCacheMode(QtWidgets.QGraphicsItem.DeviceCoordinateCache)
|
self.setCacheMode(QtWidgets.QGraphicsItem.DeviceCoordinateCache)
|
||||||
|
self.path = QPainterPath()
|
||||||
self._pi = plotitem
|
|
||||||
self.path = QtGui.QPainterPath()
|
|
||||||
self.fast_path = QtGui.QPainterPath()
|
|
||||||
|
|
||||||
self._xrange: tuple[int, int]
|
|
||||||
# self._yrange: tuple[float, float]
|
|
||||||
self._vrange = None
|
|
||||||
|
|
||||||
# TODO: don't render the full backing array each time
|
|
||||||
# self._path_data = None
|
|
||||||
self._last_bar_lines: Optional[tuple[QLineF, ...]] = None
|
self._last_bar_lines: Optional[tuple[QLineF, ...]] = None
|
||||||
|
|
||||||
# track the current length of drawable lines within the larger array
|
|
||||||
self.start_index: int = 0
|
|
||||||
self.stop_index: int = 0
|
|
||||||
|
|
||||||
# downsampler-line state
|
|
||||||
self._in_ds: bool = False
|
|
||||||
self._ds_line: Optional[FastAppendCurve] = None
|
|
||||||
self._dsi: tuple[int, int] = 0, 0
|
|
||||||
self._xs_in_px: float = 0
|
|
||||||
|
|
||||||
def draw_from_data(
|
|
||||||
self,
|
|
||||||
ohlc: np.ndarray,
|
|
||||||
start: int = 0,
|
|
||||||
|
|
||||||
) -> QtGui.QPainterPath:
|
|
||||||
'''
|
|
||||||
Draw OHLC datum graphics from a ``np.ndarray``.
|
|
||||||
|
|
||||||
This routine is usually only called to draw the initial history.
|
|
||||||
|
|
||||||
'''
|
|
||||||
hist, last = ohlc[:-1], ohlc[-1]
|
|
||||||
self.path = gen_ohlc_qpath(hist, start, self.w)
|
|
||||||
|
|
||||||
# save graphics for later reference and keep track
|
|
||||||
# of current internal "last index"
|
|
||||||
# self.start_index = len(ohlc)
|
|
||||||
index = ohlc['index']
|
|
||||||
self._xrange = (index[0], index[-1])
|
|
||||||
# self._yrange = (
|
|
||||||
# np.nanmax(ohlc['high']),
|
|
||||||
# np.nanmin(ohlc['low']),
|
|
||||||
# )
|
|
||||||
|
|
||||||
# up to last to avoid double draw of last bar
|
|
||||||
self._last_bar_lines = bar_from_ohlc_row(last, self.w)
|
|
||||||
|
|
||||||
x, y = self._ds_line_xy = ohlc_flatten(ohlc)
|
|
||||||
|
|
||||||
# TODO: figuring out the most optimial size for the ideal
|
|
||||||
# curve-path by,
|
|
||||||
# - calcing the display's max px width `.screen()`
|
|
||||||
# - drawing a curve and figuring out it's capacity:
|
|
||||||
# https://doc.qt.io/qt-5/qpainterpath.html#capacity
|
|
||||||
# - reserving that cap for each curve-mapped-to-shm with
|
|
||||||
|
|
||||||
# - leveraging clearing when needed to redraw the entire
|
|
||||||
# curve that does not release mem allocs:
|
|
||||||
# https://doc.qt.io/qt-5/qpainterpath.html#clear
|
|
||||||
curve = FastAppendCurve(
|
|
||||||
y=y,
|
|
||||||
x=x,
|
|
||||||
name='OHLC',
|
|
||||||
color=self._color,
|
|
||||||
)
|
|
||||||
curve.hide()
|
|
||||||
self._pi.addItem(curve)
|
|
||||||
self._ds_line = curve
|
|
||||||
|
|
||||||
# self._ds_xrange = (index[0], index[-1])
|
|
||||||
|
|
||||||
# trigger render
|
|
||||||
# https://doc.qt.io/qt-5/qgraphicsitem.html#update
|
|
||||||
self.update()
|
|
||||||
|
|
||||||
return self.path
|
|
||||||
|
|
||||||
def x_uppx(self) -> int:
|
def x_uppx(self) -> int:
|
||||||
if self._ds_line:
|
# we expect the downsample curve report this.
|
||||||
return self._ds_line.x_uppx()
|
|
||||||
else:
|
|
||||||
return 0
|
return 0
|
||||||
|
|
||||||
def draw_last(
|
|
||||||
self,
|
|
||||||
last: np.ndarray,
|
|
||||||
|
|
||||||
) -> None:
|
|
||||||
# generate new lines objects for updatable "current bar"
|
|
||||||
self._last_bar_lines = bar_from_ohlc_row(last, self.w)
|
|
||||||
|
|
||||||
# last bar update
|
|
||||||
i, o, h, l, last, v = last[
|
|
||||||
['index', 'open', 'high', 'low', 'close', 'volume']
|
|
||||||
]
|
|
||||||
# assert i == self.start_index - 1
|
|
||||||
# assert i == last_index
|
|
||||||
body, larm, rarm = self._last_bar_lines
|
|
||||||
|
|
||||||
# XXX: is there a faster way to modify this?
|
|
||||||
rarm.setLine(rarm.x1(), last, rarm.x2(), last)
|
|
||||||
|
|
||||||
# writer is responsible for changing open on "first" volume of bar
|
|
||||||
larm.setLine(larm.x1(), o, larm.x2(), o)
|
|
||||||
|
|
||||||
if l != h: # noqa
|
|
||||||
|
|
||||||
if body is None:
|
|
||||||
body = self._last_bar_lines[0] = QLineF(i, l, i, h)
|
|
||||||
else:
|
|
||||||
# update body
|
|
||||||
body.setLine(i, l, i, h)
|
|
||||||
|
|
||||||
# XXX: pretty sure this is causing an issue where the bar has
|
|
||||||
# a large upward move right before the next sample and the body
|
|
||||||
# is getting set to None since the next bar is flat but the shm
|
|
||||||
# array index update wasn't read by the time this code runs. Iow
|
|
||||||
# we're doing this removal of the body for a bar index that is
|
|
||||||
# now out of date / from some previous sample. It's weird
|
|
||||||
# though because i've seen it do this to bars i - 3 back?
|
|
||||||
|
|
||||||
def boundingRect(self):
|
def boundingRect(self):
|
||||||
# Qt docs: https://doc.qt.io/qt-5/qgraphicsitem.html#boundingRect
|
# Qt docs: https://doc.qt.io/qt-5/qgraphicsitem.html#boundingRect
|
||||||
|
|
||||||
|
@ -270,16 +135,6 @@ class BarItems(pg.GraphicsObject):
|
||||||
hb.bottomRight(),
|
hb.bottomRight(),
|
||||||
)
|
)
|
||||||
|
|
||||||
# fp = self.fast_path
|
|
||||||
# if fp:
|
|
||||||
# fhb = fp.controlPointRect()
|
|
||||||
# print((hb_tl, hb_br))
|
|
||||||
# print(fhb)
|
|
||||||
# hb_tl, hb_br = (
|
|
||||||
# fhb.topLeft() + hb.topLeft(),
|
|
||||||
# fhb.bottomRight() + hb.bottomRight(),
|
|
||||||
# )
|
|
||||||
|
|
||||||
# need to include last bar height or BR will be off
|
# need to include last bar height or BR will be off
|
||||||
mx_y = hb_br.y()
|
mx_y = hb_br.y()
|
||||||
mn_y = hb_tl.y()
|
mn_y = hb_tl.y()
|
||||||
|
@ -315,9 +170,6 @@ class BarItems(pg.GraphicsObject):
|
||||||
|
|
||||||
) -> None:
|
) -> None:
|
||||||
|
|
||||||
if self._in_ds:
|
|
||||||
return
|
|
||||||
|
|
||||||
profiler = pg.debug.Profiler(
|
profiler = pg.debug.Profiler(
|
||||||
disabled=not pg_profile_enabled(),
|
disabled=not pg_profile_enabled(),
|
||||||
ms_threshold=ms_slower_then,
|
ms_threshold=ms_slower_then,
|
||||||
|
@ -332,6 +184,7 @@ class BarItems(pg.GraphicsObject):
|
||||||
# lead to any perf gains other then when zoomed in to less bars
|
# lead to any perf gains other then when zoomed in to less bars
|
||||||
# in view.
|
# in view.
|
||||||
p.setPen(self.last_bar_pen)
|
p.setPen(self.last_bar_pen)
|
||||||
|
if self._last_bar_lines:
|
||||||
p.drawLines(*tuple(filter(bool, self._last_bar_lines)))
|
p.drawLines(*tuple(filter(bool, self._last_bar_lines)))
|
||||||
profiler('draw last bar')
|
profiler('draw last bar')
|
||||||
|
|
||||||
|
@ -339,6 +192,59 @@ class BarItems(pg.GraphicsObject):
|
||||||
p.drawPath(self.path)
|
p.drawPath(self.path)
|
||||||
profiler(f'draw history path: {self.path.capacity()}')
|
profiler(f'draw history path: {self.path.capacity()}')
|
||||||
|
|
||||||
# if self.fast_path:
|
def draw_last_datum(
|
||||||
# p.drawPath(self.fast_path)
|
self,
|
||||||
# profiler('draw fast path')
|
path: QPainterPath,
|
||||||
|
src_data: np.ndarray,
|
||||||
|
render_data: np.ndarray,
|
||||||
|
reset: bool,
|
||||||
|
array_key: str,
|
||||||
|
|
||||||
|
fields: list[str] = [
|
||||||
|
'index',
|
||||||
|
'open',
|
||||||
|
'high',
|
||||||
|
'low',
|
||||||
|
'close',
|
||||||
|
],
|
||||||
|
|
||||||
|
) -> None:
|
||||||
|
|
||||||
|
# relevant fields
|
||||||
|
ohlc = src_data[fields]
|
||||||
|
last_row = ohlc[-1:]
|
||||||
|
|
||||||
|
# individual values
|
||||||
|
last_row = i, o, h, l, last = ohlc[-1]
|
||||||
|
|
||||||
|
# generate new lines objects for updatable "current bar"
|
||||||
|
self._last_bar_lines = bar_from_ohlc_row(last_row)
|
||||||
|
|
||||||
|
# assert i == graphics.start_index - 1
|
||||||
|
# assert i == last_index
|
||||||
|
body, larm, rarm = self._last_bar_lines
|
||||||
|
|
||||||
|
# XXX: is there a faster way to modify this?
|
||||||
|
rarm.setLine(rarm.x1(), last, rarm.x2(), last)
|
||||||
|
|
||||||
|
# writer is responsible for changing open on "first" volume of bar
|
||||||
|
larm.setLine(larm.x1(), o, larm.x2(), o)
|
||||||
|
|
||||||
|
if l != h: # noqa
|
||||||
|
|
||||||
|
if body is None:
|
||||||
|
body = self._last_bar_lines[0] = QLineF(i, l, i, h)
|
||||||
|
else:
|
||||||
|
# update body
|
||||||
|
body.setLine(i, l, i, h)
|
||||||
|
|
||||||
|
# XXX: pretty sure this is causing an issue where the
|
||||||
|
# bar has a large upward move right before the next
|
||||||
|
# sample and the body is getting set to None since the
|
||||||
|
# next bar is flat but the shm array index update wasn't
|
||||||
|
# read by the time this code runs. Iow we're doing this
|
||||||
|
# removal of the body for a bar index that is now out of
|
||||||
|
# date / from some previous sample. It's weird though
|
||||||
|
# because i've seen it do this to bars i - 3 back?
|
||||||
|
|
||||||
|
return ohlc['index'], ohlc['close']
|
||||||
|
|
|
@ -17,31 +17,34 @@
|
||||||
Super fast ``QPainterPath`` generation related operator routines.
|
Super fast ``QPainterPath`` generation related operator routines.
|
||||||
|
|
||||||
"""
|
"""
|
||||||
|
from __future__ import annotations
|
||||||
from typing import (
|
from typing import (
|
||||||
Optional,
|
# Optional,
|
||||||
|
TYPE_CHECKING,
|
||||||
)
|
)
|
||||||
|
|
||||||
import numpy as np
|
import numpy as np
|
||||||
from numpy.lib import recfunctions as rfn
|
from numpy.lib import recfunctions as rfn
|
||||||
from numba import njit, float64, int64 # , optional
|
from numba import njit, float64, int64 # , optional
|
||||||
import pyqtgraph as pg
|
# import pyqtgraph as pg
|
||||||
from PyQt5 import QtGui
|
from PyQt5 import QtGui
|
||||||
# from PyQt5.QtCore import QLineF, QPointF
|
# from PyQt5.QtCore import QLineF, QPointF
|
||||||
|
|
||||||
from ..data._sharedmem import (
|
from ..data._sharedmem import (
|
||||||
ShmArray,
|
ShmArray,
|
||||||
)
|
)
|
||||||
from .._profile import pg_profile_enabled, ms_slower_then
|
# from .._profile import pg_profile_enabled, ms_slower_then
|
||||||
from ._compression import (
|
from ._compression import (
|
||||||
# ohlc_flatten,
|
|
||||||
ds_m4,
|
ds_m4,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
if TYPE_CHECKING:
|
||||||
|
from ._flows import Renderer
|
||||||
|
|
||||||
|
|
||||||
def xy_downsample(
|
def xy_downsample(
|
||||||
x,
|
x,
|
||||||
y,
|
y,
|
||||||
px_width,
|
|
||||||
uppx,
|
uppx,
|
||||||
|
|
||||||
x_spacer: float = 0.5,
|
x_spacer: float = 0.5,
|
||||||
|
@ -54,9 +57,7 @@ def xy_downsample(
|
||||||
bins, x, y = ds_m4(
|
bins, x, y = ds_m4(
|
||||||
x,
|
x,
|
||||||
y,
|
y,
|
||||||
px_width=px_width,
|
uppx,
|
||||||
uppx=uppx,
|
|
||||||
# log_scale=bool(uppx)
|
|
||||||
)
|
)
|
||||||
|
|
||||||
# flatten output to 1d arrays suitable for path-graphics generation.
|
# flatten output to 1d arrays suitable for path-graphics generation.
|
||||||
|
@ -142,53 +143,35 @@ def path_arrays_from_ohlc(
|
||||||
|
|
||||||
|
|
||||||
def gen_ohlc_qpath(
|
def gen_ohlc_qpath(
|
||||||
|
r: Renderer,
|
||||||
data: np.ndarray,
|
data: np.ndarray,
|
||||||
|
array_key: str, # we ignore this
|
||||||
|
vr: tuple[int, int],
|
||||||
|
|
||||||
start: int = 0, # XXX: do we need this?
|
start: int = 0, # XXX: do we need this?
|
||||||
# 0.5 is no overlap between arms, 1.0 is full overlap
|
# 0.5 is no overlap between arms, 1.0 is full overlap
|
||||||
w: float = 0.43,
|
w: float = 0.43,
|
||||||
path: Optional[QtGui.QPainterPath] = None,
|
|
||||||
|
|
||||||
) -> QtGui.QPainterPath:
|
) -> QtGui.QPainterPath:
|
||||||
|
'''
|
||||||
|
More or less direct proxy to ``path_arrays_from_ohlc()``
|
||||||
|
but with closed in kwargs for line spacing.
|
||||||
|
|
||||||
path_was_none = path is None
|
'''
|
||||||
|
|
||||||
profiler = pg.debug.Profiler(
|
|
||||||
msg='gen_qpath ohlc',
|
|
||||||
disabled=not pg_profile_enabled(),
|
|
||||||
ms_threshold=ms_slower_then,
|
|
||||||
)
|
|
||||||
|
|
||||||
x, y, c = path_arrays_from_ohlc(
|
x, y, c = path_arrays_from_ohlc(
|
||||||
data,
|
data,
|
||||||
start,
|
start,
|
||||||
bar_gap=w,
|
bar_gap=w,
|
||||||
)
|
)
|
||||||
profiler("generate stream with numba")
|
return x, y, c
|
||||||
|
|
||||||
# TODO: numba the internals of this!
|
|
||||||
path = pg.functions.arrayToQPath(
|
|
||||||
x,
|
|
||||||
y,
|
|
||||||
connect=c,
|
|
||||||
path=path,
|
|
||||||
)
|
|
||||||
|
|
||||||
# avoid mem allocs if possible
|
|
||||||
if path_was_none:
|
|
||||||
path.reserve(path.capacity())
|
|
||||||
|
|
||||||
profiler("generate path with arrayToQPath")
|
|
||||||
|
|
||||||
return path
|
|
||||||
|
|
||||||
|
|
||||||
def ohlc_to_line(
|
def ohlc_to_line(
|
||||||
ohlc_shm: ShmArray,
|
ohlc_shm: ShmArray,
|
||||||
|
data_field: str,
|
||||||
fields: list[str] = ['open', 'high', 'low', 'close']
|
fields: list[str] = ['open', 'high', 'low', 'close']
|
||||||
|
|
||||||
) -> tuple[
|
) -> tuple[
|
||||||
int, # flattened first index
|
|
||||||
int, # flattened last index
|
|
||||||
np.ndarray,
|
np.ndarray,
|
||||||
np.ndarray,
|
np.ndarray,
|
||||||
]:
|
]:
|
||||||
|
@ -221,8 +204,6 @@ def ohlc_to_line(
|
||||||
assert y_out.any()
|
assert y_out.any()
|
||||||
|
|
||||||
return (
|
return (
|
||||||
first,
|
|
||||||
last,
|
|
||||||
x_out,
|
x_out,
|
||||||
y_out,
|
y_out,
|
||||||
)
|
)
|
||||||
|
@ -239,7 +220,6 @@ def to_step_format(
|
||||||
for use by path graphics generation.
|
for use by path graphics generation.
|
||||||
|
|
||||||
'''
|
'''
|
||||||
first = shm._first.value
|
|
||||||
i = shm._array['index'].copy()
|
i = shm._array['index'].copy()
|
||||||
out = shm._array[data_field].copy()
|
out = shm._array[data_field].copy()
|
||||||
|
|
||||||
|
@ -253,4 +233,4 @@ def to_step_format(
|
||||||
|
|
||||||
# start y at origin level
|
# start y at origin level
|
||||||
y_out[0, 0] = 0
|
y_out[0, 0] = 0
|
||||||
return first, x_out, y_out
|
return x_out, y_out
|
||||||
|
|
|
@ -287,7 +287,6 @@ class MainWindow(QtGui.QMainWindow):
|
||||||
app = QtGui.QApplication.instance()
|
app = QtGui.QApplication.instance()
|
||||||
geo = self.current_screen().geometry()
|
geo = self.current_screen().geometry()
|
||||||
h, w = geo.height(), geo.width()
|
h, w = geo.height(), geo.width()
|
||||||
self.setMaximumSize(w, h)
|
|
||||||
# use approx 1/3 of the area of the screen by default
|
# use approx 1/3 of the area of the screen by default
|
||||||
self._size = round(w * .666), round(h * .666)
|
self._size = round(w * .666), round(h * .666)
|
||||||
|
|
||||||
|
|
|
@ -30,6 +30,7 @@ import uuid
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
import tractor
|
import tractor
|
||||||
import trio
|
import trio
|
||||||
|
from PyQt5.QtCore import Qt
|
||||||
|
|
||||||
from .. import config
|
from .. import config
|
||||||
from ..clearing._client import open_ems, OrderBook
|
from ..clearing._client import open_ems, OrderBook
|
||||||
|
@ -37,6 +38,7 @@ from ..clearing._allocate import (
|
||||||
mk_allocator,
|
mk_allocator,
|
||||||
Position,
|
Position,
|
||||||
)
|
)
|
||||||
|
from ._style import _font
|
||||||
from ..data._source import Symbol
|
from ..data._source import Symbol
|
||||||
from ..data.feed import Feed
|
from ..data.feed import Feed
|
||||||
from ..log import get_logger
|
from ..log import get_logger
|
||||||
|
@ -46,7 +48,8 @@ from ._position import (
|
||||||
PositionTracker,
|
PositionTracker,
|
||||||
SettingsPane,
|
SettingsPane,
|
||||||
)
|
)
|
||||||
from ._label import FormatLabel
|
from ._forms import FieldsForm
|
||||||
|
# from ._label import FormatLabel
|
||||||
from ._window import MultiStatus
|
from ._window import MultiStatus
|
||||||
from ..clearing._messages import Order, BrokerdPosition
|
from ..clearing._messages import Order, BrokerdPosition
|
||||||
from ._forms import open_form_input_handling
|
from ._forms import open_form_input_handling
|
||||||
|
@ -639,63 +642,21 @@ async def open_order_mode(
|
||||||
pp_tracker.hide_info()
|
pp_tracker.hide_info()
|
||||||
|
|
||||||
# setup order mode sidepane widgets
|
# setup order mode sidepane widgets
|
||||||
form = chart.sidepane
|
form: FieldsForm = chart.sidepane
|
||||||
vbox = form.vbox
|
form.vbox.setSpacing(
|
||||||
|
|
||||||
from textwrap import dedent
|
|
||||||
|
|
||||||
from PyQt5.QtCore import Qt
|
|
||||||
|
|
||||||
from ._style import _font, _font_small
|
|
||||||
from ..calc import humanize
|
|
||||||
|
|
||||||
feed_label = FormatLabel(
|
|
||||||
fmt_str=dedent("""
|
|
||||||
actor: **{actor_name}**\n
|
|
||||||
|_ @**{host}:{port}**\n
|
|
||||||
|_ throttle_hz: **{throttle_rate}**\n
|
|
||||||
|_ streams: **{symbols}**\n
|
|
||||||
|_ shm: **{shm}**\n
|
|
||||||
"""),
|
|
||||||
font=_font.font,
|
|
||||||
font_size=_font_small.px_size,
|
|
||||||
font_color='default_lightest',
|
|
||||||
)
|
|
||||||
|
|
||||||
form.feed_label = feed_label
|
|
||||||
|
|
||||||
# add feed info label to top
|
|
||||||
vbox.insertWidget(
|
|
||||||
0,
|
|
||||||
feed_label,
|
|
||||||
alignment=Qt.AlignBottom,
|
|
||||||
)
|
|
||||||
# vbox.setAlignment(feed_label, Qt.AlignBottom)
|
|
||||||
# vbox.setAlignment(Qt.AlignBottom)
|
|
||||||
_ = chart.height() - (
|
|
||||||
form.height() +
|
|
||||||
form.fill_bar.height()
|
|
||||||
# feed_label.height()
|
|
||||||
)
|
|
||||||
vbox.setSpacing(
|
|
||||||
int((1 + 5/8)*_font.px_size)
|
int((1 + 5/8)*_font.px_size)
|
||||||
)
|
)
|
||||||
|
|
||||||
# fill in brokerd feed info
|
from ._feedstatus import mk_feed_label
|
||||||
host, port = feed.portal.channel.raddr
|
|
||||||
if host == '127.0.0.1':
|
feed_label = mk_feed_label(
|
||||||
host = 'localhost'
|
form,
|
||||||
mpshm = feed.shm._shm
|
feed,
|
||||||
shmstr = f'{humanize(mpshm.size)}'
|
chart,
|
||||||
form.feed_label.format(
|
|
||||||
actor_name=feed.portal.channel.uid[0],
|
|
||||||
host=host,
|
|
||||||
port=port,
|
|
||||||
symbols=len(feed.symbols),
|
|
||||||
shm=shmstr,
|
|
||||||
throttle_rate=feed.throttle_rate,
|
|
||||||
)
|
)
|
||||||
|
|
||||||
|
# XXX: we set this because?
|
||||||
|
form.feed_label = feed_label
|
||||||
order_pane = SettingsPane(
|
order_pane = SettingsPane(
|
||||||
form=form,
|
form=form,
|
||||||
# XXX: ugh, so hideous...
|
# XXX: ugh, so hideous...
|
||||||
|
@ -706,6 +667,11 @@ async def open_order_mode(
|
||||||
)
|
)
|
||||||
order_pane.set_accounts(list(trackers.keys()))
|
order_pane.set_accounts(list(trackers.keys()))
|
||||||
|
|
||||||
|
form.vbox.addWidget(
|
||||||
|
feed_label,
|
||||||
|
alignment=Qt.AlignBottom,
|
||||||
|
)
|
||||||
|
|
||||||
# update pp icons
|
# update pp icons
|
||||||
for name, tracker in trackers.items():
|
for name, tracker in trackers.items():
|
||||||
order_pane.update_account_icons({name: tracker.live_pp})
|
order_pane.update_account_icons({name: tracker.live_pp})
|
||||||
|
|
|
@ -15,3 +15,7 @@
|
||||||
|
|
||||||
# ``trimeter`` for asysnc history fetching
|
# ``trimeter`` for asysnc history fetching
|
||||||
-e git+https://github.com/python-trio/trimeter.git@master#egg=trimeter
|
-e git+https://github.com/python-trio/trimeter.git@master#egg=trimeter
|
||||||
|
|
||||||
|
|
||||||
|
# ``asyncvnc`` for sending interactions to ib-gw inside docker
|
||||||
|
-e git+https://github.com/pikers/asyncvnc.git@vid_passthrough#egg=asyncvnc
|
||||||
|
|
23
setup.py
23
setup.py
|
@ -57,6 +57,7 @@ setup(
|
||||||
# from github currently (see requirements.txt)
|
# from github currently (see requirements.txt)
|
||||||
# 'trimeter', # not released yet..
|
# 'trimeter', # not released yet..
|
||||||
# 'tractor',
|
# 'tractor',
|
||||||
|
# asyncvnc,
|
||||||
|
|
||||||
# brokers
|
# brokers
|
||||||
'asks==2.4.8',
|
'asks==2.4.8',
|
||||||
|
@ -71,32 +72,34 @@ setup(
|
||||||
|
|
||||||
# UI
|
# UI
|
||||||
'PyQt5',
|
'PyQt5',
|
||||||
'pyqtgraph',
|
# 'pyqtgraph', from our fork see reqs.txt
|
||||||
'qdarkstyle >= 3.0.2',
|
'qdarkstyle >= 3.0.2', # themeing
|
||||||
# fuzzy search
|
'fuzzywuzzy[speedup]', # fuzzy search
|
||||||
'fuzzywuzzy[speedup]',
|
|
||||||
|
|
||||||
# tsdbs
|
# tsdbs
|
||||||
'pymarketstore',
|
# anyio-marketstore # from gh see reqs.txt
|
||||||
],
|
],
|
||||||
extras_require={
|
extras_require={
|
||||||
|
|
||||||
# serialization
|
|
||||||
'tsdb': [
|
'tsdb': [
|
||||||
'docker',
|
'docker',
|
||||||
],
|
],
|
||||||
|
|
||||||
},
|
},
|
||||||
tests_require=['pytest'],
|
tests_require=['pytest'],
|
||||||
python_requires=">=3.9", # literally for ``datetime.datetime.fromisoformat``...
|
python_requires=">=3.10",
|
||||||
keywords=["async", "trading", "finance", "quant", "charting"],
|
keywords=[
|
||||||
|
"async",
|
||||||
|
"trading",
|
||||||
|
"finance",
|
||||||
|
"quant",
|
||||||
|
"charting",
|
||||||
|
],
|
||||||
classifiers=[
|
classifiers=[
|
||||||
'Development Status :: 3 - Alpha',
|
'Development Status :: 3 - Alpha',
|
||||||
'License :: OSI Approved :: ',
|
'License :: OSI Approved :: ',
|
||||||
'Operating System :: POSIX :: Linux',
|
'Operating System :: POSIX :: Linux',
|
||||||
"Programming Language :: Python :: Implementation :: CPython",
|
"Programming Language :: Python :: Implementation :: CPython",
|
||||||
"Programming Language :: Python :: 3 :: Only",
|
"Programming Language :: Python :: 3 :: Only",
|
||||||
"Programming Language :: Python :: 3.9",
|
|
||||||
"Programming Language :: Python :: 3.10",
|
"Programming Language :: Python :: 3.10",
|
||||||
'Intended Audience :: Financial and Insurance Industry',
|
'Intended Audience :: Financial and Insurance Industry',
|
||||||
'Intended Audience :: Science/Research',
|
'Intended Audience :: Science/Research',
|
||||||
|
|
Loading…
Reference in New Issue