2022-06-08 15:25:17 +00:00
|
|
|
# piker: trading gear for hackers
|
|
|
|
# Copyright (C) Tyler Goodlet (in stewardship for pikers)
|
|
|
|
|
|
|
|
# This program is free software: you can redistribute it and/or modify
|
|
|
|
# it under the terms of the GNU Affero General Public License as published by
|
|
|
|
# the Free Software Foundation, either version 3 of the License, or
|
|
|
|
# (at your option) any later version.
|
|
|
|
|
|
|
|
# This program is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
# GNU Affero General Public License for more details.
|
|
|
|
|
|
|
|
# You should have received a copy of the GNU Affero General Public License
|
|
|
|
|
|
|
|
# along with this program. If not, see <https://www.gnu.org/licenses/>.
|
|
|
|
'''
|
2022-06-14 18:58:21 +00:00
|
|
|
Personal/Private position parsing, calculating, summarizing in a way
|
2022-06-08 15:25:17 +00:00
|
|
|
that doesn't try to cuk most humans who prefer to not lose their moneys..
|
2022-06-14 18:58:21 +00:00
|
|
|
(looking at you `ib` and dirt-bird friends)
|
2022-06-08 15:25:17 +00:00
|
|
|
|
|
|
|
'''
|
2022-06-13 18:11:37 +00:00
|
|
|
from contextlib import contextmanager as cm
|
2022-07-26 15:27:38 +00:00
|
|
|
from pprint import pformat
|
2022-06-13 18:11:37 +00:00
|
|
|
import os
|
|
|
|
from os import path
|
2022-06-23 20:11:50 +00:00
|
|
|
from math import copysign
|
2022-06-17 19:41:17 +00:00
|
|
|
import re
|
2022-06-23 18:59:47 +00:00
|
|
|
import time
|
2022-06-08 15:25:17 +00:00
|
|
|
from typing import (
|
2022-06-13 18:11:37 +00:00
|
|
|
Any,
|
2022-06-08 15:25:17 +00:00
|
|
|
Optional,
|
|
|
|
Union,
|
|
|
|
)
|
|
|
|
|
2022-06-18 19:30:52 +00:00
|
|
|
import pendulum
|
|
|
|
from pendulum import datetime, now
|
2022-06-22 19:41:26 +00:00
|
|
|
import tomli
|
2022-06-13 18:11:37 +00:00
|
|
|
import toml
|
2022-06-08 15:25:17 +00:00
|
|
|
|
|
|
|
from . import config
|
2022-06-13 18:11:37 +00:00
|
|
|
from .brokers import get_brokermod
|
2022-06-08 15:25:17 +00:00
|
|
|
from .clearing._messages import BrokerdPosition, Status
|
|
|
|
from .data._source import Symbol
|
2022-06-13 18:11:37 +00:00
|
|
|
from .log import get_logger
|
2022-07-19 12:26:28 +00:00
|
|
|
from .data.types import Struct
|
2022-06-13 18:11:37 +00:00
|
|
|
|
|
|
|
log = get_logger(__name__)
|
|
|
|
|
|
|
|
|
|
|
|
@cm
|
|
|
|
def open_trade_ledger(
|
|
|
|
broker: str,
|
|
|
|
account: str,
|
|
|
|
|
|
|
|
) -> str:
|
|
|
|
'''
|
|
|
|
Indempotently create and read in a trade log file from the
|
|
|
|
``<configuration_dir>/ledgers/`` directory.
|
|
|
|
|
|
|
|
Files are named per broker account of the form
|
|
|
|
``<brokername>_<accountname>.toml``. The ``accountname`` here is the
|
|
|
|
name as defined in the user's ``brokers.toml`` config.
|
|
|
|
|
|
|
|
'''
|
|
|
|
ldir = path.join(config._config_dir, 'ledgers')
|
|
|
|
if not path.isdir(ldir):
|
|
|
|
os.makedirs(ldir)
|
|
|
|
|
|
|
|
fname = f'trades_{broker}_{account}.toml'
|
|
|
|
tradesfile = path.join(ldir, fname)
|
|
|
|
|
|
|
|
if not path.isfile(tradesfile):
|
|
|
|
log.info(
|
|
|
|
f'Creating new local trades ledger: {tradesfile}'
|
|
|
|
)
|
|
|
|
with open(tradesfile, 'w') as cf:
|
|
|
|
pass # touch
|
2022-06-22 19:41:26 +00:00
|
|
|
with open(tradesfile, 'rb') as cf:
|
2022-06-23 18:59:47 +00:00
|
|
|
start = time.time()
|
2022-06-22 19:41:26 +00:00
|
|
|
ledger = tomli.load(cf)
|
2022-06-23 18:59:47 +00:00
|
|
|
print(f'Ledger load took {time.time() - start}s')
|
2022-06-16 14:52:43 +00:00
|
|
|
cpy = ledger.copy()
|
2022-06-13 18:11:37 +00:00
|
|
|
|
2022-07-22 23:01:23 +00:00
|
|
|
try:
|
|
|
|
yield cpy
|
|
|
|
finally:
|
|
|
|
if cpy != ledger:
|
|
|
|
# TODO: show diff output?
|
|
|
|
# https://stackoverflow.com/questions/12956957/print-diff-of-python-dictionaries
|
|
|
|
print(f'Updating ledger for {tradesfile}:\n')
|
|
|
|
ledger.update(cpy)
|
2022-07-21 19:28:04 +00:00
|
|
|
|
2022-07-22 23:01:23 +00:00
|
|
|
# we write on close the mutated ledger data
|
|
|
|
with open(tradesfile, 'w') as cf:
|
|
|
|
toml.dump(ledger, cf)
|
2022-06-10 17:28:34 +00:00
|
|
|
|
|
|
|
|
2022-07-11 00:00:12 +00:00
|
|
|
class Transaction(Struct, frozen=True):
|
2022-06-22 19:41:26 +00:00
|
|
|
# TODO: should this be ``.to`` (see below)?
|
|
|
|
fqsn: str
|
|
|
|
|
2022-06-14 18:58:21 +00:00
|
|
|
tid: Union[str, int] # unique transaction id
|
2022-06-10 17:28:34 +00:00
|
|
|
size: float
|
|
|
|
price: float
|
|
|
|
cost: float # commisions or other additional costs
|
2022-06-18 19:30:52 +00:00
|
|
|
dt: datetime
|
|
|
|
expiry: Optional[datetime] = None
|
2022-06-13 18:11:37 +00:00
|
|
|
|
2022-06-10 17:28:34 +00:00
|
|
|
# optional key normally derived from the broker
|
|
|
|
# backend which ensures the instrument-symbol this record
|
|
|
|
# is for is truly unique.
|
2022-06-14 18:58:21 +00:00
|
|
|
bsuid: Optional[Union[str, int]] = None
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-06-22 19:41:26 +00:00
|
|
|
# optional fqsn for the source "asset"/money symbol?
|
|
|
|
# from: Optional[str] = None
|
|
|
|
|
2022-06-08 15:25:17 +00:00
|
|
|
|
|
|
|
class Position(Struct):
|
|
|
|
'''
|
2022-06-17 19:41:17 +00:00
|
|
|
Basic pp (personal/piker position) model with attached clearing
|
|
|
|
transaction history.
|
2022-06-08 15:25:17 +00:00
|
|
|
|
|
|
|
'''
|
|
|
|
symbol: Symbol
|
|
|
|
|
2022-06-18 19:30:52 +00:00
|
|
|
# can be +ve or -ve for long/short
|
2022-06-08 15:25:17 +00:00
|
|
|
size: float
|
2022-06-18 19:30:52 +00:00
|
|
|
|
|
|
|
# "breakeven price" above or below which pnl moves above and below
|
|
|
|
# zero for the entirety of the current "trade state".
|
|
|
|
be_price: float
|
|
|
|
|
|
|
|
# unique backend symbol id
|
2022-06-14 18:58:21 +00:00
|
|
|
bsuid: str
|
2022-06-08 15:25:17 +00:00
|
|
|
|
|
|
|
# ordered record of known constituent trade messages
|
2022-06-17 19:41:17 +00:00
|
|
|
clears: dict[
|
2022-06-14 18:58:21 +00:00
|
|
|
Union[str, int, Status], # trade id
|
2022-06-19 20:30:08 +00:00
|
|
|
dict[str, Any], # transaction history summaries
|
2022-06-14 18:58:21 +00:00
|
|
|
] = {}
|
|
|
|
|
2022-06-18 19:30:52 +00:00
|
|
|
expiry: Optional[datetime] = None
|
|
|
|
|
2022-06-17 19:41:17 +00:00
|
|
|
def to_dict(self) -> dict:
|
2022-06-10 17:28:34 +00:00
|
|
|
return {
|
|
|
|
f: getattr(self, f)
|
|
|
|
for f in self.__struct_fields__
|
|
|
|
}
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
def to_pretoml(self) -> tuple[str, dict]:
|
2022-06-19 20:30:08 +00:00
|
|
|
'''
|
|
|
|
Prep this position's data contents for export to toml including
|
|
|
|
re-structuring of the ``.clears`` table to an array of
|
|
|
|
inline-subtables for better ``pps.toml`` compactness.
|
|
|
|
|
|
|
|
'''
|
2022-06-17 19:41:17 +00:00
|
|
|
d = self.to_dict()
|
|
|
|
clears = d.pop('clears')
|
2022-06-18 19:30:52 +00:00
|
|
|
expiry = d.pop('expiry')
|
2022-06-18 22:30:53 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# TODO: we need to figure out how to have one top level
|
|
|
|
# listing venue here even when the backend isn't providing
|
|
|
|
# it via the trades ledger..
|
|
|
|
# drop symbol obj in serialized form
|
|
|
|
s = d.pop('symbol')
|
|
|
|
fqsn = s.front_fqsn()
|
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
size = d.pop('size')
|
|
|
|
be_price = d.pop('be_price')
|
|
|
|
d['size'], d['be_price'] = self.audit_sizing(size, be_price)
|
|
|
|
|
2022-07-25 17:03:22 +00:00
|
|
|
if self.expiry is None:
|
|
|
|
d.pop('expiry', None)
|
|
|
|
elif expiry:
|
2022-06-18 19:30:52 +00:00
|
|
|
d['expiry'] = str(expiry)
|
2022-06-17 19:41:17 +00:00
|
|
|
|
2022-07-25 17:03:22 +00:00
|
|
|
toml_clears_list = []
|
2022-07-27 12:43:09 +00:00
|
|
|
for tid, data in sorted(
|
|
|
|
list(clears.items()),
|
|
|
|
|
|
|
|
# sort by datetime
|
|
|
|
key=lambda item: item[1]['dt'],
|
|
|
|
):
|
2022-06-18 22:30:53 +00:00
|
|
|
inline_table = toml.TomlDecoder().get_empty_inline_table()
|
|
|
|
|
2022-07-27 12:43:09 +00:00
|
|
|
inline_table['dt'] = data['dt']
|
|
|
|
|
|
|
|
# insert optional clear fields in column order
|
|
|
|
for k in ['ppu', 'accum_size']:
|
|
|
|
val = data.get(k)
|
|
|
|
if val:
|
|
|
|
inline_table[k] = val
|
|
|
|
|
|
|
|
# insert required fields
|
|
|
|
for k in ['price', 'size', 'cost']:
|
|
|
|
inline_table[k] = data[k]
|
2022-06-17 19:41:17 +00:00
|
|
|
|
2022-07-27 12:43:09 +00:00
|
|
|
inline_table['tid'] = tid
|
2022-07-25 17:03:22 +00:00
|
|
|
toml_clears_list.append(inline_table)
|
2022-06-17 19:41:17 +00:00
|
|
|
|
2022-07-25 17:03:22 +00:00
|
|
|
d['clears'] = toml_clears_list
|
2022-06-19 20:30:08 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
return fqsn, d
|
2022-06-17 19:41:17 +00:00
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
def audit_sizing(
|
|
|
|
self,
|
|
|
|
size: Optional[float] = None,
|
|
|
|
be_price: Optional[float] = None,
|
|
|
|
|
|
|
|
) -> tuple[float, float]:
|
|
|
|
'''
|
|
|
|
Audit either the `.size` and `.be_price` values or equvialent
|
|
|
|
passed in values against the clears table calculations and
|
|
|
|
return the calc-ed values if they differ and log warnings to
|
|
|
|
console.
|
|
|
|
|
|
|
|
'''
|
|
|
|
size = size or self.size
|
|
|
|
be_price = be_price or self.be_price
|
|
|
|
csize = self.calc_size()
|
2022-07-26 15:27:38 +00:00
|
|
|
cbe_price = self.calc_ppu()
|
2022-07-25 15:57:57 +00:00
|
|
|
|
|
|
|
if size != csize:
|
|
|
|
log.warning(f'size != calculated size: {size} != {csize}')
|
|
|
|
size = csize
|
|
|
|
|
|
|
|
if be_price != cbe_price:
|
|
|
|
log.warning(
|
|
|
|
f'be_price != calculated be_price: {be_price} != {cbe_price}'
|
|
|
|
)
|
|
|
|
be_price = cbe_price
|
|
|
|
|
|
|
|
return size, be_price
|
|
|
|
|
2022-06-08 15:25:17 +00:00
|
|
|
def update_from_msg(
|
|
|
|
self,
|
|
|
|
msg: BrokerdPosition,
|
|
|
|
|
|
|
|
) -> None:
|
|
|
|
|
|
|
|
# XXX: better place to do this?
|
|
|
|
symbol = self.symbol
|
|
|
|
|
|
|
|
lot_size_digits = symbol.lot_size_digits
|
2022-06-18 19:30:52 +00:00
|
|
|
be_price, size = (
|
2022-06-08 15:25:17 +00:00
|
|
|
round(
|
|
|
|
msg['avg_price'],
|
|
|
|
ndigits=symbol.tick_size_digits
|
|
|
|
),
|
|
|
|
round(
|
|
|
|
msg['size'],
|
|
|
|
ndigits=lot_size_digits
|
|
|
|
),
|
|
|
|
)
|
|
|
|
|
2022-06-18 19:30:52 +00:00
|
|
|
self.be_price = be_price
|
2022-06-08 15:25:17 +00:00
|
|
|
self.size = size
|
|
|
|
|
|
|
|
@property
|
|
|
|
def dsize(self) -> float:
|
|
|
|
'''
|
|
|
|
The "dollar" size of the pp, normally in trading (fiat) unit
|
|
|
|
terms.
|
|
|
|
|
|
|
|
'''
|
2022-06-18 19:30:52 +00:00
|
|
|
return self.be_price * self.size
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
# TODO: idea: "real LIFO" dynamic positioning.
|
|
|
|
# - when a trade takes place where the pnl for
|
|
|
|
# the (set of) trade(s) is below the breakeven price
|
|
|
|
# it may be that the trader took a +ve pnl on a short(er)
|
|
|
|
# term trade in the same account.
|
|
|
|
# - in this case we could recalc the be price to
|
|
|
|
# be reverted back to it's prior value before the nearest term
|
|
|
|
# trade was opened.?
|
|
|
|
# def lifo_price() -> float:
|
|
|
|
# ...
|
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
def calc_ppu(
|
2022-06-08 15:25:17 +00:00
|
|
|
self,
|
2022-07-25 15:57:57 +00:00
|
|
|
# include transaction cost in breakeven price
|
|
|
|
# and presume the worst case of the same cost
|
|
|
|
# to exit this transaction (even though in reality
|
|
|
|
# it will be dynamic based on exit stratetgy).
|
|
|
|
cost_scalar: float = 2,
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
) -> float:
|
2022-07-26 15:27:38 +00:00
|
|
|
'''
|
|
|
|
Compute the "price-per-unit" price for the given non-zero sized
|
|
|
|
rolling position.
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
The recurrence relation which computes this (exponential) mean
|
|
|
|
per new clear which **increases** the accumulative postiion size
|
|
|
|
is:
|
|
|
|
|
|
|
|
ppu[-1] = (
|
|
|
|
ppu[-2] * accum_size[-2]
|
|
|
|
+
|
|
|
|
ppu[-1] * size
|
|
|
|
) / accum_size[-1]
|
|
|
|
|
|
|
|
where `cost_basis` for the current step is simply the price
|
|
|
|
* size of the most recent clearing transaction.
|
|
|
|
|
|
|
|
'''
|
|
|
|
asize_h: list[float] = [] # historical accumulative size
|
|
|
|
ppu_h: list[float] = [] # historical price-per-unit
|
|
|
|
|
|
|
|
clears = list(self.clears.items())
|
|
|
|
|
|
|
|
for i, (tid, entry) in enumerate(clears):
|
2022-07-22 23:01:23 +00:00
|
|
|
|
|
|
|
clear_size = entry['size']
|
|
|
|
clear_price = entry['price']
|
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
last_accum_size = asize_h[-1] if asize_h else 0
|
|
|
|
accum_size = last_accum_size + clear_size
|
|
|
|
accum_sign = copysign(1, accum_size)
|
|
|
|
|
|
|
|
sign_change: bool = False
|
|
|
|
|
|
|
|
if accum_size == 0:
|
|
|
|
ppu_h.append(0)
|
|
|
|
asize_h.append(0)
|
|
|
|
continue
|
|
|
|
|
|
|
|
# test if the pp somehow went "passed" a net zero size state
|
|
|
|
# resulting in a change of the "sign" of the size (+ve for
|
|
|
|
# long, -ve for short).
|
|
|
|
sign_change = (
|
|
|
|
copysign(1, last_accum_size) + accum_sign == 0
|
|
|
|
and last_accum_size != 0
|
|
|
|
)
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# since we passed the net-zero-size state the new size
|
|
|
|
# after sum should be the remaining size the new
|
|
|
|
# "direction" (aka, long vs. short) for this clear.
|
|
|
|
if sign_change:
|
|
|
|
clear_size = accum_size
|
|
|
|
abs_diff = abs(accum_size)
|
|
|
|
asize_h.append(0)
|
|
|
|
ppu_h.append(0)
|
|
|
|
|
|
|
|
else:
|
|
|
|
# old size minus the new size gives us size diff with
|
|
|
|
# +ve -> increase in pp size
|
|
|
|
# -ve -> decrease in pp size
|
|
|
|
abs_diff = abs(accum_size) - abs(last_accum_size)
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
# XXX: LIFO breakeven price update. only an increaze in size
|
|
|
|
# of the position contributes the breakeven price,
|
|
|
|
# a decrease does not (i.e. the position is being made
|
|
|
|
# smaller).
|
2022-07-26 15:27:38 +00:00
|
|
|
# abs_clear_size = abs(clear_size)
|
|
|
|
abs_new_size = abs(accum_size)
|
|
|
|
|
|
|
|
if abs_diff > 0:
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
cost_basis = (
|
|
|
|
# cost basis for this clear
|
2022-07-22 23:01:23 +00:00
|
|
|
clear_price * abs(clear_size)
|
|
|
|
+
|
|
|
|
# transaction cost
|
2022-07-26 15:27:38 +00:00
|
|
|
accum_sign * cost_scalar * entry['cost']
|
2022-07-22 23:01:23 +00:00
|
|
|
)
|
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
if asize_h:
|
|
|
|
size_last = abs(asize_h[-1])
|
|
|
|
cb_last = ppu_h[-1] * size_last
|
|
|
|
ppu = (cost_basis + cb_last) / abs_new_size
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
else:
|
|
|
|
ppu = cost_basis / abs_new_size
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
ppu_h.append(ppu)
|
|
|
|
asize_h.append(accum_size)
|
2022-07-22 23:01:23 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
else:
|
|
|
|
# on "exit" clears from a given direction,
|
|
|
|
# only the size changes not the price-per-unit
|
|
|
|
# need to be updated since the ppu remains constant
|
|
|
|
# and gets weighted by the new size.
|
|
|
|
asize_h.append(accum_size)
|
|
|
|
ppu_h.append(ppu_h[-1])
|
|
|
|
|
|
|
|
return ppu_h[-1] if ppu_h else 0
|
2022-07-22 23:01:23 +00:00
|
|
|
|
|
|
|
def calc_size(self) -> float:
|
|
|
|
size: float = 0
|
|
|
|
for tid, entry in self.clears.items():
|
|
|
|
size += entry['size']
|
|
|
|
return size
|
|
|
|
|
2022-06-23 18:59:47 +00:00
|
|
|
def minimize_clears(
|
|
|
|
self,
|
|
|
|
|
|
|
|
) -> dict[str, dict]:
|
|
|
|
'''
|
|
|
|
Minimize the position's clears entries by removing
|
|
|
|
all transactions before the last net zero size to avoid
|
|
|
|
unecessary history irrelevant to the current pp state.
|
|
|
|
|
|
|
|
'''
|
2022-07-26 15:27:38 +00:00
|
|
|
size: float = 0
|
|
|
|
clears_since_zero: list[tuple(str, dict)] = []
|
2022-06-23 18:59:47 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# TODO: we might just want to always do this when iterating
|
|
|
|
# a ledger? keep a state of the last net-zero and only do the
|
|
|
|
# full iterate when no state was stashed?
|
|
|
|
|
|
|
|
# scan for the last "net zero" position by iterating
|
|
|
|
# transactions until the next net-zero size, rinse, repeat.
|
|
|
|
for tid, clear in self.clears.items():
|
|
|
|
size += clear['size']
|
|
|
|
clears_since_zero.append((tid, clear))
|
2022-06-23 18:59:47 +00:00
|
|
|
|
|
|
|
if size == 0:
|
2022-07-26 15:27:38 +00:00
|
|
|
clears_since_zero.clear()
|
2022-06-23 18:59:47 +00:00
|
|
|
|
|
|
|
self.clears = dict(clears_since_zero)
|
|
|
|
return self.clears
|
|
|
|
|
2022-07-27 14:26:50 +00:00
|
|
|
def add_clear(
|
|
|
|
self,
|
|
|
|
t: Transaction,
|
|
|
|
) -> dict:
|
|
|
|
'''
|
|
|
|
Update clearing table and populate rolling ppu and accumulative
|
|
|
|
size in both the clears entry and local attrs state.
|
|
|
|
|
|
|
|
'''
|
|
|
|
clear = self.clears[t.tid] = {
|
|
|
|
'cost': t.cost,
|
|
|
|
'price': t.price,
|
|
|
|
'size': t.size,
|
|
|
|
'dt': str(t.dt),
|
|
|
|
}
|
|
|
|
|
|
|
|
# TODO: compute these incrementally instead
|
|
|
|
# of re-looping through each time resulting in O(n**2)
|
|
|
|
# behaviour..
|
|
|
|
# compute these **after** adding the entry
|
|
|
|
# in order to make the recurrence relation math work
|
|
|
|
# inside ``.calc_size()``.
|
|
|
|
self.size = clear['accum_size'] = self.calc_size()
|
|
|
|
self.be_price = clear['ppu'] = self.calc_ppu()
|
|
|
|
|
|
|
|
return clear
|
|
|
|
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-07-18 16:23:02 +00:00
|
|
|
class PpTable(Struct):
|
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
brokername: str
|
2022-07-27 12:43:09 +00:00
|
|
|
acctid: str
|
2022-07-18 16:23:02 +00:00
|
|
|
pps: dict[str, Position]
|
|
|
|
conf: Optional[dict] = {}
|
|
|
|
|
|
|
|
def update_from_trans(
|
|
|
|
self,
|
|
|
|
trans: dict[str, Transaction],
|
2022-07-22 23:01:23 +00:00
|
|
|
cost_scalar: float = 2,
|
|
|
|
|
2022-07-18 16:23:02 +00:00
|
|
|
) -> dict[str, Position]:
|
|
|
|
|
|
|
|
pps = self.pps
|
2022-07-19 12:26:28 +00:00
|
|
|
updated: dict[str, Position] = {}
|
|
|
|
|
2022-07-18 16:23:02 +00:00
|
|
|
# lifo update all pps from records
|
2022-07-26 15:27:38 +00:00
|
|
|
for tid, t in trans.items():
|
2022-07-18 16:23:02 +00:00
|
|
|
|
|
|
|
pp = pps.setdefault(
|
2022-07-26 15:27:38 +00:00
|
|
|
t.bsuid,
|
2022-07-18 16:23:02 +00:00
|
|
|
|
|
|
|
# if no existing pp, allocate fresh one.
|
|
|
|
Position(
|
|
|
|
Symbol.from_fqsn(
|
2022-07-26 15:27:38 +00:00
|
|
|
t.fqsn,
|
2022-07-18 16:23:02 +00:00
|
|
|
info={},
|
|
|
|
),
|
|
|
|
size=0.0,
|
|
|
|
be_price=0.0,
|
2022-07-26 15:27:38 +00:00
|
|
|
bsuid=t.bsuid,
|
|
|
|
expiry=t.expiry,
|
2022-07-18 16:23:02 +00:00
|
|
|
)
|
|
|
|
)
|
|
|
|
|
|
|
|
# don't do updates for ledger records we already have
|
|
|
|
# included in the current pps state.
|
2022-07-26 15:27:38 +00:00
|
|
|
if t.tid in pp.clears:
|
2022-07-18 16:23:02 +00:00
|
|
|
# NOTE: likely you'll see repeats of the same
|
|
|
|
# ``Transaction`` passed in here if/when you are restarting
|
|
|
|
# a ``brokerd.ib`` where the API will re-report trades from
|
|
|
|
# the current session, so we need to make sure we don't
|
|
|
|
# "double count" these in pp calculations.
|
|
|
|
continue
|
|
|
|
|
2022-07-27 14:26:50 +00:00
|
|
|
# update clearing table
|
|
|
|
pp.add_clear(t)
|
2022-07-26 15:27:38 +00:00
|
|
|
updated[t.bsuid] = pp
|
2022-07-19 12:26:28 +00:00
|
|
|
|
2022-07-25 17:03:22 +00:00
|
|
|
# minimize clears tables and update sizing.
|
|
|
|
for bsuid, pp in updated.items():
|
|
|
|
pp.size, pp.be_price = pp.audit_sizing()
|
|
|
|
|
2022-07-19 12:26:28 +00:00
|
|
|
return updated
|
2022-07-18 16:23:02 +00:00
|
|
|
|
|
|
|
def dump_active(
|
|
|
|
self,
|
|
|
|
) -> tuple[
|
2022-07-26 15:27:38 +00:00
|
|
|
dict[str, Position],
|
2022-07-18 16:23:02 +00:00
|
|
|
dict[str, Position]
|
|
|
|
]:
|
|
|
|
'''
|
|
|
|
Iterate all tabulated positions, render active positions to
|
|
|
|
a ``dict`` format amenable to serialization (via TOML) and drop
|
|
|
|
from state (``.pps``) as well as return in a ``dict`` all
|
|
|
|
``Position``s which have recently closed.
|
|
|
|
|
|
|
|
'''
|
|
|
|
# NOTE: newly closed position are also important to report/return
|
|
|
|
# since a consumer, like an order mode UI ;), might want to react
|
2022-07-26 15:27:38 +00:00
|
|
|
# based on the closure (for example removing the breakeven line
|
|
|
|
# and clearing the entry from any lists/monitors).
|
2022-07-18 16:23:02 +00:00
|
|
|
closed_pp_objs: dict[str, Position] = {}
|
2022-07-26 15:27:38 +00:00
|
|
|
open_pp_objs: dict[str, Position] = {}
|
2022-07-18 16:23:02 +00:00
|
|
|
|
|
|
|
pp_objs = self.pps
|
|
|
|
for bsuid in list(pp_objs):
|
|
|
|
pp = pp_objs[bsuid]
|
|
|
|
|
|
|
|
# XXX: debug hook for size mismatches
|
2022-07-21 14:12:51 +00:00
|
|
|
# qqqbsuid = 320227571
|
|
|
|
# if bsuid == qqqbsuid:
|
2022-07-18 16:23:02 +00:00
|
|
|
# breakpoint()
|
|
|
|
|
2022-07-27 12:43:09 +00:00
|
|
|
pp.size, pp.be_price = pp.audit_sizing()
|
2022-07-18 16:23:02 +00:00
|
|
|
|
|
|
|
if (
|
2022-07-19 12:26:28 +00:00
|
|
|
# "net-zero" is a "closed" position
|
2022-07-27 12:43:09 +00:00
|
|
|
pp.size == 0
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-19 12:26:28 +00:00
|
|
|
# time-expired pps (normally derivatives) are "closed"
|
2022-07-18 16:23:02 +00:00
|
|
|
or (pp.expiry and pp.expiry < now())
|
|
|
|
):
|
2022-07-19 12:26:28 +00:00
|
|
|
# for expired cases
|
2022-07-18 16:23:02 +00:00
|
|
|
pp.size = 0
|
|
|
|
|
2022-07-19 12:26:28 +00:00
|
|
|
# NOTE: we DO NOT pop the pp here since it can still be
|
|
|
|
# used to check for duplicate clears that may come in as
|
|
|
|
# new transaction from some backend API and need to be
|
|
|
|
# ignored; the closed positions won't be written to the
|
2022-07-25 17:03:22 +00:00
|
|
|
# ``pps.toml`` since ``pp_active_entries`` above is what's
|
2022-07-19 12:26:28 +00:00
|
|
|
# written.
|
2022-07-26 15:27:38 +00:00
|
|
|
closed_pp_objs[bsuid] = pp
|
2022-07-18 16:23:02 +00:00
|
|
|
|
|
|
|
else:
|
2022-07-26 15:27:38 +00:00
|
|
|
open_pp_objs[bsuid] = pp
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
return open_pp_objs, closed_pp_objs
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
def to_toml(
|
|
|
|
self,
|
|
|
|
) -> dict[str, Any]:
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
active, closed = self.dump_active()
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# ONLY dict-serialize all active positions; those that are closed
|
|
|
|
# we don't store in the ``pps.toml``.
|
|
|
|
to_toml_dict = {}
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
for bsuid, pos in active.items():
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# keep the minimal amount of clears that make up this
|
|
|
|
# position since the last net-zero state.
|
|
|
|
pos.minimize_clears()
|
2022-06-10 21:50:29 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# serialize to pre-toml form
|
|
|
|
fqsn, asdict = pos.to_pretoml()
|
|
|
|
log.info(f'Updating active pp: {fqsn}')
|
2022-06-10 17:28:34 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# XXX: ugh, it's cuz we push the section under
|
|
|
|
# the broker name.. maybe we need to rethink this?
|
|
|
|
brokerless_key = fqsn.removeprefix(f'{self.brokername}.')
|
|
|
|
to_toml_dict[brokerless_key] = asdict
|
|
|
|
|
|
|
|
return to_toml_dict
|
2022-06-10 21:50:29 +00:00
|
|
|
|
2022-07-27 12:43:09 +00:00
|
|
|
def write_config(self) -> None:
|
|
|
|
'''
|
|
|
|
Write the current position table to the user's ``pps.toml``.
|
|
|
|
|
|
|
|
'''
|
|
|
|
# TODO: show diff output?
|
|
|
|
# https://stackoverflow.com/questions/12956957/print-diff-of-python-dictionaries
|
|
|
|
print(f'Updating ``pps.toml`` for {path}:\n')
|
|
|
|
|
|
|
|
# active, closed_pp_objs = table.dump_active()
|
|
|
|
pp_entries = self.to_toml()
|
|
|
|
self.conf[self.brokername][self.acctid] = pp_entries
|
|
|
|
|
|
|
|
# TODO: why tf haven't they already done this for inline
|
|
|
|
# tables smh..
|
|
|
|
enc = PpsEncoder(preserve=True)
|
|
|
|
# table_bs_type = type(toml.TomlDecoder().get_empty_inline_table())
|
|
|
|
enc.dump_funcs[
|
|
|
|
toml.decoder.InlineTableDict
|
|
|
|
] = enc.dump_inline_table
|
|
|
|
|
|
|
|
config.write(
|
|
|
|
self.conf,
|
|
|
|
'pps',
|
|
|
|
encoder=enc,
|
|
|
|
)
|
|
|
|
|
2022-06-14 18:58:21 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
def load_pps_from_ledger(
|
2022-06-08 15:25:17 +00:00
|
|
|
|
|
|
|
brokername: str,
|
|
|
|
acctname: str,
|
|
|
|
|
2022-06-21 16:37:33 +00:00
|
|
|
# post normalization filter on ledger entries to be processed
|
2022-06-22 19:41:26 +00:00
|
|
|
filter_by: Optional[list[dict]] = None,
|
2022-06-21 16:37:33 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
) -> tuple[
|
|
|
|
dict[str, Transaction],
|
|
|
|
dict[str, Position],
|
|
|
|
]:
|
2022-06-11 20:18:05 +00:00
|
|
|
'''
|
|
|
|
Open a ledger file by broker name and account and read in and
|
2022-07-26 15:27:38 +00:00
|
|
|
process any trade records into our normalized ``Transaction`` form
|
|
|
|
and then update the equivalent ``Pptable`` and deliver the two
|
|
|
|
bsuid-mapped dict-sets of the transactions and pps.
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-06-11 20:18:05 +00:00
|
|
|
'''
|
2022-07-26 15:27:38 +00:00
|
|
|
with (
|
|
|
|
open_trade_ledger(brokername, acctname) as ledger,
|
|
|
|
open_pps(brokername, acctname) as table,
|
|
|
|
):
|
2022-06-21 16:37:33 +00:00
|
|
|
if not ledger:
|
|
|
|
# null case, no ledger file with content
|
|
|
|
return {}
|
2022-06-15 13:55:32 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
mod = get_brokermod(brokername)
|
|
|
|
src_records: dict[str, Transaction] = mod.norm_trade_records(ledger)
|
|
|
|
|
|
|
|
if filter_by:
|
|
|
|
records = {}
|
|
|
|
bsuids = set(filter_by)
|
|
|
|
for tid, r in src_records.items():
|
|
|
|
if r.bsuid in bsuids:
|
|
|
|
records[tid] = r
|
|
|
|
else:
|
|
|
|
records = src_records
|
2022-06-21 16:37:33 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
updated = table.update_from_trans(records)
|
2022-06-21 16:37:33 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
return records, updated
|
2022-06-13 18:11:37 +00:00
|
|
|
|
|
|
|
|
2022-06-19 20:30:08 +00:00
|
|
|
# TODO: instead see if we can hack tomli and tomli-w to do the same:
|
|
|
|
# - https://github.com/hukkin/tomli
|
|
|
|
# - https://github.com/hukkin/tomli-w
|
2022-06-17 19:41:17 +00:00
|
|
|
class PpsEncoder(toml.TomlEncoder):
|
|
|
|
'''
|
|
|
|
Special "styled" encoder that makes a ``pps.toml`` redable and
|
|
|
|
compact by putting `.clears` tables inline and everything else
|
|
|
|
flat-ish.
|
|
|
|
|
|
|
|
'''
|
|
|
|
separator = ','
|
|
|
|
|
2022-06-18 22:30:53 +00:00
|
|
|
def dump_list(self, v):
|
2022-06-19 20:30:08 +00:00
|
|
|
'''
|
|
|
|
Dump an inline list with a newline after every element and
|
|
|
|
with consideration for denoted inline table types.
|
2022-06-18 22:30:53 +00:00
|
|
|
|
2022-06-19 20:30:08 +00:00
|
|
|
'''
|
2022-06-18 22:30:53 +00:00
|
|
|
retval = "[\n"
|
|
|
|
for u in v:
|
|
|
|
if isinstance(u, toml.decoder.InlineTableDict):
|
|
|
|
out = self.dump_inline_table(u)
|
|
|
|
else:
|
|
|
|
out = str(self.dump_value(u))
|
|
|
|
|
|
|
|
retval += " " + out + "," + "\n"
|
|
|
|
retval += "]"
|
|
|
|
return retval
|
|
|
|
|
2022-06-17 19:41:17 +00:00
|
|
|
def dump_inline_table(self, section):
|
|
|
|
"""Preserve inline table in its compact syntax instead of expanding
|
|
|
|
into subsection.
|
|
|
|
https://github.com/toml-lang/toml#user-content-inline-table
|
|
|
|
"""
|
|
|
|
val_list = []
|
|
|
|
for k, v in section.items():
|
|
|
|
# if isinstance(v, toml.decoder.InlineTableDict):
|
|
|
|
if isinstance(v, dict):
|
|
|
|
val = self.dump_inline_table(v)
|
|
|
|
else:
|
|
|
|
val = str(self.dump_value(v))
|
|
|
|
|
|
|
|
val_list.append(k + " = " + val)
|
|
|
|
|
|
|
|
retval = "{ " + ", ".join(val_list) + " }"
|
|
|
|
return retval
|
|
|
|
|
|
|
|
def dump_sections(self, o, sup):
|
|
|
|
retstr = ""
|
|
|
|
if sup != "" and sup[-1] != ".":
|
|
|
|
sup += '.'
|
|
|
|
retdict = self._dict()
|
|
|
|
arraystr = ""
|
|
|
|
for section in o:
|
|
|
|
qsection = str(section)
|
|
|
|
value = o[section]
|
|
|
|
|
|
|
|
if not re.match(r'^[A-Za-z0-9_-]+$', section):
|
|
|
|
qsection = toml.encoder._dump_str(section)
|
|
|
|
|
|
|
|
# arrayoftables = False
|
|
|
|
if (
|
|
|
|
self.preserve
|
|
|
|
and isinstance(value, toml.decoder.InlineTableDict)
|
|
|
|
):
|
|
|
|
retstr += (
|
|
|
|
qsection
|
|
|
|
+
|
|
|
|
" = "
|
|
|
|
+
|
|
|
|
self.dump_inline_table(o[section])
|
|
|
|
+
|
|
|
|
'\n' # only on the final terminating left brace
|
|
|
|
)
|
|
|
|
|
|
|
|
# XXX: this code i'm pretty sure is just blatantly bad
|
|
|
|
# and/or wrong..
|
|
|
|
# if isinstance(o[section], list):
|
|
|
|
# for a in o[section]:
|
|
|
|
# if isinstance(a, dict):
|
|
|
|
# arrayoftables = True
|
|
|
|
# if arrayoftables:
|
|
|
|
# for a in o[section]:
|
|
|
|
# arraytabstr = "\n"
|
|
|
|
# arraystr += "[[" + sup + qsection + "]]\n"
|
|
|
|
# s, d = self.dump_sections(a, sup + qsection)
|
|
|
|
# if s:
|
|
|
|
# if s[0] == "[":
|
|
|
|
# arraytabstr += s
|
|
|
|
# else:
|
|
|
|
# arraystr += s
|
|
|
|
# while d:
|
|
|
|
# newd = self._dict()
|
|
|
|
# for dsec in d:
|
|
|
|
# s1, d1 = self.dump_sections(d[dsec], sup +
|
|
|
|
# qsection + "." +
|
|
|
|
# dsec)
|
|
|
|
# if s1:
|
|
|
|
# arraytabstr += ("[" + sup + qsection +
|
|
|
|
# "." + dsec + "]\n")
|
|
|
|
# arraytabstr += s1
|
|
|
|
# for s1 in d1:
|
|
|
|
# newd[dsec + "." + s1] = d1[s1]
|
|
|
|
# d = newd
|
|
|
|
# arraystr += arraytabstr
|
|
|
|
|
|
|
|
elif isinstance(value, dict):
|
|
|
|
retdict[qsection] = o[section]
|
|
|
|
|
|
|
|
elif o[section] is not None:
|
|
|
|
retstr += (
|
|
|
|
qsection
|
|
|
|
+
|
|
|
|
" = "
|
|
|
|
+
|
|
|
|
str(self.dump_value(o[section]))
|
|
|
|
)
|
|
|
|
|
|
|
|
# if not isinstance(value, dict):
|
|
|
|
if not isinstance(value, toml.decoder.InlineTableDict):
|
|
|
|
# inline tables should not contain newlines:
|
|
|
|
# https://toml.io/en/v1.0.0#inline-table
|
|
|
|
retstr += '\n'
|
|
|
|
|
|
|
|
else:
|
|
|
|
raise ValueError(value)
|
|
|
|
|
|
|
|
retstr += arraystr
|
|
|
|
return (retstr, retdict)
|
|
|
|
|
|
|
|
|
2022-07-18 16:23:02 +00:00
|
|
|
@cm
|
|
|
|
def open_pps(
|
|
|
|
brokername: str,
|
|
|
|
acctid: str,
|
2022-07-21 14:12:51 +00:00
|
|
|
write_on_exit: bool = True,
|
2022-07-18 16:23:02 +00:00
|
|
|
|
2022-07-21 14:12:51 +00:00
|
|
|
) -> PpTable:
|
2022-07-18 16:23:02 +00:00
|
|
|
'''
|
|
|
|
Read out broker-specific position entries from
|
|
|
|
incremental update file: ``pps.toml``.
|
|
|
|
|
2022-06-19 20:30:08 +00:00
|
|
|
'''
|
2022-06-10 21:50:29 +00:00
|
|
|
conf, path = config.load('pps')
|
|
|
|
brokersection = conf.setdefault(brokername, {})
|
2022-06-15 13:55:32 +00:00
|
|
|
pps = brokersection.setdefault(acctid, {})
|
2022-07-19 12:26:28 +00:00
|
|
|
|
2022-07-26 15:27:38 +00:00
|
|
|
# TODO: ideally we can pass in an existing
|
|
|
|
# pps state to this right? such that we
|
|
|
|
# don't have to do a ledger reload all the
|
|
|
|
# time.. a couple ideas I can think of,
|
|
|
|
# - mirror this in some client side actor which
|
|
|
|
# does the actual ledger updates (say the paper
|
|
|
|
# engine proc if we decide to always spawn it?),
|
|
|
|
# - do diffs against updates from the ledger writer
|
|
|
|
# actor and the in-mem state here?
|
|
|
|
|
2022-06-21 16:37:33 +00:00
|
|
|
pp_objs = {}
|
2022-07-26 15:27:38 +00:00
|
|
|
table = PpTable(
|
|
|
|
brokername,
|
2022-07-27 12:43:09 +00:00
|
|
|
acctid,
|
2022-07-26 15:27:38 +00:00
|
|
|
pp_objs,
|
|
|
|
conf=conf,
|
|
|
|
)
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-07-18 16:23:02 +00:00
|
|
|
# unmarshal/load ``pps.toml`` config entries into object form
|
|
|
|
# and update `PpTable` obj entries.
|
2022-06-21 16:37:33 +00:00
|
|
|
for fqsn, entry in pps.items():
|
2022-06-22 19:41:26 +00:00
|
|
|
bsuid = entry['bsuid']
|
2022-06-21 16:37:33 +00:00
|
|
|
|
|
|
|
# convert clears sub-tables (only in this form
|
|
|
|
# for toml re-presentation) back into a master table.
|
|
|
|
clears_list = entry['clears']
|
|
|
|
|
|
|
|
# index clears entries in "object" form by tid in a top
|
|
|
|
# level dict instead of a list (as is presented in our
|
|
|
|
# ``pps.toml``).
|
2022-07-01 20:12:09 +00:00
|
|
|
pp = pp_objs.get(bsuid)
|
|
|
|
if pp:
|
|
|
|
clears = pp.clears
|
|
|
|
else:
|
|
|
|
clears = {}
|
|
|
|
|
2022-06-21 16:37:33 +00:00
|
|
|
for clears_table in clears_list:
|
|
|
|
tid = clears_table.pop('tid')
|
|
|
|
clears[tid] = clears_table
|
|
|
|
|
2022-06-22 19:41:26 +00:00
|
|
|
size = entry['size']
|
2022-07-25 15:57:57 +00:00
|
|
|
be_price = entry['be_price']
|
2022-06-22 19:41:26 +00:00
|
|
|
|
2022-06-21 16:37:33 +00:00
|
|
|
expiry = entry.get('expiry')
|
|
|
|
if expiry:
|
|
|
|
expiry = pendulum.parse(expiry)
|
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
pp = pp_objs[bsuid] = Position(
|
2022-06-21 16:37:33 +00:00
|
|
|
Symbol.from_fqsn(fqsn, info={}),
|
2022-06-22 19:41:26 +00:00
|
|
|
size=size,
|
2022-07-25 15:57:57 +00:00
|
|
|
be_price=be_price,
|
2022-06-21 16:37:33 +00:00
|
|
|
expiry=expiry,
|
|
|
|
bsuid=entry['bsuid'],
|
|
|
|
|
|
|
|
# XXX: super critical, we need to be sure to include
|
|
|
|
# all pps.toml clears to avoid reusing clears that were
|
|
|
|
# already included in the current incremental update
|
|
|
|
# state, since today's records may have already been
|
|
|
|
# processed!
|
|
|
|
clears=clears,
|
|
|
|
)
|
2022-06-08 15:25:17 +00:00
|
|
|
|
2022-07-25 15:57:57 +00:00
|
|
|
# audit entries loaded from toml
|
|
|
|
pp.size, pp.be_price = pp.audit_sizing()
|
|
|
|
|
2022-07-22 23:01:23 +00:00
|
|
|
try:
|
|
|
|
yield table
|
|
|
|
finally:
|
|
|
|
if write_on_exit:
|
2022-07-27 12:43:09 +00:00
|
|
|
table.write_config()
|
2022-06-19 20:30:08 +00:00
|
|
|
|
|
|
|
|
2022-06-10 21:50:29 +00:00
|
|
|
if __name__ == '__main__':
|
2022-06-13 18:11:37 +00:00
|
|
|
import sys
|
|
|
|
|
|
|
|
args = sys.argv
|
|
|
|
assert len(args) > 1, 'Specifiy account(s) from `brokers.toml`'
|
|
|
|
args = args[1:]
|
|
|
|
for acctid in args:
|
|
|
|
broker, name = acctid.split('.')
|
2022-07-26 15:27:38 +00:00
|
|
|
trans, updated_pps = load_pps_from_ledger(broker, name)
|
|
|
|
print(
|
|
|
|
f'Processing transactions into pps for {broker}:{acctid}\n'
|
|
|
|
f'{pformat(trans)}\n\n'
|
|
|
|
f'{pformat(updated_pps)}'
|
|
|
|
)
|