2

GitHub - thecolngroup/alphakit: Framework for developing sound algorithmic tradi...

 2 years ago
source link: https://github.com/thecolngroup/alphakit/
Go to the source link to view the article. You can view the picture content, updated content and better typesetting reading experience. If the link is broken, please click the button below to view the snapshot at that time.

Alphakit

Introducing a framework for algorithmic trading in Go and serverless cloud

           /$$           /$$                 /$$       /$$   /$$    
          | $$          | $$                | $$      |__/  | $$    
  /$$$$$$ | $$  /$$$$$$ | $$$$$$$   /$$$$$$ | $$   /$$ /$$ /$$$$$$  
 |____  $$| $$ /$$__  $$| $$__  $$ |____  $$| $$  /$$/| $$|_  $$_/  
  /$$$$$$$| $$| $$  \ $$| $$  \ $$  /$$$$$$$| $$$$$$/ | $$  | $$    
 /$$__  $$| $$| $$  | $$| $$  | $$ /$$__  $$| $$_  $$ | $$  | $$ /$$
|  $$$$$$$| $$| $$$$$$$/| $$  | $$|  $$$$$$$| $$ \  $$| $$  |  $$$$/
 \_______/|__/| $$____/ |__/  |__/ \_______/|__/  \__/|__/   \___/  
              | $$                                                  
              | $$                                                  
              |__/                                                  

Companion code repository for the forthcoming book Zero to Algo

"Master the latest features of Go and learn how to design, validate and deploy sound algorithmic trading strategies."

Inspiration

The majority of open source algo trading frameworks, especially in Go, focus purely on trade execution - that's a sure fire way to get rekt. The most important precursor to a successful trading system is researching and validating practical market edges - which is the focus of alphakit. Furthermore, I wanted a composable architecture that could easily be executed serverless in the cloud using features such as cloud functions and messaging queues.

What's included?

A complete starter kit for developing algorithmic trading strategies in the Go language:

  • Example buy-and-hold and trend-following algos
  • Backtest and walk-forward engine to evaluate algos
  • Performance reports and metrics
  • Brute force parameter optimization method
  • Command app to execute research studies from a config file
  • Scaffold for serverless production deployment in the cloud (coming soon)
  • Uses latest Go language (1.18) features including generics
  • Idiomatic Go style using community accepted best practices
  • Pragmatic use of concurrency, go routines and channels
  • Extensive test coverage where it matters

Install

go get "github.com/thecolngroup/alphakit"

Getting started

warning API is pre v1 and is not stable

The canonical example that brings together many of the framework components is in the optimize package and reproduced below. A further well documented example in the backtest package demonstrates how to use a simulated dealer to study algos without an optimizer.

func Example() {
	// Verbose error handling omitted for brevity

	// Identify the bot (algo) to optimize by supplying a factory function
	// Here we're using the classic moving average (MA) cross variant of trend bot
	bot := trend.MakeCrossBotFromConfig

	// Define the parameter space to optimize
	// Param names must match those expected by the MakeBot function passed to optimizer
	// Here we're optimizing the lookback period of a fast and slow MA
	// and the Market Meanness Index (MMI) filter
	paramSpace := ParamMap{
		"mafastlength": []any{30, 90, 180},
		"maslowlength": []any{90, 180, 360},
		"mmilength":    []any{200, 300},
	}

	// Read price samples to use for optimization
	btc, _ := market.ReadKlinesFromCSV("testdata/btcusdt-1h/")
	eth, _ := market.ReadKlinesFromCSV("testdata/ethusdt-1h/")
	priceSamples := [][]market.Kline{btc, eth}

	// Create a new brute style optimizer with a default simulated dealer (no broker costs)
	// The default optimization objective is the param set with the highest sharpe ratio
	optimizer := NewBruteOptimizer()
	optimizer.SampleSplitPct = 0.5   // Use first 50% as in-sample training data, and remainder for out-of-sample validation
	optimizer.WarmupBarCount = 360 // Set as maximum lookback of your param space
	optimizer.MakeBot = bot        // Tell the optimizer which bot to use

	// Prepare the optimizer and get an estimate on the number of trials (backtests) required
	trialCount, _ := optimizer.Prepare(paramSpace, priceSamples)
	fmt.Printf("%d backtest trials to run during optimization\n", trialCount)

	// Start the optimization process and monitor with a receive-only channel
	// Trials will execute concurrently with a default worker pool matching the num of CPUs
	trials, _ := optimizer.Start(context.Background())
	for range trials {
	  // Monitor for errors and progress
	}

	// Inspect the study results following optimization
	study := optimizer.Study()

	// Read out the optimal param set and results
	optimaPSet := study.Validation[0]
	fmt.Printf("Optima params: fast: %d slow: %d MMI: %d\n",
		optimaPSet.Params["mafastlength"], optimaPSet.Params["maslowlength"], optimaPSet.Params["mmilength"])
	optimaResult := study.ValidationResults[optimaPSet.ID]
	fmt.Printf("Optima sharpe ratio is %.2f", optimaResult.Sharpe)

	// Output:
	// 38 backtest trials to run during optimization
	// Optima params: fast: 30 slow: 90 MMI: 200
	// Optima sharpe ratio is 2.46
}

Fundamental architecture patterns

The core assumption underlying the framework is that price data enters the system at a defined interval. Each time a new kline arrives it triggers an evaluation process owned by a bot that may result in 1 or more new orders being issued to a dealer.

Every component that participates in this processing implements the market.Receiver interface and accepts a kline (and a context to control long running operations).

The broker package offers an API to mediate the interaction between bot and trading venue. A bot creates market positions by placing orders through an implementation of Dealer. A simulated dealer in the backtest package (also a price receiver) allows you study and validate algos outsde of an optimizer.

In future releases new Dealer implementations will enable you to connect to specific trading venues.

Working with price data

The price data used in the unit tests and examples is sourced from Binance. It's a good source of clean crypto data going back to late 2017. See https://github.com/binance/binance-public-data/.

Alphakit offers an API for price data in the market package. The primary representation is in the form of a candlestick (OHLC) - also known as a kline. CSVKlineReader expects the timestamp denoting the start of the kline interval to be provided in unix millisecond format. Convenience functions for reading individual CSV files or walking a directory are also included.

Performance reports

Package perf provides comprehensive performance reporting for your algo, enabling you to track industry standard metrics such as CAGR, return rate, sharpe ratio, and drawdowns.

To create a new report use the equity history and trade history data from a dealer.

Trading costs

Many algos appear to be viable until you correctly factor in trading costs! Package backtest offers a PerpCoster implementation that simulates typical costs you might expect when trading crypto perpetual futures, including an hourly funding rate fee. See the tests in package backtest to understand how costs are applied during backtesting.

Building a trading bot

In the trader package you will find a couple of example bots: hodl and trend. The hodl bot is useful for benchmarking an asset, and the trend bot serves as a template for developing your own algo.

The following notes refer to how the bot in the trend package operates.

Prediction

trader.Predicter is a simple interface that returns a value between -1 and 1. A value of 1 signals maximum confidence in opening a long position, whilst -1 maximum confidence in opening a short position. 0 indicates no directional bias. Other values between -1 and 1 indicate varying confidence in direction.

CrossPredicter uses a fast and slow moving average cross with a Market Meanness Index (MMI) filter to determine the prediction.

ApexPredicter uses peak and valley detection in a smoothed price series with an MMI filter.

To understand more about trend following and MMI this is a great starting point: https://financial-hacker.com/trend-and-exploiting-it/

The trend bot interprets the prediction value according to a set of threshold values for opening and closing positions, namely:

  • EnterLong
  • EnterShort
  • ExitLong
  • ExitShort

By varying these threshold values you can create asymmetric entry and exit conditions.

Risk Management

Package risk provides methods to calculate unit risk, used as an input to position sizing and stop loss specification.

By default 'full-risk' will be used which assumes no stop-loss. As an alternative a standard deviation method is also provided.

Money Management

Package money provides methods to size a position. By default (and recommended for initial backtesting) a position based on a fixed capital amount is used.

A more sophisticated option is to a use a fixed fraction method given by the SafeFSizer type. You can determine the optimal value of 'f' by using the OptimalF or Kelly value from a performance report.

Command app: studyrun

The command app studyrun enables you to execute optimization studies by specifying a .toml config file.

See the test in cmd/studyrun to understand the syntax and play with a working example.

If you wish to use your own custom bots with the command app you'll need to update the type registry inside internal/studyrun. In a later release we may implement a plug-in architecture to negate this step.

The command app will execute an optimization study using BruteOptimizer and dump out the results in .csv format.

Connecting to a live trading venue

Future releases will provide implementations of broker.Dealer for specific trading venues. Contributions welcome!

Further reading

Contributing

Please fork and raise a PR or submit an issue.

Contributions should comply with:


About Joyk


Aggregate valuable and interesting links.
Joyk means Joy of geeK