Skip to content

Workflow

KaizenFlow workflow explanation

This document is a roadmap of most activities that Quants, Quant devs, and DevOps can perform using KaizenFlow.

For each activity we point to the relevant resources (e.g., documents in docs, notebooks) in the repo.

A high-level description of KaizenFlow is KaizenFlow White Paper

Work organization

Set-up

  • TODO(gp): Add pointers to the docs we ask to read during the on-boarding

Documentation_meta

Quant workflows

The life of a Quant is spent between:

  • Exploring the raw data
  • Computing features
  • Building models to predict output given features
  • Assessing models

These activities are mapped in KaizenFlow as follows:

  • Exploring the raw data
  • This is performed by reading data using DataPull in a notebook and performing exploratory analysis
  • Computing features
  • This is performed by reading data using DataPull in a notebook and creating some DataFlow nodes
  • Building models to predict output given features
  • This is performed by connecting DataFlow nodes into a Dag
  • Assessing models
  • This is performed by running data through a Dag in a notebook or in a Python script and post-processing the results in an analysis notebook
  • Comparing models
  • The parameters of a model are exposed through a Config and then sweep over Config lists

DataPull

Universe

Dataset signature

DataFlow

Meta

DAG

System

Quant dev workflows

DataPull

DataFlow

TradingOps workflows

Trading execution

Intro

Components

Testing

Procedures

MLOps workflows

Deploying

Monitoring

DevOps workflows

The documentation outlines the architecture and deployment processes for the Kaizen Infrastructure, leveraging a blend of AWS services, Kubernetes for container orchestration, and traditional EC2 for virtualized computing. Emphasizing Infrastructure as Code (IaC), the project employs Terraform for provisioning and Ansible for configuration, ensuring a maintainable and replicable environment.

Overview

Current set-up description

Set up infra