Skip to main content

ML Infrastructure Overview

The Hokusai ML infrastructure powers the protocol's ability to evaluate and reward data contributions that improve AI models. This production-ready system includes a Metaflow-based data pipeline for model evaluation and attestation generation, with plans to expand into a reusable ML platform package.

Architecture Overview

Core Capabilities

The Hokusai ML pipeline is a production-ready Metaflow-based system that:

  • Evaluates Contributions: Measures how contributed data improves model performance
  • Generates Attestations: Creates cryptographic proofs of improvement (DeltaOne scores)
  • Integrates with Blockchain: Produces outputs suitable for on-chain verification

Key Features

  1. Automated Evaluation: Compare baseline vs improved models
  2. Privacy Protection: PII detection and data validation
  3. Reproducibility: Deterministic pipeline execution
  4. Scalability: Handle large datasets via streaming

Pipeline Flow

Current Implementation

The ML pipeline includes several integrated components:

Model Registry

MLFlow-based model management for tracking experiments and versions:

# Current implementation
from src.pipeline.model_registry import ModelRegistry

registry = ModelRegistry()
model_info = registry.log_model(
model=improved_model,
metrics=evaluation_results,
data_version=contribution_id
)

Evaluation Framework

Automated comparison of baseline vs improved models:

# Pipeline evaluation step
@step
def evaluate_models(self):
baseline_metrics = evaluate(self.baseline_model, test_data)
improved_metrics = evaluate(self.improved_model, test_data)
self.deltaone_score = compute_deltaone(baseline_metrics, improved_metrics)

Attestation Generation

Cryptographic proofs of model improvement:

# Generate verifiable attestation
attestation = {
"model_id": model_id,
"deltaone_score": deltaone_score,
"contributor": eth_address,
"timestamp": timestamp,
"signature": generate_signature(...)
}

Use Cases

For Data Contributors

  • Submit datasets to improve specific models
  • Track contribution impact via DeltaOne scores
  • Earn rewards based on measurable improvements

For Model Developers

  • Access high-quality training data
  • Leverage automated evaluation infrastructure
  • Deploy improved models with confidence

For Protocol Integrators

  • Connect to Hokusai's evaluation infrastructure
  • Submit data contributions programmatically
  • Track rewards and attestations on-chain

Technical Stack

  • Orchestration: Metaflow for pipeline management
  • ML Framework: MLFlow for experiment tracking
  • Storage: S3-compatible object storage
  • Compute: Kubernetes for scalable execution
  • Monitoring: Integrated logging and metrics

Getting Started

Running the Pipeline

# Clone the repository
git clone https://github.com/Hokusai-protocol/hokusai-data-pipeline
cd hokusai-data-pipeline

# Install dependencies
pip install -r requirements.txt

# Run evaluation
python -m src.pipeline.hokusai_pipeline run \
--contributed-data=your_data.csv \
--eth-address=0x... \
--model-type=gpt-3.5-turbo

Quick Testing

# Run with test data
python -m src.pipeline.hokusai_pipeline run \
--dry-run \
--contributed-data=data/test_fixtures/test_queries.csv

Key Features

FeatureDescriptionStatus
Data ValidationPII detection, schema validation✅ Active
Model TrainingAutomated fine-tuning with contributed data✅ Active
EvaluationBaseline vs improved model comparison✅ Active
DeltaOne ScoringQuantified improvement metrics✅ Active
AttestationCryptographic proof generation✅ Active
MLFlow IntegrationExperiment tracking and model registry✅ Active
Streaming SupportHandle large datasets efficiently✅ Active

Future: ML Platform Package

Note: The hokusai-ml-platform package is currently under development. This section describes the planned architecture and APIs.

Vision

Transform the Hokusai data pipeline's ML infrastructure into a standalone platform that any project can use to:

  • Manage model versions and deployments
  • Run A/B tests between models
  • Track performance improvements
  • Generate attestations for contributions
  • Build inference pipelines with caching

Planned Components

Key Platform Features (Coming Soon)

🚀 Production Ready

  • Battle-tested components from Hokusai data pipeline
  • Comprehensive error handling and logging
  • Performance optimized for high throughput
  • Built-in monitoring and metrics

🔄 Model Lifecycle Management

  • Version control for models
  • Automated rollback on performance degradation
  • Staging environments (dev, staging, production)
  • Model lineage tracking

📊 A/B Testing Framework

# Future API example
from hokusai.core.ab_testing import ModelTrafficRouter, ABTestConfig

router = ModelTrafficRouter()
test = ABTestConfig(
model_a="lead-scorer/1.0.0",
model_b="lead-scorer/1.1.0",
traffic_split={"a": 0.8, "b": 0.2},
metrics_to_track=["conversion_rate", "latency"]
)
router.create_test(test)

Next Steps