Prize Pool: $100,000
- 1st Place: $20,000
- 2nd Place: $10,000
- 3rd Place: $10,000
- 4th-15th Place: $5,000 each
Timeline:
- Entry Deadline: September 29, 2025
- Team Merger Deadline: September 29, 2025
- Final Submission Deadline: October 6, 2025
- Format: Notebook submissions only
- Runtime: β€ 8 hours (9 hours during forecasting phase)
- Internet: Disabled during execution
- Submissions: 5 per day maximum, 2 final submissions for judging
- Team Size: Maximum 5 members
- Reproducibility: Complete code delivery required for winners
- Documentation: Detailed methodology and architecture description
- License: Non-exclusive license granted to sponsor
- External Data: Allowed if "reasonably accessible to all"
-
Data Understanding
- Analyze 1977 features across LME, JPX, US, FX markets
- Understand 424 target configurations
- Identify data quality issues and patterns
-
Baseline Development
- Implement memory-efficient pipeline
- Create robust cross-validation
- Establish performance benchmarks
-
Community Engagement
- Join official Discord
- Study public notebooks
- Consider team formation
-
Feature Engineering
- Domain-specific features (spreads, ratios, correlations)
- Time-series features (lags, rolling statistics)
- Cross-asset relationships
-
Model Development
- Gradient boosting optimization
- Ensemble methods
- Hyperparameter tuning
-
Memory & Runtime Optimization
- Feature selection (top 500-1000 features)
- Data type optimization
- Batch processing
-
Multi-Target Optimization
- Target-specific feature engineering
- Multi-output model tuning
- Target correlation analysis
-
Stability & Generalization
- Overfitting prevention
- Robust validation strategies
- Model ensemble diversity
-
Competition-Specific Tuning
- Sharpe ratio variant optimization
- Stability-focused evaluation
- Cross-validation with gaps
-
Submission Preparation
- Notebook format conversion
- Runtime optimization
- Memory efficiency verification
-
Quality Assurance
- Reproducibility testing
- Performance validation
- Competition compliance check
# Memory optimization strategies
- Reduce data types (float64 β float32 β float16)
- Feature selection (correlation-based)
- Chunk processing for large datasets
- Batch predictions
- Garbage collection# Runtime optimization strategies
- Conservative model parameters
- Early stopping in gradient boosting
- Single-threaded processing
- Efficient data loading
- Minimal cross-validation during submission# Overfitting prevention strategies
- Time-series CV with gaps
- Regularization (L1/L2)
- Feature sampling
- Data sampling
- Model ensemble diversity# Sharpe ratio variant optimization
- Mean Spearman correlation / Standard deviation
- Focus on stability (lower variance)
- Multi-target performance
- Robust evaluation across time periods# Key components
from src.memory_optimization import create_memory_efficient_pipeline
from src.robust_validation import time_series_cv_robust
from src.multi_target import parse_target_pairs
# Usage
train_opt, labels_opt, test_opt, features = create_memory_efficient_pipeline(
train_df, train_labels, target_pairs, test_df,
max_features=500, # Conservative for runtime
max_targets=None # Use all targets
)# Run competition submission
python run_competition_submission.py
# Or use notebook template
jupyter notebook notebooks/competition_submission_template.ipynb- Memory Usage: < 16 GB
- Runtime: < 6 hours (2-hour buffer)
- CV Score: > 0.5 (competition metric)
- Stability: CV std < 0.1
- Memory Usage: < 12 GB
- Runtime: < 4 hours
- CV Score: > 1.0
- Stability: CV std < 0.05
- Memory Management: Major challenge (14 replies in "Reduce data size")
- Overfitting: Significant concern (4 comments)
- Team Formation: Active (multiple "Looking for team" posts)
- Getting Started: Official Discord available
- Download real competition data immediately
- Join official Discord for community support
- Study top public notebooks for insights
- Consider team formation for better results
- Focus on memory optimization first
- Public: Based on public test set (visible during competition)
- Private: Based on private test set (final ranking)
- Strategy: Don't overfit to public leaderboard
# Robust validation approach
- Time-series CV with gaps (prevent data leakage)
- Multiple validation periods
- Stability-focused evaluation
- Overfitting detection- Memory Issues: Implement aggressive memory optimization
- Runtime Exceeded: Conservative model parameters
- Overfitting: Robust validation and regularization
- Reproducibility: Complete documentation and code
- Late Submission: Submit early and iterate
- Rule Violations: Review rules carefully
- Team Issues: Clear communication and agreements
- External Dependencies: Avoid expensive external data/tools
- Download and analyze real competition data
- Implement memory-efficient baseline
- Join official Discord community
- Submit first entry
- Achieve baseline performance
- Optimize for memory and runtime
- Study top public notebooks
- Consider team formation
- Implement advanced feature engineering
- Optimize for competition metric
- Achieve competitive performance
- Prepare final submission strategy
- Final model optimization
- Competition submission preparation
- Performance validation
- Documentation completion
Remember: This is a marathon, not a sprint. Focus on stability, reproducibility, and incremental improvements rather than chasing the public leaderboard.