Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
604 changes: 604 additions & 0 deletions deep_research/README.md

Large diffs are not rendered by default.

Empty file added deep_research/__init__.py
Empty file.
79 changes: 79 additions & 0 deletions deep_research/configs/balanced_research.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,79 @@
# Deep Research Pipeline Configuration - Balanced Mode
enable_cache: true

# ZenML MCP
model:
name: "deep_research"
description: "Parallelized ZenML pipelines for deep research on a given query."
tags:
[
"research",
"exa",
"tavily",
"openrouter",
"sambanova",
"langfuse",
"balanced",
]
use_cases: "Research on a given query."

# Langfuse project name for LLM tracking
langfuse_project_name: "deep-research"

# Research parameters for balanced research
parameters:
query: "Default research query"

steps:
initial_query_decomposition_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"
max_sub_questions: 10 # Balanced number of sub-questions

process_sub_question_step:
parameters:
llm_model_search: "sambanova/Meta-Llama-3.3-70B-Instruct"
llm_model_synthesis: "sambanova/DeepSeek-R1-Distill-Llama-70B"
cap_search_length: 20000 # Standard cap for search length

cross_viewpoint_analysis_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"
viewpoint_categories:
[
"scientific",
"political",
"economic",
"social",
"ethical",
"historical",
] # Standard viewpoints

generate_reflection_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"

get_research_approval_step:
parameters:
timeout: 3600 # 1 hour timeout
max_queries: 2 # Moderate additional queries

execute_approved_searches_step:
parameters:
llm_model: "sambanova/Meta-Llama-3.3-70B-Instruct"
cap_search_length: 20000

pydantic_final_report_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"

# Environment settings
settings:
docker:
requirements:
- openai>=1.0.0
- tavily-python>=0.2.8
- PyYAML>=6.0
- click>=8.0.0
- pydantic>=2.0.0
- typing_extensions>=4.0.0
43 changes: 43 additions & 0 deletions deep_research/configs/compare_viewpoints.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,43 @@
# Deep Research Pipeline Configuration - Compare Viewpoints
enable_cache: true

# ZenML MCP
model:
name: "deep_research"
description: "Parallelized ZenML pipelines for deep research on a given query."
tags:
[
"research",
"exa",
"tavily",
"openrouter",
"sambanova",
"langfuse",
"viewpoints",
]
use_cases: "Research on a given query."

# Langfuse project name for LLM tracking
langfuse_project_name: "deep-research"

# Research parameters for comparing different viewpoints
steps:
initial_query_decomposition_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"

cross_viewpoint_analysis_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"
viewpoint_categories: ["scientific", "political", "economic", "social", "ethical", "historical"]

# Environment settings
settings:
docker:
requirements:
- openai>=1.0.0
- tavily-python>=0.2.8
- PyYAML>=6.0
- click>=8.0.0
- pydantic>=2.0.0
- typing_extensions>=4.0.0
45 changes: 45 additions & 0 deletions deep_research/configs/daily_trends.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,45 @@
# Deep Research Pipeline Configuration - Daily Trends Research
enable_cache: false # Disable cache to always get fresh results for daily trends

# ZenML MCP
model:
name: "deep_research"
description: "Parallelized ZenML pipelines for deep research on a given query."
tags:
[
"research",
"exa",
"tavily",
"openrouter",
"sambanova",
"langfuse",
"daily_trends",
]
use_cases: "Research on a given query."

# Langfuse project name for LLM tracking
langfuse_project_name: "deep-research"

# Research parameters for daily trending topics
parameters:
query: "Latest developments in artificial intelligence"

steps:
initial_query_decomposition_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"

process_sub_question_step:
parameters:
llm_model_search: "sambanova/Meta-Llama-3.3-70B-Instruct"
llm_model_synthesis: "sambanova/DeepSeek-R1-Distill-Llama-70B"
# Environment settings
settings:
docker:
requirements:
- openai>=1.0.0
- tavily-python>=0.2.8
- PyYAML>=6.0
- click>=8.0.0
- pydantic>=2.0.0
- typing_extensions>=4.0.0
81 changes: 81 additions & 0 deletions deep_research/configs/deep_research.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,81 @@
# Deep Research Pipeline Configuration - Deep Comprehensive Mode
enable_cache: false # Disable cache for fresh comprehensive analysis

# ZenML MCP
model:
name: "deep_research"
description: "Parallelized ZenML pipelines for deep research on a given query."
tags:
[
"research",
"exa",
"tavily",
"openrouter",
"sambanova",
"langfuse",
"deep",
]
use_cases: "Research on a given query."

# Langfuse project name for LLM tracking
langfuse_project_name: "deep-research"

# Research parameters for deep comprehensive research
parameters:
query: "Default research query"

steps:
initial_query_decomposition_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"
max_sub_questions: 15 # Maximum sub-questions for comprehensive analysis

process_sub_question_step:
parameters:
llm_model_search: "sambanova/Meta-Llama-3.3-70B-Instruct"
llm_model_synthesis: "sambanova/DeepSeek-R1-Distill-Llama-70B"
cap_search_length: 30000 # Higher cap for more comprehensive data

cross_viewpoint_analysis_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"
viewpoint_categories:
[
"scientific",
"political",
"economic",
"social",
"ethical",
"historical",
"technological",
"philosophical",
] # Extended viewpoints for comprehensive analysis

generate_reflection_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"

get_research_approval_step:
parameters:
timeout: 7200 # 2 hour timeout for deep research
max_queries: 4 # Maximum additional queries for deep mode

execute_approved_searches_step:
parameters:
llm_model: "sambanova/Meta-Llama-3.3-70B-Instruct"
cap_search_length: 30000

pydantic_final_report_step:
parameters:
llm_model: "sambanova/DeepSeek-R1-Distill-Llama-70B"

# Environment settings
settings:
docker:
requirements:
- openai>=1.0.0
- tavily-python>=0.2.8
- PyYAML>=6.0
- click>=8.0.0
- pydantic>=2.0.0
- typing_extensions>=4.0.0
71 changes: 71 additions & 0 deletions deep_research/configs/enhanced_research.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,71 @@
# Enhanced Deep Research Pipeline Configuration
enable_cache: false

# ZenML MCP
model:
name: "deep_research"
description: "Parallelized ZenML pipelines for deep research on a given query."
tags:
[
"research",
"exa",
"tavily",
"openrouter",
"sambanova",
"langfuse",
"enhanced",
]
use_cases: "Research on a given query."

# Research query parameters
query: "Climate change policy debates"

# Langfuse project name for LLM tracking
langfuse_project_name: "deep-research"

# Step configurations
steps:
initial_query_decomposition_step:
parameters:
llm_model: "openrouter/google/gemini-2.5-flash-preview-05-20"

cross_viewpoint_analysis_step:
parameters:
llm_model: "openrouter/google/gemini-2.5-flash-preview-05-20"
viewpoint_categories:
[
"scientific",
"political",
"economic",
"social",
"ethical",
"historical",
]

generate_reflection_step:
parameters:
llm_model: "openrouter/google/gemini-2.5-flash-preview-05-20"

get_research_approval_step:
parameters:
timeout: 3600
max_queries: 2

execute_approved_searches_step:
parameters:
llm_model: "openrouter/google/gemini-2.5-flash-preview-05-20"

pydantic_final_report_step:
parameters:
llm_model: "openrouter/google/gemini-2.5-flash-preview-05-20"

# Environment settings
settings:
docker:
requirements:
- openai>=1.0.0
- tavily-python>=0.2.8
- PyYAML>=6.0
- click>=8.0.0
- pydantic>=2.0.0
- typing_extensions>=4.0.0
Loading
Loading