Skip to content

Conversation

@charlieduzstuf
Copy link

@charlieduzstuf charlieduzstuf commented Oct 8, 2025

Doing some of this myself, and as an experiment I'm also giving part of this task to the AI itself. So far it seems a bit sloppy and more like pseudo-code, but I'll give it more of a kick in the coming week(s)

…dling

- Add HuggingFaceChatClient for running models from Hugging Face Hub
- Introduce LocalModelClient base class for common local model functionality
- Update README with detailed local model configuration options
- Add huggingface_hub dependency to requirements.txt
- Improve environment variable handling for local models
Implement GGUF model discovery functionality that searches for models in a specified directory. When GGUF_MODEL_DIR is set, the system will automatically find and load GGUF models by name without requiring full paths.

Update documentation to reflect new GGUF model discovery feature and clarify optional environment variables.
…lization

- Move GGUF model search function below imports and conditional imports
- Expand AVAILABLE_LLMS list with additional model variants
- Rename chat namespace to _openai_compatible_chat for clarity
- Simplify Hugging Face model parsing logic and update env var names
Added section comments to group and clarify the different model providers in the AVAILABLE_LLMS list for better readability and maintainability.
Add support for multiple new LLM providers (anthropic, bedrock, vertex_ai, deepseek, openrouter, ollama, localai, llama-cpp) and implement automatic GGUF model discovery with configurable local models directory. Include path detection logic and comprehensive provider-specific model name formatting.
@charlieduzstuf charlieduzstuf marked this pull request as ready for review October 9, 2025 09:39
@charlieduzstuf charlieduzstuf marked this pull request as draft October 9, 2025 09:40
Add support for --model_dir parameter to allow GGUF model discovery in a specified directory when using the 'local' provider. This enables interactive model selection when no specific model is provided and improves error handling for missing models or directories.
The example usage was uncommented and updated with a more specific prompt to demonstrate the LLM client functionality. This helps verify the client setup is working as expected.
Update response content access to use direct 'content' attribute instead of nested structure
Add model path construction for llama-2 model
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant