You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I’m using Vanna.ai in my architecture with ChromaDB, Gemini, and BigQuery. Users chat via a WebSocket, and I send the chat history as context to the LLM in the vn.ask() (I am calling the 4 functions in a sequence as mentioned on the documentation) function.
The first time it works fine and gives correct answers, but as the conversation history grows, I start getting direct answers instead of SQL queries (which are wrong). I have a retry mechanism in place that tries up to 5 times if the generated SQL is invalid, but the output remains the same each time.
I’m wondering:
Why does the generate_sql() function return direct answers instead of SQL queries?
Why does the retry mechanism always return the same output?
Could this be due to caching or model hallucinations, or is there another reason? Any suggestions on how to fix this?
I have attached my model architecture for better understanding.
Any help will be very useful, It will be even better if someone from Vanna could directly get in touch with us.
The text was updated successfully, but these errors were encountered:
I’m using Vanna.ai in my architecture with ChromaDB, Gemini, and BigQuery. Users chat via a WebSocket, and I send the chat history as context to the LLM in the
vn.ask()
(I am calling the 4 functions in a sequence as mentioned on the documentation) function.The first time it works fine and gives correct answers, but as the conversation history grows, I start getting direct answers instead of SQL queries (which are wrong). I have a retry mechanism in place that tries up to 5 times if the generated SQL is invalid, but the output remains the same each time.
I’m wondering:
Why does the generate_sql() function return direct answers instead of SQL queries?
Why does the retry mechanism always return the same output?
Could this be due to caching or model hallucinations, or is there another reason? Any suggestions on how to fix this?
I have attached my model architecture for better understanding.
Any help will be very useful, It will be even better if someone from Vanna could directly get in touch with us.
The text was updated successfully, but these errors were encountered: