From 87e5fa520071416a95141f2af32756d8bebd6e65 Mon Sep 17 00:00:00 2001 From: Sapphire Koser Date: Thu, 17 Oct 2024 03:22:19 -0700 Subject: [PATCH] fix: use get_temperature in base LLM generate (#1520) Matches the behavior in agenerate, avoiding some API call failures --- src/ragas/llms/base.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/src/ragas/llms/base.py b/src/ragas/llms/base.py index 6d6051213..eb9d26532 100644 --- a/src/ragas/llms/base.py +++ b/src/ragas/llms/base.py @@ -88,7 +88,7 @@ async def generate( """Generate text using the given event loop.""" if temperature is None: - temperature = 1e-8 + temperature = self.get_temperature(n) if is_async: agenerate_text_with_retry = add_async_retry( @@ -280,7 +280,7 @@ async def agenerate_text( callbacks: Callbacks = None, ) -> LLMResult: if temperature is None: - temperature = 1e-8 + temperature = self.get_temperature(n) kwargs = self.check_args(n, temperature, stop, callbacks) li_response = await self.llm.acomplete(prompt.to_string(), **kwargs)