You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I wonder why there is such a limitation in llamacpp? if self.config.task == "text-generation": if input_shapes["batch_size"] != 1: raise ValueError("Batch size must be 1 for LlamaCpp text generation")
In '/optimum_benchmark/backends/llama_cpp/backend.py'