Skip to content

本地部署codeqwen-1_5-7b-chat-q5_k_m.gguf 出现'max_tokens' or 'max_completion_tokens' is too large: 64000 #272

@nessajming

Description

@nessajming

Error: 400 'max_tokens' or 'max_completion_tokens' is too large: 64000. This model's maximum context length is 64001 tokens and your request has 24235 input tokens (64000 > 64001 - 24235). None

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions