Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add q8_0 models to download-ggml-model.sh #2589

Merged
merged 1 commit into from
Nov 28, 2024

Conversation

mrienstra
Copy link
Contributor

Models names as per v1.7.2 announcement.

PR adds -q8_0 models for tiny, base, small, medium, large-v2, & large-v3-turbo.

With the changes in this PR, the remaining differences between the model list in the v1.7.2 announcement and models/download-ggml-model.sh are:

Only in models/download-ggml-model.sh

  • tiny.en
  • tiny.en-q5_1
  • base.en-q5_1
  • small.en
  • small.en-tdrz
  • small.en-q5_1
  • medium.en
  • medium.en-q5_0
  • large-v1
  • large-v3
  • large-v3-q5_0

Only in v1.7.2 announcement

  • tiny
  • tiny-q5_0
  • base-q5_0
  • small-q5_0
  • medium-q5_1
  • medium-dis
  • large-v2-q5_1
  • large-v2-dis

@ggerganov ggerganov merged commit a9d06ce into ggerganov:master Nov 28, 2024
42 of 44 checks passed
bygreencn added a commit to bygreencn/whisper.cpp that referenced this pull request Dec 3, 2024
* ggerganov/master: (447 commits)
  ruby : Add low-level methods to transcribe (ggerganov#2585)
  models : add `q8_0` models to `download-ggml-model.sh` (ggerganov#2589)
  ruby : Follow source tree change (ggerganov#2580)
  whisper : use backend registry (#0)
  ggml/sched : do not skip views in pre-assignments
  whisper : adapt to new ggml (wip)
  talk-llama : sync llama.cpp
  sync : ggml
  ggml : sync resolve (skip) (#0)
  Add required ggml-base and backend libs to cmake pkg (llama/10407)
  cuda : fix CUDA_FLAGS not being applied (llama/10403)
  sycl : Add option to set the SYCL architecture for all targets (llama/10266)
  vulkan: Optimize soft_max (llama/10301)
  sycl: Revert MUL_MAT_OP support changes (llama/10385)
  cuda : only use native when supported by cmake (llama/10389)
  vulkan: remove use of null initializer (llama/10372)
  metal : fox offset integer overflows in im2col (ggml/1015)
  Vulkan: Fix device info output format specifiers (llama/10366)
  metal : add `GGML_UNARY_OP_ELU` kernel (ggml/1018)
  CUDA: fix MMV kernel being used for FP16 src1 (llama/10357)
  ...
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants