mirror of
https://github.com/ggml-org/whisper.cpp.git
synced 2025-09-15 13:28:35 +08:00
* whisper : migrate to ggml-backend * whisper : fix logit reading * whisper : fix tensor allocation during load * whisper : fix beam-search with CUDA * whisper : free backends + fix compile warning * whisper : print when CUDA is enabled * whisper : fix CoreML * make : clean-up * talk : fix compile warning * whisper : support ggml_conv with CUDA and Metal (#1473) * ggml : add CUDA support for ggml_conv * whisper : remove ggml_repeat for conv bias + single backend * cuda : fix im2col kernel * metal : add im2col support + mul mat-vec f16 x f16 * bench-all : add q4 models * whisper : clean-up * quantize-all : fix * ggml : im2col opts * whisper : avoid whisper_model_data wrapper * whisper : add note that ggml_mul_mat_pad does not work with CUDA * whisper : factor out graph compute in common function * whisper : fixes * whisper : fix UB with measure buffers * whisper : try to fix the parallel whisper_state functionality (#1479) * whisper : try to fix the parallel whisper_state functionality * whisper : fix multi-state Metal * whisper : free backend instances in whisper_state |
||
|---|---|---|
| .. | ||
| addon.node | ||
| bench | ||
| bench.wasm | ||
| command | ||
| command.wasm | ||
| lsp | ||
| main | ||
| quantize | ||
| stream | ||
| stream.wasm | ||
| talk | ||
| talk-llama | ||
| talk.wasm | ||
| whisper.android | ||
| whisper.nvim | ||
| whisper.objc | ||
| whisper.swiftui | ||
| whisper.wasm | ||
| CMakeLists.txt | ||
| common-ggml.cpp | ||
| common-ggml.h | ||
| common-sdl.cpp | ||
| common-sdl.h | ||
| common.cpp | ||
| common.h | ||
| dr_wav.h | ||
| generate-karaoke.sh | ||
| helpers.js | ||
| livestream.sh | ||
| twitch.sh | ||
| yt-wsp.sh | ||