]>
git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/log
Georgi Gerganov [Sun, 25 Jun 2023 11:22:21 +0000 (14:22 +0300)]
ggml : sync latest ggml lib
Bo-Yi Wu [Sun, 25 Jun 2023 11:07:55 +0000 (19:07 +0800)]
go : improve progress reporting and callback handling (#1024)
- Rename `cb` to `callNewSegment` in the `Process` function
- Add `callProgress` as a new parameter to the `Process` function
- Introduce `ProgressCallback` type for reporting progress during processing
- Update `Whisper_full` function to include `progressCallback` parameter
- Add `registerProgressCallback` function and `cbProgress` map for handling progress callbacks
Signed-off-by: appleboy <redacted>
byte-6174 [Sun, 25 Jun 2023 10:59:48 +0000 (06:59 -0400)]
make : update cuBLAS build both x86 and aarch64 (#1015)
make cuBLAS compilation compatible with x86 as well as aarch64.
KP Kaiser [Sun, 25 Jun 2023 10:57:18 +0000 (06:57 -0400)]
make : fix for CUDA native not working as an option on Ubuntu (#1012)
faker [Sun, 25 Jun 2023 10:52:29 +0000 (18:52 +0800)]
main : exit gracefully when invalid params are passed
* Refactor whisper_params_parse to return false on failure
* Updated help flag behavior
faker [Sun, 25 Jun 2023 10:51:59 +0000 (18:51 +0800)]
main : gracefully exit when invalid params are passed (#1002)
* Refactor whisper_params_parse to return false on failure
* Updated help flag behavior
Akash Mahajan [Sun, 25 Jun 2023 10:50:14 +0000 (03:50 -0700)]
py : make convert-pt-to-ggml.py backwards compatible with older vocab.json tokenizer files (#1001)
* patch checkpoint convert script to keep compatibility with older hf_transformers whisper tokenizer
* typo fix
Larry Battle [Sun, 25 Jun 2023 10:46:44 +0000 (05:46 -0500)]
readme : corrected syntax for markdown link (#995)
Nicholas Albion [Tue, 6 Jun 2023 00:27:26 +0000 (10:27 +1000)]
updated java README
Nicholas Albion [Thu, 1 Jun 2023 12:45:00 +0000 (22:45 +1000)]
`speak` scripts for Windows
Nicholas Albion [Thu, 1 Jun 2023 06:53:56 +0000 (16:53 +1000)]
updated README for java
geniusnut [Wed, 31 May 2023 07:13:14 +0000 (15:13 +0800)]
whisper.android : support decode wav file has 2 channels (#972)
Nicholas Albion [Sun, 28 May 2023 23:38:58 +0000 (09:38 +1000)]
Feature/java bindings2 (#944)
* Java needs to call `whisper_full_default_params_by_ref()`, returning struct by val does not seem to work.
* added convenience methods to WhisperFullParams
* Remove unused WhisperJavaParams
genevera (she/her) [Sat, 27 May 2023 07:40:28 +0000 (03:40 -0400)]
models : fix README.md (#964)
Fixes typo on line 76 of models/README.md
DGdev91 [Wed, 24 May 2023 18:11:01 +0000 (20:11 +0200)]
examples : update elevenlabs scripts to use official python API (#837)
* Update elevenlabs example to use ufficial python API
* Update elevenlabs example to use official python API
0xsourcecode [Wed, 24 May 2023 08:23:51 +0000 (04:23 -0400)]
readme : highlight OpenBLAS support (#956)
* highlight openblas support
* Update README.md
Georgi Gerganov [Tue, 23 May 2023 11:04:39 +0000 (14:04 +0300)]
talk-llama : sync latest llama.cpp (close #922, close #954)
Alexey Kharlamov [Sat, 20 May 2023 18:23:45 +0000 (19:23 +0100)]
cmake : build with any BLAS compatible library (#927)
* Build with any BLAS library
* ci: Removed explicit CUDA nvcc path
Georgi Gerganov [Sat, 20 May 2023 17:00:06 +0000 (20:00 +0300)]
ggml : update WASM SIMD
Georgi Gerganov [Sat, 20 May 2023 15:56:30 +0000 (18:56 +0300)]
ggml : sync latest ggml repo
- new Q4 and Q8 quantization
- updated CUDA
Nicholas Albion [Sat, 20 May 2023 15:25:02 +0000 (01:25 +1000)]
bindings : add java bindings (#931)
* WIP - java bindings
* updated README
* failed attempt at JNI
* fullTranscribe() test passes
* tested on Ubuntu 20
* link to Java bindings
Elkana Bardugo [Sat, 20 May 2023 15:17:54 +0000 (18:17 +0300)]
whisper : fix hebrew language code (#935)
Ahmad Bilal [Mon, 15 May 2023 15:36:06 +0000 (20:36 +0500)]
coreml : add support of large-v1 model (#926)
Georgi Gerganov [Sun, 14 May 2023 16:06:45 +0000 (19:06 +0300)]
release : v1.4.2
Georgi Gerganov [Sun, 14 May 2023 15:56:46 +0000 (18:56 +0300)]
ggml : add AVX dot products
Georgi Gerganov [Sun, 14 May 2023 15:46:19 +0000 (18:46 +0300)]
talk-llama : fix build + sync latest llama.cpp
Jhen-Jie Hong [Sun, 14 May 2023 15:11:08 +0000 (23:11 +0800)]
readme : improve Core ML model conversion guidance (#915)
Georgi Gerganov [Sun, 14 May 2023 15:09:44 +0000 (18:09 +0300)]
coreml : support quantized model files
Georgi Gerganov [Sun, 14 May 2023 15:04:23 +0000 (18:04 +0300)]
ggml : sync latest ggml
- New Q4 and Q5 formats
- Various improvements
Rich Jones [Sun, 14 May 2023 14:54:57 +0000 (16:54 +0200)]
main : fix help for --no-timestamps arg (#908)
Georgi Gerganov [Sun, 14 May 2023 07:01:52 +0000 (10:01 +0300)]
extra : update ggml sync script
Jhen-Jie Hong [Sun, 14 May 2023 06:47:02 +0000 (14:47 +0800)]
whisper.objc : enable Core ML in example & fix segmentation fault (#910)
* coreml : update endcoder header import path
* coreml : force objc_arc in whisper-encoder.mm
* whisper.objc : create coreml/ group link
* whisper.objc : add coreml model link
* whisper.objc : update readme
* coreml : use -fobjc-arc for coreml/whisper-encoder.mm
* ci: create dummy .mlmodelc for pass ios build
* whisper.objc : update readme
---------
Co-authored-by: Georgi Gerganov <redacted>
Georgi Gerganov [Sun, 14 May 2023 06:42:19 +0000 (09:42 +0300)]
coreml : fix seg fault, double free (#919, #917, #899)
Georgi Gerganov [Tue, 9 May 2023 15:38:12 +0000 (18:38 +0300)]
coreml : fix memory leak (#899)
Jonathan Soo [Mon, 8 May 2023 18:08:09 +0000 (14:08 -0400)]
cmake : fix define used for COREML_ALLOW_FALLBACK (#893)
Luis Herrera [Mon, 8 May 2023 17:59:21 +0000 (12:59 -0500)]
talk-llama : only copy used KV cache in get / set state (#890)
---------
Co-authored-by: ejones <redacted>
Clifford Heath [Mon, 8 May 2023 17:58:36 +0000 (03:58 +1000)]
readme : add instructions on converting to GGML + "--no-config" to wget (#874)
ZaBlazzingZephyrus [Mon, 8 May 2023 17:45:53 +0000 (00:45 +0700)]
cmake : fix options disabling AVX and AVX2 flags (#885)
Georgi Gerganov [Thu, 4 May 2023 16:31:04 +0000 (19:31 +0300)]
cmake : add options to disable CPU flags (#860)
RelatedTitle [Wed, 3 May 2023 20:47:37 +0000 (14:47 -0600)]
ci : add cuBLAS build workflow and fix error causing lines in CMakeLists (#867)
* Add windows build with cuBLAS
* Remove error causing lines for cuBLAS on Windows
Vulcan [Wed, 3 May 2023 16:24:43 +0000 (21:54 +0530)]
readme : partial OpenCL GPU support via CLBlast (#863)
* ggml : CLBlast support as in llama.cpp
Building with CLBlast speeds up whisper.cpp ~2x on low end / older AMD APUs (CPU with integrated GPU) such as the A9.
Usage:
WHISPER_CLBLAST=1 make
* CMake/Makefile : CLBlast support as in llama.cpp
Building with CLBlast speeds up whisper.cpp ~2x on low end / older AMD APUs (CPU with integrated GPU) such as the A9.
Usage:
```
Makefile:
cd whisper.cpp
WHISPER_CLBLAST=1 make
CMake:
cd whisper.cpp ; mkdir build ; cd build
cmake -DWHISPER_CLBLAST=ON ..
make
```
* Update README.md
Added OpenCL Build Instructions
* Instruction: Partial OpenCL GPU support via CLBlast
Added build instructions and examples for Make and CMake to support OpenCL enabled GPUs.
Vulcan [Tue, 2 May 2023 19:50:32 +0000 (01:20 +0530)]
build : CLBlast support as in llama.cpp (#862)
* ggml : CLBlast support as in llama.cpp
Building with CLBlast speeds up whisper.cpp ~2x on low end / older AMD APUs (CPU with integrated GPU) such as the A9.
Usage:
WHISPER_CLBLAST=1 make
* CMake/Makefile : CLBlast support as in llama.cpp
Building with CLBlast speeds up whisper.cpp ~2x on low end / older AMD APUs (CPU with integrated GPU) such as the A9.
Usage:
```
Makefile:
cd whisper.cpp
WHISPER_CLBLAST=1 make
CMake:
cd whisper.cpp ; mkdir build ; cd build
cmake -DWHISPER_CLBLAST=ON ..
make
```
Georgi Gerganov [Tue, 2 May 2023 18:47:12 +0000 (21:47 +0300)]
ggml : fix 32-bit ARM build + quantization
Georgi Gerganov [Tue, 2 May 2023 18:23:54 +0000 (21:23 +0300)]
ggml : sync ggml (clBLAST + tensor names)
Luis Herrera [Tue, 2 May 2023 17:05:27 +0000 (12:05 -0500)]
talk-llama : fix session prompt load (#854)
CRD716 [Tue, 2 May 2023 16:51:52 +0000 (11:51 -0500)]
whisper : add detect-language mode (#853)
* add detectlanguage flag
* renaming and help
* no idea why that last one didn't commit
* run language detection if dl is set
* help message fix
* various fixes
* fix quitting
* fix language being english on print
Luis Herrera [Mon, 1 May 2023 17:18:10 +0000 (12:18 -0500)]
talk-llama : add --session support (#845)
* feat: adding session support
* readme: adding --session info in examples/talk-llama
* llama: adding session fixes
* readme: updating session doc
* talk-llama: update the value of need_to_save_session to true in order to save the session in the subsequent interaction
* talk-llama: adding missing function which updates session_tokens
Georgi Gerganov [Mon, 1 May 2023 11:44:39 +0000 (14:44 +0300)]
bench : improve benchmarks
Georgi Gerganov [Mon, 1 May 2023 07:03:56 +0000 (10:03 +0300)]
whisper : add memory sizes for Q8_0 (close #846)
Baffin Lee [Mon, 1 May 2023 06:28:05 +0000 (14:28 +0800)]
whisper.wasm : fix typo in readme (#832)
Georgi Gerganov [Sun, 30 Apr 2023 19:57:42 +0000 (22:57 +0300)]
release : v1.4.1
Georgi Gerganov [Sun, 30 Apr 2023 19:50:04 +0000 (22:50 +0300)]
whisper : fix quantize bug (#842)
* whisper : debug
* whisper : fix bug during quantization
Georgi Gerganov [Sun, 30 Apr 2023 19:27:30 +0000 (22:27 +0300)]
ggml : fix UB (int << 31)
Georgi Gerganov [Sun, 30 Apr 2023 16:23:37 +0000 (19:23 +0300)]
release : v1.4.0
Georgi Gerganov [Sun, 30 Apr 2023 16:12:49 +0000 (19:12 +0300)]
examples : fix + refactor Levenshtein distance
Georgi Gerganov [Sun, 30 Apr 2023 15:51:57 +0000 (18:51 +0300)]
whisper : add integer quantization support (#540)
* whisper : add integer quantization support
* examples : add common-ggml + prepare to add "quantize" tool
* whisper : quantization tool ready
* whisper : fix F32 support
* whisper : try to fix shared lib linkage
* wasm : update quantized models to Q5
* bench.wasm : remove "medium" button
* bench.wasm : fix custom model button
* ggml : add Q5_0 and Q5_1 WASM SIMD
* wasm : add quantized models to all WASM examples
* wasm : bump DB version number to 2
* talk-llama : update example to latest llama.cpp
* node : increase test timeout to 10s
* readme : add information for model quantization
* wasm : add links to other examples
Georgi Gerganov [Sun, 30 Apr 2023 09:14:33 +0000 (12:14 +0300)]
whisper : add GPU support via cuBLAS (#834)
* make : add WHISPER_CUBLAS
* make : fix CUBLAS build
* whisper : disable Flash Attention + adjust memory buffers
* whisper : remove old commented code
* readme : add cuBLAS instructions
* cmake : add WHISPER_CUBLAS option
* gitignore : ignore build-cublas
Georgi Gerganov [Sat, 29 Apr 2023 18:37:23 +0000 (21:37 +0300)]
ggml : fix WASM build
Georgi Gerganov [Sat, 29 Apr 2023 18:33:33 +0000 (21:33 +0300)]
ggml : fix 32-bit ARM NEON (#836)
* ggml : add support for 32-bit ARM
* ggml : fix
* ggml : fix
Georgi Gerganov [Sat, 29 Apr 2023 18:14:09 +0000 (21:14 +0300)]
ggml : use vzip instead of vuzp for consistency
Georgi Gerganov [Sat, 29 Apr 2023 17:21:25 +0000 (20:21 +0300)]
ggml : fix WASM build
Georgi Gerganov [Sat, 29 Apr 2023 16:30:22 +0000 (19:30 +0300)]
ggml : sync with ggml repo (warning fixes + asserts)
Thijs Raymakers [Sat, 29 Apr 2023 15:55:37 +0000 (17:55 +0200)]
whisper : use correct seek_end when offset is used (#833)
Whenever an `offset_ms` is provided, the value of `seek_end` is
calculated incorrectly. This causes Whisper to keep transcribing
after the end of the file.
The current behavior looks like
```
[00:34:40.000 --> 00:34:47.000] This is an example audio file.
[00:34:47.000 --> 00:34:49.000] The text has been redacted
[00:34:49.000 --> 00:34:51.000] This is the end of the audio.
[00:34:51.000 --> 00:34:52.000] ***
[00:34:52.000 --> 00:34:53.000] ***
[00:34:53.000 --> 00:34:54.000] ***
[00:34:55.000 --> 00:34:56.000] ***
...
```
The expected behavior should be
```
[00:34:40.000 --> 00:34:47.000] This is an example audio file.
[00:34:47.000 --> 00:34:49.000] The text has been redacted
[00:34:49.000 --> 00:34:51.000] This is the end of the audio.
- end of program -
```
This commit changes the calculation of the `seek_end` variable to
only add `seek_start` if a custom `duration_ms` is provided.
Otherwise, it defaults to the end of the file.
Signed-off-by: Thijs Raymakers <redacted>
Georgi Gerganov [Sat, 29 Apr 2023 09:32:18 +0000 (12:32 +0300)]
tests : add "threads" to run-tests.sh
Georgi Gerganov [Sat, 29 Apr 2023 09:32:05 +0000 (12:32 +0300)]
extra : add sync-ggml.sh script
Georgi Gerganov [Sat, 29 Apr 2023 09:31:52 +0000 (12:31 +0300)]
ggml : sync latest ggml + llama.cpp updates (quantization)
Zollner [Sat, 29 Apr 2023 08:00:20 +0000 (16:00 +0800)]
whisper.android : add some tips (#816)
Georgi Gerganov [Sat, 29 Apr 2023 07:55:24 +0000 (10:55 +0300)]
build : add WHISPER_COREML_ALLOW_FALLBACK to make / CMake (#812)
Canis Lupus [Sat, 29 Apr 2023 07:49:02 +0000 (08:49 +0100)]
whisper : allow non-CoreML fallback when Core ML cannot be loaded (#812)
if the Core ML model cannot be loaded, continue without Core ML instead of
returning. This allows a single build to transcribe using Core ML models
where available, and regular models when not.
Georgi Gerganov [Sat, 29 Apr 2023 07:42:14 +0000 (10:42 +0300)]
whisper : fix bug from previous commit
Georgi Gerganov [Sat, 29 Apr 2023 07:36:50 +0000 (10:36 +0300)]
whisper : avoid designated initializers
AsukaMinato [Sat, 29 Apr 2023 07:06:25 +0000 (16:06 +0900)]
minor : improve C++ and Python style (#768)
* use some STL functions
* use self.field than setattr, use pathlib.Path
* recover some format
* const some iter
* Keep the original
* 2 space
Georgi Gerganov [Fri, 28 Apr 2023 19:41:29 +0000 (22:41 +0300)]
readme : add logo
Laytan Laats [Sun, 23 Apr 2023 16:01:59 +0000 (18:01 +0200)]
main : escape quotes in csv output (#815)
Taras Glek [Sun, 23 Apr 2023 14:00:30 +0000 (17:00 +0300)]
stream : flush upon finishing inference (#811)
Philipp Zabel [Sun, 23 Apr 2023 13:52:52 +0000 (15:52 +0200)]
examples : add missing #include <cstdint> (#798)
common.cpp uses uint8_t and uint64_t, which are defined in <cstdint>.
Tauseef Mohiuddin [Sun, 23 Apr 2023 13:47:30 +0000 (08:47 -0500)]
main : update escape_double_quotes() function (#776)
Updated the escape_double_quotes() function such that the function now escapes both double quotes and backslashes in the input string.
Changes Made:
- Renamed the function to escape_quotes_and_backslashes
- Modified the condition in the first loop to increment the value of 'escaped_length' for both double quotes and backslashes.
- Modified the condition in second loop to add a backslash before the current character if it is a double quote or a backslash.
Resolves: #769
Georgi Gerganov [Sat, 15 Apr 2023 14:30:44 +0000 (17:30 +0300)]
release : v1.3.0
Georgi Gerganov [Sat, 15 Apr 2023 14:18:43 +0000 (17:18 +0300)]
whisper : pad audio instead of spectrogram (#579)
Also, fallback only if more temperatures are available and if we are
at least 3 seconds before the end of the audio
Georgi Gerganov [Sat, 15 Apr 2023 13:04:07 +0000 (16:04 +0300)]
whisper : restore decoder temperature fallbacks
I disabled this because there were many complaints about slow decoding.
The current implementation does not allow batching the decoders when
using the "best of" or "beam size" parameters, so the decoding time is
proportional to the number of decoders, which is obviously not great.
However, now there are even more complaints about wrong decodings and
repetition.
So, making a compromise by re-enabling the fallbacks, but defaulting to
just 2 "best of" / "beam size" decoders. Also, the temperature step is
increased from 0.2 to 0.4 - i.e. from maximum of 5 fallbacks to maximum
of 2.
Also, the stream example now has fallbacks enabled by default.
close #471 #477 #508 #612 #719 #731
Jhen-Jie Hong [Sat, 15 Apr 2023 11:21:58 +0000 (19:21 +0800)]
ggml, ci : fix build on whisper.android (ARM_NEON) + add CI (#764)
* ggml : fix undefined symbol by remove inline handle
* ggml : make own ggml_aligned_malloc function
* ci: add ios/android build
Georgi Gerganov [Sat, 15 Apr 2023 11:18:46 +0000 (14:18 +0300)]
whisper : slightly faster Log Mel computation + n-1 FFT threads (#568)
Georgi Gerganov [Sat, 15 Apr 2023 10:30:36 +0000 (13:30 +0300)]
readme : fix link
Georgi Gerganov [Sat, 15 Apr 2023 10:30:07 +0000 (13:30 +0300)]
readme : add usage instructions for Core ML
Georgi Gerganov [Sat, 15 Apr 2023 10:21:27 +0000 (13:21 +0300)]
whisper : add Core ML support (#566)
* coreml : use Core ML encoder inference
* coreml : simlpify whisper_encode + log messages
* whisper : resolve rebase conflicts
* coreml : add scripts for CoreML model generation
* bench-all : recognize COREML flag
Maximiliano Levi [Fri, 14 Apr 2023 19:35:34 +0000 (16:35 -0300)]
whisper : do not launch log_mel threads when n_thread is 1 (#763)
AfryMask [Fri, 14 Apr 2023 17:35:03 +0000 (01:35 +0800)]
whisper : fix the bug related to word splitting errors in the "tokenize" function. (#760)
Co-authored-by: AfryMask <redacted>
Aaron Taylor [Fri, 14 Apr 2023 17:24:00 +0000 (13:24 -0400)]
readme : add SwiftWhisper to listed bindings (#755)
Georgi Gerganov [Fri, 14 Apr 2023 17:13:47 +0000 (20:13 +0300)]
gitignore : add .test
Bader-eddine Ouaich [Fri, 14 Apr 2023 17:05:56 +0000 (17:05 +0000)]
whisper : fix potential memory leaks (#740)
* fix potential memory leak if whisper_init_state failed
* fix potential memory leak if gpt2_init failed
Anton Kostin [Fri, 14 Apr 2023 17:04:42 +0000 (00:04 +0700)]
license : update year (#739)
GitAritron [Fri, 14 Apr 2023 17:03:16 +0000 (20:03 +0300)]
whisper : fix typos in whisper.h (#737)
Fixed a couple of typos (in comments, so nothing major). Keep up the great work 😄
Ali Alameh [Fri, 14 Apr 2023 17:02:18 +0000 (20:02 +0300)]
stream : support language auto-detect (#501)
#445 fix Language auto-detect "auto" flag does not work using the stream tool
Alex Evgrashin [Fri, 14 Apr 2023 16:59:44 +0000 (19:59 +0300)]
readme : add unity bindings (#733)
DGdev91 [Fri, 14 Apr 2023 16:53:58 +0000 (18:53 +0200)]
talk, talk-llama : add basic example script for eleven-labs tts (#728)
Ivan Gorin [Fri, 14 Apr 2023 16:50:39 +0000 (19:50 +0300)]
models : change convert-pt-to-ggml to use .tiktoken tokenizer files (#725)
LittleLoli [Fri, 14 Apr 2023 16:36:38 +0000 (00:36 +0800)]
cmake : add msvc compiler args /utf-8 fix error C3688 (#721)
* force msvc compiler use utf-8 encode
* only enable on msvc
Maciek [Fri, 14 Apr 2023 16:36:09 +0000 (18:36 +0200)]
talk-llama : correct default speak.sh path (#720)
There is `speak.sh` file in `./examples/talk-llama` as described in README.
However `./examples/talk/speak.sh` is used in `talk-llama.cpp`, this commit corrects that.
LittleLoli [Fri, 14 Apr 2023 16:35:33 +0000 (00:35 +0800)]
main : add lrc output support (#718)
* add lrc output support.
* fix wrong comment
Sam [Fri, 14 Apr 2023 16:33:06 +0000 (16:33 +0000)]
readme : make the quick start instructions clearer. (#716)
Users wanting to make use of this implementation of the whisper model with no prior knowledge of C/C++ may download the Whisper model but fail to use of the "make" command as specified given that they forgot or didn't know they needed to clone the repository first. Hope this modification clears things up.