]> git.djapps.eu Git - pkg/ggml/sources/whisper.cpp/commit
model : Qwen3 Next (llama/16095)
authorPiotr Wilkin (ilintar) <redacted>
Fri, 28 Nov 2025 11:02:56 +0000 (12:02 +0100)
committerGeorgi Gerganov <redacted>
Fri, 12 Dec 2025 15:53:11 +0000 (17:53 +0200)
commit43441ff58aa9e29d75fb99d288598310ea3b881d
treef83b4f98c5e4be3b6b198c1dc72181907b97466b
parent37e4c2ed3a77826e592f3dddcdf7c16ca3b0c72c
model : Qwen3 Next (llama/16095)

* Qwen3 Next - cleaned up version

* Whitespaces and stuff

* Correct minor errors

* Update src/llama-model.cpp

Co-authored-by: Sigbjørn Skjæret <redacted>
* Misc. fixes.

* Clean up code, add missing hybrid qualifier

* Did someone transpose the SOLVE_TRI result matrix? Perhaps...

* Whitespace

* Proper tensors for cb calls

* Use llama-graph.h vertical alignment

* BROKEN: chunking

* Set new tensors as inputs.

* Proper chunk logic

* It's the circle of life...

* More shenanigans for n_seq > 1

* Nail in the coffin?

* Fix Windows build

* Eh, one fails on Windows, the other fails on Mac... just use general capture.

* quant : cleanup

* model : cleanup

* qwen3 : cleanup

* cont : cleanup

* cont : cleanup

* ggml : revert change

* qwen3 : cleanup

* cont : cleanup

* Readd cmath

* qwen3 : fix typo

* Update convert_hf_to_gguf.py

Co-authored-by: Sigbjørn Skjæret <redacted>
* Usual suspects

* fix my bad suggestion

---------

Co-authored-by: Sigbjørn Skjæret <redacted>
Co-authored-by: Georgi Gerganov <redacted>
ggml/src/ggml-cpu/ops.cpp