You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+4Lines changed: 4 additions & 0 deletions
Original file line number
Diff line number
Diff line change
@@ -11,6 +11,10 @@ Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++
11
11
12
12
### Hot topics
13
13
14
+
-## IMPORTANT: Tokenizer fixes and API change (developers and projects using `llama.cpp` built-in tokenization must read): https://github.com/ggerganov/llama.cpp/pull/2810
15
+
16
+
-## GGUFv2 adds support for 64-bit sizes + backwards compatible: https://github.com/ggerganov/llama.cpp/pull/2821
17
+
14
18
- Added support for Falcon models: https://github.com/ggerganov/llama.cpp/pull/2717
15
19
16
20
- A new file format has been introduced: [GGUF](https://github.com/ggerganov/llama.cpp/pull/2398)
0 commit comments