Skip to content

Commit 6fd6a0a

Browse files
authored
New version: ggml.llamacpp version b8334 (microsoft#348618)
1 parent 4f16812 commit 6fd6a0a

3 files changed

Lines changed: 98 additions & 0 deletions

File tree

Lines changed: 30 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,30 @@
1+
# Created with komac v2.15.0
2+
# yaml-language-server: $schema=https://aka.ms/winget-manifest.installer.1.12.0.schema.json
3+
4+
PackageIdentifier: ggml.llamacpp
5+
PackageVersion: b8334
6+
InstallerType: zip
7+
NestedInstallerType: portable
8+
NestedInstallerFiles:
9+
- RelativeFilePath: llama-batched-bench.exe
10+
- RelativeFilePath: llama-bench.exe
11+
- RelativeFilePath: llama-cli.exe
12+
- RelativeFilePath: llama-gguf-split.exe
13+
- RelativeFilePath: llama-imatrix.exe
14+
- RelativeFilePath: llama-mtmd-cli.exe
15+
- RelativeFilePath: llama-perplexity.exe
16+
- RelativeFilePath: llama-quantize.exe
17+
- RelativeFilePath: llama-server.exe
18+
- RelativeFilePath: llama-tokenize.exe
19+
- RelativeFilePath: llama-tts.exe
20+
Dependencies:
21+
PackageDependencies:
22+
- PackageIdentifier: Microsoft.VCRedist.2015+.x64
23+
ReleaseDate: 2026-03-14
24+
ArchiveBinariesDependOnPath: true
25+
Installers:
26+
- Architecture: x64
27+
InstallerUrl: https://github.com/ggml-org/llama.cpp/releases/download/b8334/llama-b8334-bin-win-vulkan-x64.zip
28+
InstallerSha256: 9F1C11B6875416BBE08F1528C0E4C8C49C82F0056A794297774572884312BDE6
29+
ManifestType: installer
30+
ManifestVersion: 1.12.0
Lines changed: 60 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,60 @@
1+
# Created with komac v2.15.0
2+
# yaml-language-server: $schema=https://aka.ms/winget-manifest.defaultLocale.1.12.0.schema.json
3+
4+
PackageIdentifier: ggml.llamacpp
5+
PackageVersion: b8334
6+
PackageLocale: en-US
7+
Publisher: ggml
8+
PublisherUrl: https://github.com/ggml-org
9+
PublisherSupportUrl: https://github.com/ggml-org/llama.cpp/issues
10+
PackageName: llama.cpp
11+
PackageUrl: https://github.com/ggml-org/llama.cpp
12+
License: MIT
13+
LicenseUrl: https://github.com/ggml-org/llama.cpp/blob/HEAD/LICENSE
14+
ShortDescription: LLM inference in C/C++
15+
Tags:
16+
- ggml
17+
- llama
18+
ReleaseNotes: |-
19+
tools : enable kvu in perplexity for hellaswag, winogrande, multiple-choice (#19954)
20+
llama-perplexity -hf unsloth/Qwen3-0.6B-GGUF:Q4_K_M -f winogrande-debiased-eval.csv --winogrande
21+
winogrande_score : tokenizing selected tasks
22+
winogrande_score : calculating winogrande score over selected tasks.
23+
split_equal: sequential split is not supported when there are coupled sequences in the input batch (you may need to use the -kvu flag)
24+
decode: failed to find a memory slot for batch of size 46
25+
failed to decode the batch, n_batch = 2048, ret = 1
26+
winogrande_score: llama_decode() failed
27+
same for hellaswag:
28+
split_equal: sequential split is not supported when there are coupled sequences in the input batch (you may need to use the -kvu flag)
29+
decode: failed to find a memory slot for batch of size 99
30+
failed to decode the batch, n_batch = 2048, ret = 1
31+
hellaswag_score: llama_decode() failed
32+
Signed-off-by: Adrien Gallouët angt@huggingface.co
33+
macOS/iOS:
34+
- macOS Apple Silicon (arm64)
35+
- macOS Intel (x64)
36+
- iOS XCFramework
37+
Linux:
38+
- Ubuntu x64 (CPU)
39+
- Ubuntu x64 (Vulkan)
40+
- Ubuntu x64 (ROCm 7.2)
41+
- Ubuntu s390x (CPU)
42+
Windows:
43+
- Windows x64 (CPU)
44+
- Windows arm64 (CPU)
45+
- Windows x64 (CUDA 12) - CUDA 12.4 DLLs
46+
- Windows x64 (CUDA 13) - CUDA 13.1 DLLs
47+
- Windows x64 (Vulkan)
48+
- Windows x64 (SYCL)
49+
- Windows x64 (HIP)
50+
openEuler:
51+
- openEuler x86 (310p)
52+
- openEuler x86 (910b, ACL Graph)
53+
- openEuler aarch64 (310p)
54+
- openEuler aarch64 (910b, ACL Graph)
55+
ReleaseNotesUrl: https://github.com/ggml-org/llama.cpp/releases/tag/b8334
56+
Documentations:
57+
- DocumentLabel: Wiki
58+
DocumentUrl: https://github.com/ggml-org/llama.cpp/wiki
59+
ManifestType: defaultLocale
60+
ManifestVersion: 1.12.0
Lines changed: 8 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,8 @@
1+
# Created with komac v2.15.0
2+
# yaml-language-server: $schema=https://aka.ms/winget-manifest.version.1.12.0.schema.json
3+
4+
PackageIdentifier: ggml.llamacpp
5+
PackageVersion: b8334
6+
DefaultLocale: en-US
7+
ManifestType: version
8+
ManifestVersion: 1.12.0

0 commit comments

Comments
 (0)