Skip to content

Commit 53ed6de

Browse files
Merge pull request #504 from janhq/update-dev-from-master-2026-05-03-01-07
Sync master with upstream release b9010
2 parents febf7fb + d05fe1d commit 53ed6de

42 files changed

Lines changed: 4000 additions & 933 deletions

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

.github/ISSUE_TEMPLATE/010-bug-compilation.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,8 @@ body:
1212
after recreating the CMake build directory and with `-DGGML_CCACHE=OFF`.
1313
If the compilation succeeds with ccache disabled you should be able to permanently fix the issue
1414
by clearing `~/.cache/ccache` (on Linux).
15+
16+
Please fill out this template yourself, copypasting language model outputs is [strictly prohibited](https://github.com/ggml-org/llama.cpp/blob/master/CONTRIBUTING.md#ai-usage-policy).
1517
- type: textarea
1618
id: commit
1719
attributes:

.github/ISSUE_TEMPLATE/011-bug-results.yml

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
name: Bug (model use)
2-
description: Something goes wrong when using a model (in general, not specific to a single llama.cpp module).
2+
description: Something goes wrong when running a model (crashes, garbled outputs, etc.).
33
title: "Eval bug: "
44
labels: ["bug-unconfirmed", "model evaluation"]
55
body:
@@ -12,6 +12,8 @@ body:
1212
If you encountered the issue while using an external UI (e.g. ollama),
1313
please reproduce your issue using one of the examples/binaries in this repository.
1414
The `llama-completion` binary can be used for simple and reproducible model inference.
15+
16+
Please fill out this template yourself, copypasting language model outputs is [strictly prohibited](https://github.com/ggml-org/llama.cpp/blob/master/CONTRIBUTING.md#ai-usage-policy).
1517
- type: textarea
1618
id: version
1719
attributes:

.github/ISSUE_TEMPLATE/019-bug-misc.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -10,6 +10,8 @@ body:
1010
This issue template is intended for miscellaneous bugs that don't fit into any other category.
1111
If you encountered the issue while using an external UI (e.g. ollama),
1212
please reproduce your issue using one of the examples/binaries in this repository.
13+
14+
Please fill out this template yourself, copypasting language model outputs is [strictly prohibited](https://github.com/ggml-org/llama.cpp/blob/master/CONTRIBUTING.md#ai-usage-policy).
1315
- type: textarea
1416
id: version
1517
attributes:

.github/ISSUE_TEMPLATE/020-enhancement.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@ body:
88
value: |
99
[Please post your idea first in Discussion if there is not yet a consensus for this enhancement request. This will help to keep this issue tracker focused on enhancements that the community has agreed needs to be implemented.](https://github.com/ggml-org/llama.cpp/discussions/categories/ideas)
1010
11+
Please fill out this template yourself, copypasting language model outputs is [strictly prohibited](https://github.com/ggml-org/llama.cpp/blob/master/CONTRIBUTING.md#ai-usage-policy).
12+
1113
- type: checkboxes
1214
id: prerequisites
1315
attributes:

.github/ISSUE_TEMPLATE/030-research.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -8,6 +8,8 @@ body:
88
value: |
99
Don't forget to check for any [duplicate research issue tickets](https://github.com/ggml-org/llama.cpp/issues?q=is%3Aopen+is%3Aissue+label%3A%22research+%F0%9F%94%AC%22)
1010
11+
Please fill out this template yourself, copypasting language model outputs is [strictly prohibited](https://github.com/ggml-org/llama.cpp/blob/master/CONTRIBUTING.md#ai-usage-policy).
12+
1113
- type: checkboxes
1214
id: research-stage
1315
attributes:

.github/ISSUE_TEMPLATE/040-refactor.yml

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -9,6 +9,8 @@ body:
99
Don't forget to [check for existing refactor issue tickets](https://github.com/ggml-org/llama.cpp/issues?q=is%3Aopen+is%3Aissue+label%3Arefactoring) in case it's already covered.
1010
Also you may want to check [Pull request refactor label as well](https://github.com/ggml-org/llama.cpp/pulls?q=is%3Aopen+is%3Apr+label%3Arefactoring) for duplicates too.
1111
12+
Please fill out this template yourself, copypasting language model outputs is [strictly prohibited](https://github.com/ggml-org/llama.cpp/blob/master/CONTRIBUTING.md#ai-usage-policy).
13+
1214
- type: textarea
1315
id: background-description
1416
attributes:

convert_hf_to_gguf.py

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -13232,17 +13232,18 @@ class LazyTorchTensor(gguf.LazyBase):
1323213232
}
1323313233

1323413234
# only used when byteswapping data. Only correct size is needed
13235+
# TODO: uncomment uint64, uint32, and uint16, ref: https://github.com/pytorch/pytorch/issues/58734
1323513236
_dtype_byteswap_map: dict[torch.dtype, type] = {
1323613237
torch.float64: np.float64,
1323713238
torch.float32: np.float32,
1323813239
torch.bfloat16: np.float16,
1323913240
torch.float16: np.float16,
1324013241
torch.int64: np.int64,
13241-
torch.uint64: np.uint64,
13242+
# torch.uint64: np.uint64,
1324213243
torch.int32: np.int32,
13243-
torch.uint32: np.uint32,
13244+
# torch.uint32: np.uint32,
1324413245
torch.int16: np.int16,
13245-
torch.uint16: np.uint16,
13246+
# torch.uint16: np.uint16,
1324613247
torch.int8: np.int8,
1324713248
torch.uint8: np.uint8,
1324813249
torch.bool: np.uint8,

ggml/CMakeLists.txt

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ project("ggml" C CXX ASM)
55
### GGML Version
66
set(GGML_VERSION_MAJOR 0)
77
set(GGML_VERSION_MINOR 10)
8-
set(GGML_VERSION_PATCH 1)
8+
set(GGML_VERSION_PATCH 2)
99
set(GGML_VERSION_BASE "${GGML_VERSION_MAJOR}.${GGML_VERSION_MINOR}.${GGML_VERSION_PATCH}")
1010

1111
list(APPEND CMAKE_MODULE_PATH "${CMAKE_CURRENT_SOURCE_DIR}/cmake/")

ggml/src/ggml-cuda/ggml-cuda.cu

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -5431,8 +5431,8 @@ ggml_backend_reg_t ggml_backend_cuda_reg() {
54315431
CUDA_CHECK(cudaGetDeviceProperties(&prop, i));
54325432
dev_ctx->description = prop.name;
54335433

5434-
char pci_bus_id[16] = {};
5435-
snprintf(pci_bus_id, sizeof(pci_bus_id), "%04x:%02x:%02x.0", prop.pciDomainID, prop.pciBusID, prop.pciDeviceID);
5434+
char pci_bus_id[32] = {};
5435+
CUDA_CHECK(cudaDeviceGetPCIBusId(pci_bus_id, sizeof(pci_bus_id), i));
54365436
dev_ctx->pci_bus_id = pci_bus_id;
54375437
dev_ctx->op_offload_min_batch_size = min_batch_size;
54385438

ggml/src/ggml-cuda/vendors/hip.h

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -55,6 +55,7 @@
5555
#define cudaDeviceDisablePeerAccess hipDeviceDisablePeerAccess
5656
#define cudaDeviceEnablePeerAccess hipDeviceEnablePeerAccess
5757
#define cudaDeviceGetAttribute hipDeviceGetAttribute
58+
#define cudaDeviceGetPCIBusId hipDeviceGetPCIBusId
5859
#define cudaDeviceProp hipDeviceProp_t
5960
#define cudaDeviceSynchronize hipDeviceSynchronize
6061
#define cudaError_t hipError_t

0 commit comments

Comments
 (0)