Skip to content

Commit 65c570e

Browse files
committed
fix: bugs
1 parent 492dc59 commit 65c570e

1 file changed

Lines changed: 3 additions & 1 deletion

File tree

src/bindings/Llama.ts

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -137,7 +137,7 @@ export class Llama {
137137
this._supportsMmap = bindings.getSupportsMmap();
138138
this._gpuSupportsMmap = bindings.getGpuSupportsMmap();
139139
this._supportsMlock = bindings.getSupportsMlock();
140-
this._mathCores = bindings.getMathCores();
140+
this._mathCores = Math.floor(bindings.getMathCores());
141141
this._consts = bindings.getConsts();
142142
this._vramOrchestrator = vramOrchestrator;
143143
this._vramPadding = vramPadding;
@@ -690,6 +690,8 @@ function getTransformedLogLevel(level: LlamaLogLevel, message: string, gpu: Buil
690690
return LlamaLogLevel.info;
691691
else if (level === LlamaLogLevel.warn && message.startsWith("llama_init_from_model: model default pooling_type is [0], but [-1] was specified"))
692692
return LlamaLogLevel.info;
693+
else if (level === LlamaLogLevel.warn && message.startsWith("llama_context: n_ctx_seq (") && message.endsWith("- the full capacity of the model will not be utilized"))
694+
return LlamaLogLevel.info;
693695
else if (gpu === false && level === LlamaLogLevel.warn && message.startsWith("llama_adapter_lora_init_impl: lora for '") && message.endsWith("' cannot use buft 'CPU_REPACK', fallback to CPU"))
694696
return LlamaLogLevel.info;
695697
else if (gpu === "metal" && level === LlamaLogLevel.warn && message.startsWith("ggml_metal_device_init: tensor API disabled for"))

0 commit comments

Comments
 (0)