Skip to content

Bonsai 4B: Uncaught Error: Can't create a session. ERROR_CODE: 6, ERROR_MESSAGE: std::bad_alloc #1653

@thomas-0816

Description

@thomas-0816

System Info

package.json:

{
  "dependencies": {
    "@huggingface/transformers": "^4.1.0"
  }
}

no bundlers used
KUbuntu 24.04
Chromium 146.0.7680.177 (chromium --enable-unsafe-webgpu --enable-features=Vulkan)
Waterfox 6.6.11 (dom.webgpu.enabled=true)
Ryzen 7 PRO 7840U (32 GB Ram, GTT configured with 24 GB)

Environment/Platform

  • Website/web-app
  • Browser extension
  • Server-side (e.g., Node.js, Deno, Bun)
  • Desktop app (e.g., Electron)
  • Other (e.g., VSCode extension)

Description

Error comes when reading Bonsai 4B or 8B, no error on Bonsai 1.7B, no error on Qwen3 0.6B.

package.json:

{
  "dependencies": {
    "@huggingface/transformers": "^4.1.0"
  }
}

example code:

<!DOCTYPE html>
<html lang="de">
    <head>
        <title>Test</title>
        <link rel="icon" href="data:,">
    </head>
    <body>
        <pre id="output"></pre>
        <script type="module">
            import { pipeline, env, TextStreamer } from "/node_modules/@huggingface/transformers/dist/transformers.js";
            env.localModelPath = '/models/'
            env.backends.onnx.wasm.wasmPaths = '/node_modules/onnxruntime-web/dist/';
            env.allowRemoteModels = false;
            env.allowLocalModels = true;
            env.useBrowserCache = false;
            env.useWasmCache = false;
            const output = document.getElementById('output');
            output.innerText += `${new Date().toLocaleTimeString()} Bonsai-4B-ONNX ...\n`;
            const model = await pipeline("text-generation", "Bonsai-4B-ONNX", { dtype: "q1f16", device: "webgpu" }); // or dtype: "q1"
            const messages = [
                { role: "system", content: "You are a helpful assistant." },
                { role: "user", content: "Write a quick sort algorithm in php 8, only the code. /no_think" },
            ];
            output.innerText += `${new Date().toLocaleTimeString()} loaded.\n`;
            const result = await model(messages, {
                temperature: 0.7, top_k: 20, top_p: 0.8, min_p: 0.0,
                max_new_tokens: 4096,
                do_sample: true,
                streamer: new TextStreamer(model.tokenizer, {
                    skip_prompt: false, skip_special_tokens: true,
                    callback_function: (text) => output.innerText += text
                }),
            });
            output.innerText += `\n${new Date().toLocaleTimeString()} done.`;
            model.dispose();
        </script>
    </body>
</html>

Waterfox 6.6.11 error message:

Uncaught Error: Can't create a session. ERROR_CODE: 6, ERROR_MESSAGE: std::bad_alloc
    $ http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:10106
    Qt http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11007
    Ds http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11306
    loadModel http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11380
    createInferenceSessionHandler http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11434
    create http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:8093
    load http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:12420
    promise callback*createInferenceSession http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:12425
    constructSessions http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:23319
    constructSessions http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:23310
    from_pretrained http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:25194
    from_pretrained http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:32197
    pipeline2 http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:33953
    async* http://127.0.0.1:8081/transformers.html:19
transformers.js:10106:13
    $ http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:10106
    Qt http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11007
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    Ds http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11306
    loadModel http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11380
    createInferenceSessionHandler http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:11434
    create http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:8093
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    load http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:12420
    (Async: promise callback)
    createInferenceSession http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:12425
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    constructSessions http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:23319
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    map self-hosted:175
    constructSessions http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:23310
    from_pretrained http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:25194
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    from_pretrained http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:32197
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    pipeline2 http://127.0.0.1:8081/node_modules/@huggingface/transformers/dist/transformers.js:33953
    InterpretGeneratorResume self-hosted:1425
    AsyncFunctionNext self-hosted:800
    (Async: async)
    <anonymous> http://127.0.0.1:8081/transformers.html:19

Chromium 146.0.7680.177 error message:

transformers.js:10106 Uncaught Error: Can't create a session. ERROR_CODE: 6, ERROR_MESSAGE: std::bad_alloc
    at $ (transformers.js:10106:13)
    at Qt (transformers.js:11007:88)
    at async pr.loadModel (transformers.js:11380:182)
    at async mr.createInferenceSessionHandler (transformers.js:11434:14)
    at async a.create (transformers.js:8093:37)
    at async createInferenceSession (transformers.js:12425:19)
    at async transformers.js:23319:25
    at async Promise.all (index 0)
    at async constructSessions (transformers.js:23309:5)
    at async Promise.all (index 0)

Setup:

bun install @huggingface/transformers
mkdir models; cd models
apt-get install git-lfs
git clone https://huggingface.co/onnx-community/Bonsai-1.7B-ONNX
git clone https://huggingface.co/onnx-community/Bonsai-4B-ONNX
git clone https://huggingface.co/onnx-community/Bonsai-8B-ONNX

Reproduction

see description

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type
    No fields configured for issues without a type.

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions