Skip to content

Commit 1465ff0

Browse files
authored
reduce max context length (#2)
1 parent c8afa6f commit 1465ff0

1 file changed

Lines changed: 3 additions & 1 deletion

File tree

src/provider.ts

Lines changed: 3 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -21,7 +21,9 @@ const DEFAULT_MAX_OUTPUT_TOKENS = 16000;
2121
// Token estimates for gpt‑oss are correct as we use the appropriate tokenizer.
2222
// For Qwen we must first create the tokenizer from the model, as it does not use tiktoken.
2323
// As a workaround, we also use the gpt‑oss tokenizer for now and reduce the max context length here.
24-
const DEFAULT_CONTEXT_LENGTH = 120000;
24+
//
25+
// Further reduced to avoid running into rate limits for free users.
26+
const DEFAULT_CONTEXT_LENGTH = 96000;
2527

2628
/**
2729
* VS Code Chat provider backed by Privatemode OpenAI API.

0 commit comments

Comments
 (0)