@@ -51,13 +51,23 @@ class ParametersEvaluationClassifyParametersJudge(TypedDict, total=False):
5151 external_base_url : str
5252 """Base URL for external judge models. Must be OpenAI-compatible base URL."""
5353
54+ max_tokens : int
55+ """Maximum number of tokens the judge model can generate.
56+
57+ Defaults to 32768. Increase for reasoning models (e.g. Gemini, o-series) that
58+ consume output token budget for chain-of-thought.
59+ """
60+
5461 num_workers : int
5562 """Number of concurrent workers for inference requests.
5663
5764 Overrides the default concurrency for this model. Useful for tuning throughput
5865 when using proxy endpoints (e.g. OpenRouter) or rate-limited external APIs.
5966 """
6067
68+ temperature : float
69+ """Sampling temperature for the judge model. Defaults to 0.05."""
70+
6171
6272class ParametersEvaluationClassifyParametersModelToEvaluateEvaluationModelRequest (TypedDict , total = False ):
6373 input_template : Required [str ]
@@ -129,13 +139,23 @@ class ParametersEvaluationScoreParametersJudge(TypedDict, total=False):
129139 external_base_url : str
130140 """Base URL for external judge models. Must be OpenAI-compatible base URL."""
131141
142+ max_tokens : int
143+ """Maximum number of tokens the judge model can generate.
144+
145+ Defaults to 32768. Increase for reasoning models (e.g. Gemini, o-series) that
146+ consume output token budget for chain-of-thought.
147+ """
148+
132149 num_workers : int
133150 """Number of concurrent workers for inference requests.
134151
135152 Overrides the default concurrency for this model. Useful for tuning throughput
136153 when using proxy endpoints (e.g. OpenRouter) or rate-limited external APIs.
137154 """
138155
156+ temperature : float
157+ """Sampling temperature for the judge model. Defaults to 0.05."""
158+
139159
140160class ParametersEvaluationScoreParametersModelToEvaluateEvaluationModelRequest (TypedDict , total = False ):
141161 input_template : Required [str ]
@@ -210,13 +230,23 @@ class ParametersEvaluationCompareParametersJudge(TypedDict, total=False):
210230 external_base_url : str
211231 """Base URL for external judge models. Must be OpenAI-compatible base URL."""
212232
233+ max_tokens : int
234+ """Maximum number of tokens the judge model can generate.
235+
236+ Defaults to 32768. Increase for reasoning models (e.g. Gemini, o-series) that
237+ consume output token budget for chain-of-thought.
238+ """
239+
213240 num_workers : int
214241 """Number of concurrent workers for inference requests.
215242
216243 Overrides the default concurrency for this model. Useful for tuning throughput
217244 when using proxy endpoints (e.g. OpenRouter) or rate-limited external APIs.
218245 """
219246
247+ temperature : float
248+ """Sampling temperature for the judge model. Defaults to 0.05."""
249+
220250
221251class ParametersEvaluationCompareParametersModelAEvaluationModelRequest (TypedDict , total = False ):
222252 input_template : Required [str ]
0 commit comments