|
63 | 63 | "model = AzureAIChatCompletionsModel(\n", |
64 | 64 | " endpoint=os.environ[\"AZURE_INFERENCE_ENDPOINT\"],\n", |
65 | 65 | " credential=os.environ[\"AZURE_INFERENCE_CREDENTIAL\"],\n", |
66 | | - " model_name=\"mistral-large-2407\",\n", |
| 66 | + " model=\"mistral-large-2407\",\n", |
67 | 67 | ")" |
68 | 68 | ] |
69 | 69 | }, |
|
189 | 189 | "producer = AzureAIChatCompletionsModel(\n", |
190 | 190 | " endpoint=os.environ[\"AZURE_INFERENCE_ENDPOINT\"],\n", |
191 | 191 | " credential=os.environ[\"AZURE_INFERENCE_CREDENTIAL\"],\n", |
192 | | - " model_name=\"mistral-large-2407\",\n", |
| 192 | + " model=\"mistral-large-2407\",\n", |
193 | 193 | ")\n", |
194 | 194 | "\n", |
195 | 195 | "verifier = AzureAIChatCompletionsModel(\n", |
196 | 196 | " endpoint=os.environ[\"AZURE_INFERENCE_ENDPOINT\"],\n", |
197 | 197 | " credential=os.environ[\"AZURE_INFERENCE_CREDENTIAL\"],\n", |
198 | | - " model_name=\"mistral-small\",\n", |
| 198 | + " model=\"mistral-small\",\n", |
199 | 199 | ")" |
200 | 200 | ] |
201 | 201 | }, |
|
372 | 372 | "model = AzureAIChatCompletionsModel(\n", |
373 | 373 | " endpoint=os.environ[\"AZURE_INFERENCE_ENDPOINT\"],\n", |
374 | 374 | " credential=os.environ[\"AZURE_INFERENCE_CREDENTIAL\"],\n", |
375 | | - " model_name=\"mistral-large-2407\",\n", |
| 375 | + " model=\"mistral-large-2407\",\n", |
376 | 376 | " client_kwargs={\"logging_enable\": True},\n", |
377 | 377 | ")" |
378 | 378 | ] |
|
0 commit comments