You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This example demonstrates how to call the inference endpoint using a minimal client that only uses only an inference key (no client credentials needed).
This example demonstrates how to call the inference async endpoint using a minimal client that only uses only an inference key (no client credentials needed).
# Create an inference client that uses only the inference key, without client credentials
9
+
inference_client=InferenceClient(
10
+
inference_key=DATACRUNCH_INFERENCE_KEY,
11
+
endpoint_base_url=DATACRUNCH_ENDPOINT_BASE_URL
12
+
)
13
+
14
+
# Make a synchronous request to the endpoint.
15
+
# This example demonstrates calling a SGLang deployment which serves LLMs using an OpenAI-compatible API format
16
+
data= {
17
+
"model": "deepseek-ai/deepseek-llm-7b-chat",
18
+
"prompt": "Is consciousness fundamentally computational, or is there something more to subjective experience that cannot be reduced to information processing?",
0 commit comments