Skip to content

Commit d44be10

Browse files
committed
Regenerate template files
1 parent 936cb5a commit d44be10

6 files changed

Lines changed: 12 additions & 12 deletions

File tree

src/inference_endpoint/config/templates/concurrency_template.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,5 +17,5 @@ settings:
1717
type: concurrency # Load pattern type | options: max_throughput, poisson, concurrency, burst, step
1818
target_concurrency: 32 # Concurrent requests
1919
endpoint_config:
20-
endpoints: # Endpoint URL(s)
21-
- 'http://localhost:8000'
20+
endpoints: # Endpoint URL(s). Must include scheme, e.g. 'http://host:port'.
21+
- '<ENDPOINT_URL eg: http://localhost:8000>'

src/inference_endpoint/config/templates/concurrency_template_full.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -69,8 +69,8 @@ settings:
6969
min_required_connections: -1 # Min connections to initialize (-1=auto, 0=disabled)
7070
worker_gc_mode: relaxed # Worker GC strategy | options: disabled, relaxed, system
7171
endpoint_config:
72-
endpoints: # Endpoint URL(s)
73-
- 'http://localhost:8000'
72+
endpoints: # Endpoint URL(s). Must include scheme, e.g. 'http://host:port'.
73+
- '<ENDPOINT_URL eg: http://localhost:8000>'
7474
api_key: null # API key
7575
api_type: openai # API type: openai or sglang | options: openai, sglang
7676
report_dir: null # Report output directory

src/inference_endpoint/config/templates/offline_template.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,5 +14,5 @@ settings:
1414
max_duration_ms: 0 # Maximum test duration in ms (0 for no limit)
1515
n_samples_to_issue: null # Sample count override
1616
endpoint_config:
17-
endpoints: # Endpoint URL(s)
18-
- 'http://localhost:8000'
17+
endpoints: # Endpoint URL(s). Must include scheme, e.g. 'http://host:port'.
18+
- '<ENDPOINT_URL eg: http://localhost:8000>'

src/inference_endpoint/config/templates/offline_template_full.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -69,8 +69,8 @@ settings:
6969
min_required_connections: -1 # Min connections to initialize (-1=auto, 0=disabled)
7070
worker_gc_mode: relaxed # Worker GC strategy | options: disabled, relaxed, system
7171
endpoint_config:
72-
endpoints: # Endpoint URL(s)
73-
- 'http://localhost:8000'
72+
endpoints: # Endpoint URL(s). Must include scheme, e.g. 'http://host:port'.
73+
- '<ENDPOINT_URL eg: http://localhost:8000>'
7474
api_key: null # API key
7575
api_type: openai # API type: openai or sglang | options: openai, sglang
7676
report_dir: null # Report output directory

src/inference_endpoint/config/templates/online_template.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -17,5 +17,5 @@ settings:
1717
type: poisson # Load pattern type | options: max_throughput, poisson, concurrency, burst, step
1818
target_qps: 10.0 # Target QPS
1919
endpoint_config:
20-
endpoints: # Endpoint URL(s)
21-
- 'http://localhost:8000'
20+
endpoints: # Endpoint URL(s). Must include scheme, e.g. 'http://host:port'.
21+
- '<ENDPOINT_URL eg: http://localhost:8000>'

src/inference_endpoint/config/templates/online_template_full.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -69,8 +69,8 @@ settings:
6969
min_required_connections: -1 # Min connections to initialize (-1=auto, 0=disabled)
7070
worker_gc_mode: relaxed # Worker GC strategy | options: disabled, relaxed, system
7171
endpoint_config:
72-
endpoints: # Endpoint URL(s)
73-
- 'http://localhost:8000'
72+
endpoints: # Endpoint URL(s). Must include scheme, e.g. 'http://host:port'.
73+
- '<ENDPOINT_URL eg: http://localhost:8000>'
7474
api_key: null # API key
7575
api_type: openai # API type: openai or sglang | options: openai, sglang
7676
report_dir: null # Report output directory

0 commit comments

Comments
 (0)