Skip to content

Commit 6785a65

Browse files
authored
fix: update model ids and documentation links for switch (#997)
Signed-off-by: Jake LoRocco <jake.lorocco@ibm.com>
1 parent c57aba1 commit 6785a65

10 files changed

Lines changed: 37 additions & 27 deletions

File tree

docs/docs/advanced/intrinsics.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -36,12 +36,12 @@ Or, with a Granite Switch model via the OpenAI backend:
3636

3737
```python
3838
from mellea.backends.openai import OpenAIBackend
39-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
39+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
4040
from mellea.formatters import TemplateFormatter
4141

4242
backend = OpenAIBackend(
43-
model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name,
44-
formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name),
43+
model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name,
44+
formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name),
4545
base_url="http://localhost:8000/v1", # vLLM server
4646
api_key="EMPTY",
4747
load_embedded_adapters=True,

docs/docs/integrations/openai.md

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -278,12 +278,12 @@ Then create a backend with `load_embedded_adapters=True`:
278278

279279
```python
280280
from mellea.backends.openai import OpenAIBackend
281-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
281+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
282282
from mellea.formatters import TemplateFormatter
283283

284284
backend = OpenAIBackend(
285-
model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name,
286-
formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name),
285+
model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name,
286+
formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name),
287287
base_url="http://localhost:8000/v1",
288288
api_key="EMPTY",
289289
load_embedded_adapters=True,
@@ -305,20 +305,20 @@ For more control, load adapters manually with `load_embedded_adapters=False`:
305305
```python
306306
from mellea.backends.adapters.adapter import EmbeddedIntrinsicAdapter
307307
from mellea.backends.openai import OpenAIBackend
308-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
308+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
309309
from mellea.formatters import TemplateFormatter
310310

311311
backend = OpenAIBackend(
312-
model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name,
313-
formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name),
312+
model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name,
313+
formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name),
314314
base_url="http://localhost:8000/v1",
315315
api_key="EMPTY",
316316
load_embedded_adapters=False,
317317
)
318318

319319
# Load a single adapter from the model's HuggingFace repo
320320
adapters = EmbeddedIntrinsicAdapter.from_hub(
321-
IBM_GRANITE_SWITCH_4_1_3B.hf_model_name,
321+
IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name,
322322
intrinsic_name="answerability",
323323
)
324324
for adapter in adapters:
@@ -347,4 +347,4 @@ local servers, list available models from the server's API or UI.
347347

348348
**See also:** [Backends and Configuration](../how-to/backends-and-configuration) |
349349
[Enforce Structured Output](../how-to/enforce-structured-output) |
350-
[Official Granite Switch Documentation](GRANITE_SWITCH_DOCS)
350+
[Official Granite Switch Documentation](https://github.com/generative-computing/granite-switch)

docs/docs/reference/glossary.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -295,7 +295,7 @@ See: [Making Agents Reliable](../tutorials/04-making-agents-reliable)
295295

296296
A Granite model variant with LoRA and aLoRA adapters pre-baked into the model weights. When served via vLLM and accessed through `OpenAIBackend` with `load_embedded_adapters=True`, these embedded adapters enable [Intrinsics](../advanced/intrinsics) (RAG quality checks, requirement validation, safety evaluation) without runtime adapter loading. Only intrinsics embedded in the model are available — check the model's `adapter_index.json`.
297297

298-
See: [Official Granite Switch Documentation](GRANITE_SWITCH_DOCS) |
298+
See: [Official Granite Switch Documentation](https://github.com/generative-computing/granite-switch) |
299299
[Intrinsics](../advanced/intrinsics) |
300300
[OpenAI and OpenAI-Compatible APIs](../integrations/openai)
301301

docs/examples/granite-switch/answerability_openai.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,14 +26,14 @@
2626
print(f"Skipped: vLLM server not reachable at {VLLM_BASE_URL}", file=sys.stderr)
2727
raise SystemExit(1)
2828

29-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
29+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
3030
from mellea.backends.openai import OpenAIBackend
3131
from mellea.formatters import TemplateFormatter
3232
from mellea.stdlib.components import Document, Message
3333
from mellea.stdlib.components.intrinsic import rag
3434
from mellea.stdlib.context import ChatContext
3535

36-
SWITCH_MODEL_ID = IBM_GRANITE_SWITCH_4_1_3B.hf_model_name
36+
SWITCH_MODEL_ID = IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name
3737
assert SWITCH_MODEL_ID is not None
3838

3939
backend = OpenAIBackend(

docs/examples/granite-switch/hallucination_detection_openai.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -26,14 +26,14 @@
2626
print(f"Skipped: vLLM server not reachable at {VLLM_BASE_URL}", file=sys.stderr)
2727
raise SystemExit(1)
2828

29-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
29+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
3030
from mellea.backends.openai import OpenAIBackend
3131
from mellea.formatters import TemplateFormatter
3232
from mellea.stdlib.components import Document, Message
3333
from mellea.stdlib.components.intrinsic import rag
3434
from mellea.stdlib.context import ChatContext
3535

36-
SWITCH_MODEL_ID = IBM_GRANITE_SWITCH_4_1_3B.hf_model_name
36+
SWITCH_MODEL_ID = IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name
3737
assert SWITCH_MODEL_ID is not None
3838

3939
backend = OpenAIBackend(

docs/examples/granite-switch/manual_adapter_loading.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -36,14 +36,14 @@
3636
raise SystemExit(1)
3737

3838
from mellea.backends.adapters.adapter import EmbeddedIntrinsicAdapter
39-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
39+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
4040
from mellea.backends.openai import OpenAIBackend
4141
from mellea.formatters import TemplateFormatter
4242
from mellea.stdlib.components import Document, Message
4343
from mellea.stdlib.components.intrinsic import rag
4444
from mellea.stdlib.context import ChatContext
4545

46-
SWITCH_MODEL_ID = IBM_GRANITE_SWITCH_4_1_3B.hf_model_name
46+
SWITCH_MODEL_ID = IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name
4747
assert SWITCH_MODEL_ID is not None
4848

4949
# Create the backend WITHOUT auto-loading adapters.

docs/examples/intrinsics/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -85,7 +85,7 @@ out, new_ctx = mfuncs.act(
8585
OpenAIBackends also support a type of embedded adapter for Granite Switch models:
8686
```python
8787
backend = OpenAIBackend(
88-
model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name,
88+
model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name,
8989
load_embedded_adapters=True, # Auto-loads adapters from huggingface repo.
9090
...
9191
)

docs/examples/intrinsics/intrinsics.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -14,12 +14,12 @@
1414
# Requires the adapter for this intrinsic to be embedded in the Granite Switch
1515
# model. See docs/examples/granite-switch/ for a full runnable example.
1616
# from mellea.backends.openai import OpenAIBackend
17-
# from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
17+
# from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
1818
# from mellea.formatters import TemplateFormatter
1919
#
2020
# backend = OpenAIBackend(
21-
# model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name,
22-
# formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B.hf_model_name),
21+
# model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name,
22+
# formatter=TemplateFormatter(model_id=IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name),
2323
# base_url="http://localhost:8000/v1", # vLLM server URL
2424
# api_key="EMPTY",
2525
# load_embedded_adapters=True,

mellea/backends/model_ids.py

Lines changed: 13 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -135,10 +135,20 @@ class ModelIdentifier:
135135
)
136136

137137
# Pre-Built Granite Switch Models
138-
IBM_GRANITE_SWITCH_4_1_3B = ModelIdentifier(
139-
hf_model_name="GrizleeBer/gs-test-1" # TODO: Placeholder. Change this value.
138+
IBM_GRANITE_SWITCH_4_1_3B_PREVIEW = ModelIdentifier(
139+
hf_model_name="ibm-granite/granite-switch-4.1-3b-preview"
140140
)
141-
"""Granite Switch.""" # Document what adapters are included by default here.
141+
"""Granite Switch Preview Model. Adapters: `citations`, `query_rewrite`, `query_clarification`, `hallucination_detection`, `answerability`, `policy-guardrails`, `guardian-core`, `uncertainty`, `requirement-check`, `context-attribution`, `factuality-detection`, `factuality-correction`.""" # Document what adapters are included by default here.
142+
143+
IBM_GRANITE_SWITCH_4_1_8B_PREVIEW = ModelIdentifier(
144+
hf_model_name="ibm-granite/granite-switch-4.1-8b-preview"
145+
)
146+
"""Granite Switch Preview Model. Adapters: `citations`, `query_rewrite`, `query_clarification`, `hallucination_detection`, `answerability`, `policy-guardrails`, `guardian-core`, `uncertainty`, `requirement-check`, `context-attribution`, `factuality-detection`, `factuality-correction`.""" # Document what adapters are included by default here.
147+
148+
IBM_GRANITE_SWITCH_4_1_30B_PREVIEW = ModelIdentifier(
149+
hf_model_name="ibm-granite/granite-switch-4.1-30b-preview"
150+
)
151+
"""Granite Switch Preview Model. Adapters: `citations`, `query_rewrite`, `query_clarification`, `hallucination_detection`, `answerability`, `policy-guardrails`, `guardian-core`, `uncertainty`, `requirement-check`, `context-attribution`, `factuality-detection`, `factuality-correction`.""" # Document what adapters are included by default here.
142152

143153
#####################
144154
#### Meta models ####

test/backends/test_openai_intrinsics.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@
3737
# ---------------------------------------------------------------------------
3838
# Imports (after markers so collection-time skips fire first)
3939
# ---------------------------------------------------------------------------
40-
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B
40+
from mellea.backends.model_ids import IBM_GRANITE_SWITCH_4_1_3B_PREVIEW
4141
from mellea.backends.openai import OpenAIBackend
4242
from mellea.formatters import TemplateFormatter
4343
from mellea.stdlib import functional as mfuncs
@@ -54,7 +54,7 @@
5454
# Configuration
5555
# ---------------------------------------------------------------------------
5656
SWITCH_MODEL_ID = os.environ.get(
57-
"GRANITE_SWITCH_MODEL_ID", IBM_GRANITE_SWITCH_4_1_3B.hf_model_name
57+
"GRANITE_SWITCH_MODEL_ID", IBM_GRANITE_SWITCH_4_1_3B_PREVIEW.hf_model_name
5858
)
5959

6060

0 commit comments

Comments
 (0)