Skip to content

Commit 8931a2a

Browse files
authored
fix(gallery): normalize inconsistent tag casing/plurals across gallery models (#9574)
- embeddings → embedding (6 models): aligns with the WebUI filter button defined in core/http/views/models.html ({ term: 'embedding', ... }), so models like nomic-embed-text-v1.5 now appear under the Embedding filter - TTS → tts (5 models), ASR → asr (2 models): lowercase, per existing convention used by 161+ models - CPU/Cpu → cpu (17 models), GPU → gpu (17 models): lowercase, per existing convention used by 666+ models - dedupe duplicate tag entries on 3 models that already had repeated tags (gpt-oss-20b had gguf x2; arcee-ai/AFM-4.5B had gpu x2; one Qwen model had default x2) Closes #9247
1 parent e16e758 commit 8931a2a

1 file changed

Lines changed: 21 additions & 24 deletions

File tree

gallery/index.yaml

Lines changed: 21 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -743,7 +743,6 @@
743743
- https://huggingface.co/mradermacher/Qwen3.5-27B-Claude-4.6-Opus-Reasoning-Distilled-heretic-i1-GGUF
744744
tags:
745745
- default
746-
- default
747746
overrides:
748747
parameters:
749748
model: llama-cpp/models/Qwen3.-27B-Claude-4.6-Opus-Reasoning-Distilled-heretic.i1-Q4_K_M.gguf
@@ -1915,7 +1914,7 @@
19151914
Qwen3-TTS is a high-quality text-to-speech model supporting custom voice, voice design, and voice cloning.
19161915
tags:
19171916
- text-to-speech
1918-
- TTS
1917+
- tts
19191918
license: apache-2.0
19201919
icon: https://cdn-avatars.huggingface.co/v1/production/uploads/620760a26e3b7210c2ff1943/-s1gyJfvbE1RgO5iBeNOi.png
19211920
name: "qwen3-tts-1.7b-custom-voice"
@@ -1947,7 +1946,7 @@
19471946
Fish Speech S2-Pro is a high-quality text-to-speech model supporting voice cloning via reference audio. Uses a two-stage pipeline: text to semantic tokens (LLaMA-based) then semantic to audio (DAC decoder).
19481947
tags:
19491948
- text-to-speech
1950-
- TTS
1949+
- tts
19511950
- voice-cloning
19521951
license: apache-2.0
19531952
icon: https://huggingface.co/fishaudio/s2-pro/resolve/main/overview.png
@@ -1966,7 +1965,7 @@
19661965
Qwen3-ASR is an automatic speech recognition model supporting multiple languages and batch inference.
19671966
tags:
19681967
- speech-recognition
1969-
- ASR
1968+
- asr
19701969
license: apache-2.0
19711970
icon: https://cdn-avatars.huggingface.co/v1/production/uploads/620760a26e3b7210c2ff1943/-s1gyJfvbE1RgO5iBeNOi.png
19721971
name: "qwen3-asr-1.7b"
@@ -2575,7 +2574,7 @@
25752574
license: mit
25762575
tags:
25772576
- text-to-speech
2578-
- TTS
2577+
- tts
25792578
name: "vibevoice"
25802579
urls:
25812580
- https://github.com/microsoft/VibeVoice
@@ -2609,7 +2608,7 @@
26092608
license: mit
26102609
tags:
26112610
- text-to-speech
2612-
- TTS
2611+
- tts
26132612
name: "pocket-tts"
26142613
urls:
26152614
- https://github.com/kyutai-labs/pocket-tts
@@ -3057,8 +3056,8 @@
30573056
license: apache-2.0
30583057
tags:
30593058
- gguf
3060-
- GPU
3061-
- CPU
3059+
- gpu
3060+
- cpu
30623061
- text-to-text
30633062
- jamba
30643063
- mamba
@@ -3082,8 +3081,8 @@
30823081
icon: https://cdn-avatars.huggingface.co/v1/production/uploads/639bcaa2445b133a4e942436/CEW-OjXkRkDNmTxSu8Egh.png
30833082
tags:
30843083
- gguf
3085-
- GPU
3086-
- CPU
3084+
- gpu
3085+
- cpu
30873086
- text-to-text
30883087
urls:
30893088
- https://huggingface.co/ibm-granite/granite-4.0-h-small
@@ -3145,8 +3144,8 @@
31453144
license: apache-2.0
31463145
tags:
31473146
- gguf
3148-
- GPU
3149-
- CPU
3147+
- gpu
3148+
- cpu
31503149
- text-to-text
31513150
icon: https://cdn-avatars.huggingface.co/v1/production/uploads/64f187a2cc1c03340ac30498/TYYUxK8xD1AxExFMWqbZD.png
31523151
urls:
@@ -3169,8 +3168,8 @@
31693168
license: mit
31703169
tags:
31713170
- gguf
3172-
- GPU
3173-
- CPU
3171+
- gpu
3172+
- cpu
31743173
- text-to-text
31753174
icon: https://cdn-uploads.huggingface.co/production/uploads/634262af8d8089ebaefd410e/9Bnn2AnIjfQFWBGkhDNmI.png
31763175
name: "aurore-reveil_koto-small-7b-it"
@@ -3197,8 +3196,8 @@
31973196
tags:
31983197
- multimodal
31993198
- gguf
3200-
- GPU
3201-
- Cpu
3199+
- gpu
3200+
- cpu
32023201
- image-to-text
32033202
- text-to-text
32043203
description: |
@@ -3819,7 +3818,6 @@
38193818
- gguf
38203819
- gpu
38213820
- cpu
3822-
- gguf
38233821
- openai
38243822
icon: https://raw.githubusercontent.com/openai/gpt-oss/main/docs/gpt-oss-20b.svg
38253823
urls:
@@ -4005,7 +4003,6 @@
40054003
tags:
40064004
- gguf
40074005
- gpu
4008-
- gpu
40094006
- text-generation
40104007
description: |
40114008
AFM-4.5B is a 4.5 billion parameter instruction-tuned model developed by Arcee.ai, designed for enterprise-grade performance across diverse deployment environments from cloud to edge. The base model was trained on a dataset of 8 trillion tokens, comprising 6.5 trillion tokens of general pretraining data followed by 1.5 trillion tokens of midtraining data with enhanced focus on mathematical reasoning and code generation. Following pretraining, the model underwent supervised fine-tuning on high-quality instruction datasets. The instruction-tuned model was further refined through reinforcement learning on verifiable rewards as well as for human preference. We use a modified version of TorchTitan for pretraining, Axolotl for supervised fine-tuning, and a modified version of Verifiers for reinforcement learning.
@@ -9112,7 +9109,7 @@
91129109
description: |
91139110
Granite-Embedding-107M-Multilingual is a 107M parameter dense biencoder embedding model from the Granite Embeddings suite that can be used to generate high quality text embeddings. This model produces embedding vectors of size 384 and is trained using a combination of open source relevance-pair datasets with permissive, enterprise-friendly license, and IBM collected and generated datasets. This model is developed using contrastive finetuning, knowledge distillation and model merging for improved performance.
91149111
tags:
9115-
- embeddings
9112+
- embedding
91169113
overrides:
91179114
backend: llama-cpp
91189115
embeddings: true
@@ -9130,7 +9127,7 @@
91309127
description: |
91319128
Granite-Embedding-125m-English is a 125M parameter dense biencoder embedding model from the Granite Embeddings suite that can be used to generate high quality text embeddings. This model produces embedding vectors of size 768. Compared to most other open-source models, this model was only trained using open-source relevance-pair datasets with permissive, enterprise-friendly license, plus IBM collected and generated datasets. While maintaining competitive scores on academic benchmarks such as BEIR, this model also performs well on many enterprise use cases. This model is developed using retrieval oriented pretraining, contrastive finetuning and knowledge distillation.
91329129
tags:
9133-
- embeddings
9130+
- embedding
91349131
overrides:
91359132
embeddings: true
91369133
parameters:
@@ -9147,7 +9144,7 @@
91479144
description: |
91489145
EmbeddingGemma 300M is a lightweight, high-quality embedding model from Google, based on the Gemma architecture. It produces 1024-dimensional embeddings optimized for retrieval and semantic similarity tasks. This GGUF version uses QAT (Quantization-Aware Training) Q8_0 quantization for efficient inference.
91499146
tags:
9150-
- embeddings
9147+
- embedding
91519148
overrides:
91529149
backend: llama-cpp
91539150
embeddings: true
@@ -15923,7 +15920,7 @@
1592315920
tags:
1592415921
- gpu
1592515922
- cpu
15926-
- embeddings
15923+
- embedding
1592715924
- python
1592815925
name: "all-MiniLM-L6-v2"
1592915926
url: "github:mudler/LocalAI/gallery/sentencetransformers.yaml@master"
@@ -16776,7 +16773,7 @@
1677616773
description: |
1677716774
llama3.2 embeddings model. Using as drop-in replacement for bert-embeddings
1677816775
tags:
16779-
- embeddings
16776+
- embedding
1678016777
overrides:
1678116778
embeddings: true
1678216779
parameters:
@@ -18499,7 +18496,7 @@
1849918496
description: |
1850018497
Resizable Production Embeddings with Matryoshka Representation Learning
1850118498
tags:
18502-
- embeddings
18499+
- embedding
1850318500
overrides:
1850418501
embeddings: true
1850518502
parameters:

0 commit comments

Comments
 (0)