Skip to content

Commit 0ec26f7

Browse files
authored
chore: remove auto warmup where possible (#409)
1 parent 35b128f commit 0ec26f7

10 files changed

Lines changed: 0 additions & 31 deletions

integrations/astradb.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -94,7 +94,6 @@ documents = [Document(content="There are over 7,000 languages spoken around the
9494
Document(content="In certain parts of the world, like the Maldives, Puerto Rico, and San Diego, you can witness the phenomenon of bioluminescent waves.")]
9595

9696
document_embedder = SentenceTransformersDocumentEmbedder(model=model)
97-
document_embedder.warm_up()
9897
documents_with_embeddings = document_embedder.run(documents)
9998

10099
document_store.write_documents(documents_with_embeddings.get("documents"))

integrations/couchbase-document-store.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -212,7 +212,6 @@ from haystack.components.embedders import SentenceTransformersDocumentEmbedder
212212
documents = [Document(content="Alice has been living in New York City for the past 5 years.")]
213213

214214
document_embedder = SentenceTransformersDocumentEmbedder(model="sentence-transformers/all-MiniLM-L6-v2")
215-
document_embedder.warm_up() # will download the model during first run
216215
documents_with_embeddings = document_embedder.run(documents)
217216

218217
document_store.write_documents(documents_with_embeddings.get("documents"))
@@ -342,7 +341,6 @@ documents = [
342341
model_name = "sentence-transformers/all-MiniLM-L6-v2"
343342

344343
document_embedder = SentenceTransformersDocumentEmbedder(model=model_name)
345-
document_embedder.warm_up()
346344
documents_with_embeddings = document_embedder.run(documents)
347345

348346
document_store.write_documents(documents_with_embeddings.get("documents"))
@@ -525,7 +523,6 @@ documents = [
525523

526524
model_name = "sentence-transformers/all-MiniLM-L6-v2"
527525
document_embedder = SentenceTransformersDocumentEmbedder(model=model_name)
528-
document_embedder.warm_up()
529526
documents_with_embeddings = document_embedder.run(documents)
530527

531528
document_store.write_documents(documents_with_embeddings.get("documents"))

integrations/hanlp.md

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -69,9 +69,6 @@ splitter = ChineseDocumentSplitter(
6969
respect_sentence_boundary=True
7070
)
7171

72-
# Warm up the component (loads the necessary models)
73-
splitter.warm_up()
74-
7572
result = splitter.run(documents=[doc])
7673
print(result["documents"])
7774
```
@@ -126,7 +123,6 @@ splitter = ChineseDocumentSplitter(
126123
split_overlap=3,
127124
respect_sentence_boundary=True
128125
)
129-
splitter.warm_up()
130126
result = splitter.run(documents=[doc])
131127
```
132128

@@ -147,7 +143,6 @@ splitter = ChineseDocumentSplitter(
147143
split_by="function",
148144
splitting_function=custom_chinese_split
149145
)
150-
splitter.warm_up()
151146
result = splitter.run(documents=[doc])
152147
```
153148

integrations/instructor-embedder.md

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -89,7 +89,6 @@ text_embedder = InstructorTextEmbedder(
8989
model="hkunlp/instructor-base", instruction=instruction,
9090
device=ComponentDevice.from_str("cpu"),
9191
)
92-
text_embedder.warm_up()
9392
result = text_embedder.run(text)
9493
print(f"Embedding: {result['embedding']}")
9594
print(f"Embedding Dimension: {len(result['embedding'])}")
@@ -112,8 +111,6 @@ doc_embedder = InstructorDocumentEmbedder(
112111
device=ComponentDevice.from_str("cpu"),
113112
)
114113

115-
doc_embedder.warm_up()
116-
117114
# Text taken from PubMed QA Dataset (https://huggingface.co/datasets/pubmed_qa)
118115
document_list = [
119116
Document(
@@ -174,8 +171,6 @@ doc_embedder = InstructorDocumentEmbedder(
174171
batch_size=32,
175172
device=ComponentDevice.from_str("cpu"),
176173
)
177-
# Warm up the embedder (loading the pre-trained model)
178-
doc_embedder.warm_up()
179174

180175
# Create an indexing pipeline
181176
indexing_pipeline = Pipeline()
@@ -218,8 +213,6 @@ text_embedder = InstructorTextEmbedder(
218213
instruction=query_embedding_instruction,
219214
device=ComponentDevice.from_str("cpu"),
220215
)
221-
# Load the text embedding model
222-
text_embedder.warm_up()
223216

224217
# Create a query pipeline
225218
query_pipeline = Pipeline()

integrations/langfuse.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -116,7 +116,6 @@ def get_pipeline(document_store: InMemoryDocumentStore):
116116
document_store = InMemoryDocumentStore()
117117
dataset = load_dataset("bilgeyucel/seven-wonders", split="train")
118118
embedder = SentenceTransformersDocumentEmbedder("sentence-transformers/all-MiniLM-L6-v2")
119-
embedder.warm_up()
120119
docs_with_embeddings = embedder.run([Document(**ds) for ds in dataset]).get("documents") or [] # type: ignore
121120
document_store.write_documents(docs_with_embeddings)
122121

integrations/llama_cpp.md

Lines changed: 0 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -105,7 +105,6 @@ generator = LlamaCppGenerator(
105105
model_kwargs={"n_gpu_layers": -1},
106106
generation_kwargs={"max_tokens": 128, "temperature": 0.1},
107107
)
108-
generator.warm_up()
109108
prompt = f"Who is the best American actor?"
110109
result = generator.run(prompt)
111110
```
@@ -129,7 +128,6 @@ generator = LlamaCppGenerator(
129128
n_batch=128,
130129
model_kwargs={"n_gpu_layers": -1}
131130
)
132-
generator.warm_up()
133131
prompt = f"Who is the best American actor?"
134132
result = generator.run(prompt, generation_kwargs={"max_tokens": 128})
135133
generated_text = result["replies"][0]
@@ -153,7 +151,6 @@ generator = LlamaCppGenerator(
153151
n_batch=128,
154152
generation_kwargs={"max_tokens": 128, "temperature": 0.1},
155153
)
156-
generator.warm_up()
157154
prompt = f"Who is the best American actor?"
158155
result = generator.run(prompt)
159156
```
@@ -168,7 +165,6 @@ generator = LlamaCppGenerator(
168165
n_ctx=512,
169166
n_batch=128,
170167
)
171-
generator.warm_up()
172168
prompt = f"Who is the best American actor?"
173169
result = generator.run(
174170
prompt,

integrations/neo4j-document-store.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -153,7 +153,6 @@ documents = [
153153
Document(content="My name is Morgan and I live in Paris.", meta={"release_date": "2018-12-09"})]
154154

155155
document_embedder = SentenceTransformersDocumentEmbedder(model="sentence-transformers/all-MiniLM-L6-v2")
156-
document_embedder.warm_up()
157156
documents_with_embeddings = document_embedder.run(documents)
158157

159158
document_store.write_documents(documents_with_embeddings.get("documents"))

integrations/nvidia.md

Lines changed: 0 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,6 @@ from haystack_integrations.components.embedders.nvidia import NvidiaTextEmbedder
8686
text_to_embed = "I love pizza!"
8787

8888
text_embedder = NvidiaTextEmbedder(model="nvidia/llama-3.2-nv-embedqa-1b-v2")
89-
text_embedder.warm_up()
9089

9190
print(text_embedder.run(text_to_embed))
9291
# {'embedding': [-0.02264290489256382, -0.03457780182361603, ...}
@@ -105,7 +104,6 @@ documents = [
105104
]
106105

107106
document_embedder = NvidiaDocumentEmbedder(model="nvidia/llama-3.2-nv-embedqa-1b-v2")
108-
document_embedder.warm_up()
109107
document_embedder.run(documents=documents)
110108
# {'documents': [Document(id=..., content: 'Pizza is made with dough and cheese', embedding: vector of size 2048), ...], 'meta': {'usage': {'prompt_tokens': 36, 'total_tokens': 36}}}
111109
```
@@ -123,7 +121,6 @@ generator = NvidiaGenerator(
123121
"max_tokens": 1024,
124122
},
125123
)
126-
generator.warm_up()
127124

128125
result = generator.run(prompt="When was the Golden Gate Bridge built?")
129126
print(result["replies"])
@@ -160,7 +157,6 @@ from haystack_integrations.components.rankers.nvidia import NvidiaRanker
160157
ranker = NvidiaRanker(
161158
api_key=Secret.from_env_var("NVIDIA_API_KEY"),
162159
)
163-
ranker.warm_up()
164160

165161
query = "What is the capital of Germany?"
166162
documents = [
@@ -248,7 +244,6 @@ text_embedder = NvidiaTextEmbedder(model="nvidia/llama-3.2-nv-embedqa-1b-v2")
248244
retriever = InMemoryEmbeddingRetriever(document_store=document_store)
249245
prompt_builder = PromptBuilder(template=prompt)
250246
generator = NvidiaGenerator(model="meta/llama-3.1-70b-instruct")
251-
generator.warm_up()
252247

253248
rag_pipeline = Pipeline()
254249

integrations/opensearch-document-store.md

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -98,7 +98,6 @@ docs = [
9898

9999
# Embed the documents and add them to the document store
100100
doc_embedder = SentenceTransformersDocumentEmbedder(model="sentence-transformers/all-MiniLM-L6-v2")
101-
doc_embedder.warm_up()
102101
docs = doc_embedder.run(docs)
103102

104103
# Write the documents to the OpenSearch document store

integrations/snowflake.md

Lines changed: 0 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -100,8 +100,6 @@ The `SnowflakeTableRetriever` supports three authentication methods:
100100

101101
Ensure you have `select` access to the tables before querying the database. More details [here](https://docs.snowflake.com/en/user-guide/security-access-control-privileges):
102102
```python
103-
# Warm up the component so it connects to the database
104-
executor.warm_up()
105103

106104
# Run the retriever
107105
response = executor.run(query="""select * from database_name.schema_name.table_name""")
@@ -113,7 +111,6 @@ executor = SnowflakeTableRetriever(
113111
schema_name="<SCHEMA-NAME>",
114112
database ="<DB-NAME>"
115113
)
116-
executor.warm_up()
117114

118115
response = executor.run(query="""select * from table_name""")
119116
```

0 commit comments

Comments
 (0)