Skip to content

Commit 8dc90d6

Browse files
authored
docs: remove DynamicChatPromptBuilder from code snippets (#10890)
1 parent 7e9c43e commit 8dc90d6

10 files changed

Lines changed: 40 additions & 40 deletions

File tree

docs-website/docs/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.18/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7171
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7272
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7373

74-
from haystack.components.builders import DynamicChatPromptBuilder
74+
from haystack.components.builders import ChatPromptBuilder
7575
from haystack.components.generators.chat import OpenAIChatGenerator
7676
from haystack.dataclasses import ChatMessage
7777
from haystack import Pipeline
@@ -81,8 +81,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8181
if __name__ == "__main__":
8282
pipe = Pipeline()
8383
pipe.add_component("tracer", LangfuseConnector("Chat example"))
84-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
85-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
84+
pipe.add_component("prompt_builder", ChatPromptBuilder())
85+
pipe.add_component("llm", OpenAIChatGenerator())
8686

8787
pipe.connect("prompt_builder.prompt", "llm.messages")
8888

@@ -97,7 +97,7 @@ if __name__ == "__main__":
9797
data={
9898
"prompt_builder": {
9999
"template_variables": {"location": "Berlin"},
100-
"prompt_source": messages,
100+
"template": messages,
101101
},
102102
},
103103
)

docs-website/versioned_docs/version-2.19/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.20/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.21/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.22/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.23/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.24/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.25/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

docs-website/versioned_docs/version-2.26/pipeline-components/connectors/langfuseconnector.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -74,7 +74,7 @@ os.environ["LANGFUSE_HOST"] = "https://cloud.langfuse.com"
7474
os.environ["TOKENIZERS_PARALLELISM"] = "false"
7575
os.environ["HAYSTACK_CONTENT_TRACING_ENABLED"] = "true"
7676

77-
from haystack.components.builders import DynamicChatPromptBuilder
77+
from haystack.components.builders import ChatPromptBuilder
7878
from haystack.components.generators.chat import OpenAIChatGenerator
7979
from haystack.dataclasses import ChatMessage
8080
from haystack import Pipeline
@@ -84,8 +84,8 @@ from haystack_integrations.components.connectors.langfuse import LangfuseConnect
8484
if __name__ == "__main__":
8585
pipe = Pipeline()
8686
pipe.add_component("tracer", LangfuseConnector("Chat example"))
87-
pipe.add_component("prompt_builder", DynamicChatPromptBuilder())
88-
pipe.add_component("llm", OpenAIChatGenerator(model="gpt-3.5-turbo"))
87+
pipe.add_component("prompt_builder", ChatPromptBuilder())
88+
pipe.add_component("llm", OpenAIChatGenerator())
8989

9090
pipe.connect("prompt_builder.prompt", "llm.messages")
9191

@@ -100,7 +100,7 @@ if __name__ == "__main__":
100100
data={
101101
"prompt_builder": {
102102
"template_variables": {"location": "Berlin"},
103-
"prompt_source": messages,
103+
"template": messages,
104104
},
105105
},
106106
)

0 commit comments

Comments
 (0)