Summary and motivation
I would like to propose adding an official HPC-AI integration to haystack-core-integrations.
HPC-AI provides an OpenAI-compatible chat completions API, so it fits well with Haystack’s existing pattern for provider-specific OpenAI-compatible integrations. A dedicated integration would give users a clearer setup, provider-specific docs, and tested behavior instead of relying on a generic OpenAI configuration.
This integration would mainly support users who want to use HPC-AI-hosted models in Haystack pipelines and agents, especially:
minimax/minimax-m2.5
moonshotai/kimi-k2.5
Detailed design
The integration would be added under integrations/hpc_ai and released as hpc-ai-haystack.
The initial scope would be intentionally small:
- add a single public component:
HPCAIChatGenerator
- do not add a non-chat generator in the first version
HPCAIChatGenerator would reuse Haystack’s existing OpenAIChatGenerator behavior with HPC-AI-specific defaults:
api_key=Secret.from_env_var("HPC_AI_API_KEY")
api_base_url=os.getenv("HPC_AI_BASE_URL", "https://api.hpc-ai.com/inference/v1")
model="minimax/minimax-m2.5"
The integration would declare these officially supported models:
SUPPORTED_MODELS = ["minimax/minimax-m2.5", "moonshotai/kimi-k2.5"]
Because the API is OpenAI-compatible, the integration would reuse Haystack’s existing support for:
- chat completions
- streaming callbacks
- tool calling
- structured outputs
- sync and async usage
If accepted, the implementation would also include the usual repo wiring for a new integration:
- tests and CI workflow
- integration README and examples
- labeler update
- README inventory update
- companion docs in
haystack
- integration page in
haystack-integrations
To keep the scope easy to review and maintain, the first version would be chat-only and limited to the two models above.
Checklist
If the request is accepted, ensure the following checklist is complete before closing this issue.
Tasks
Summary and motivation
I would like to propose adding an official
HPC-AIintegration tohaystack-core-integrations.HPC-AI provides an OpenAI-compatible chat completions API, so it fits well with Haystack’s existing pattern for provider-specific OpenAI-compatible integrations. A dedicated integration would give users a clearer setup, provider-specific docs, and tested behavior instead of relying on a generic OpenAI configuration.
This integration would mainly support users who want to use HPC-AI-hosted models in Haystack pipelines and agents, especially:
minimax/minimax-m2.5moonshotai/kimi-k2.5Detailed design
The integration would be added under
integrations/hpc_aiand released ashpc-ai-haystack.The initial scope would be intentionally small:
HPCAIChatGeneratorHPCAIChatGeneratorwould reuse Haystack’s existingOpenAIChatGeneratorbehavior with HPC-AI-specific defaults:api_key=Secret.from_env_var("HPC_AI_API_KEY")api_base_url=os.getenv("HPC_AI_BASE_URL", "https://api.hpc-ai.com/inference/v1")model="minimax/minimax-m2.5"The integration would declare these officially supported models:
SUPPORTED_MODELS = ["minimax/minimax-m2.5", "moonshotai/kimi-k2.5"]Because the API is OpenAI-compatible, the integration would reuse Haystack’s existing support for:
If accepted, the implementation would also include the usual repo wiring for a new integration:
haystackhaystack-integrationsTo keep the scope easy to review and maintain, the first version would be chat-only and limited to the two models above.
Checklist
If the request is accepted, ensure the following checklist is complete before closing this issue.
Tasks
mainbranchintegration:hpc-aihas been added to the list of labels for this repository