@@ -42,12 +42,9 @@ for token, tool, tool_bool in T.handle_streaming(stream) :
4242│ ├── deepseek(api_key: str) -> openai.OpenAI
4343│ ├── openrouter(api_key: str) -> openai.OpenAI
4444│ │
45- │ ├── veniceai_request(client:openai.OpenAI, messages:list[dict], model:str, temperature:float,
46- │ │ max_tokens:int, tools: list[dict], include_venice_system_prompt:bool=False, **kwargs) -> openai.Stream
47- │ ├── generic_request(client:openai.OpenAI, messages:list[dict], model:str, temperature:float,
48- │ │ max_tokens:int, tools:list[dict], **kwargs) -> openai.Stream
49- │ └── openrouter_request(client:openai.OpenAI, messages:list[dict], model:str, temperature:float,
50- │ max_tokens:int, tools:list[dict], **kwargs) -> openai.Stream
45+ │ ├── veniceai_request(client:openai.OpenAI, messages:list[dict], model:str, temperature:float, max_tokens:int, tools: list[dict], include_venice_system_prompt:bool=False, **kwargs) -> openai.Stream
46+ │ ├── generic_request(client:openai.OpenAI, messages:list[dict], model:str, temperature:float, max_tokens:int, tools:list[dict], **kwargs) -> openai.Stream
47+ │ └── openrouter_request(client:openai.OpenAI, messages:list[dict], model:str, temperature:float, max_tokens:int, tools:list[dict], **kwargs) -> openai.Stream
5148│
5249├── handle_streaming(stream: openai.Stream) -> generator(token: str or None, tool: list, tool_bool: bool)
5350├── handle_tool_call(tool_call: dict) -> tuple[str, str, dict, str]
0 commit comments