@@ -177,29 +177,20 @@ def __init__(self, llm_model, prompt_template_file, inference_type, vllm_endpoin
177177 self .vllm_endpoint = vllm_endpoint
178178
179179 def init_prompt (self , model_path , prompt_content = None , prompt_template_file = None , enable_think = False ):
180- # using the prompt template enhancement strategy(only tested on Qwen2-7B-Instruction) if template_enhance_on is true
181- template_enhance_on = True if "Qwen2" in self .model_id else False
182180 if prompt_content :
183181 return get_prompt_template (model_path , prompt_content , prompt_template_file , enable_think )
184182 elif prompt_template_file is None :
185183 print ("There is no template file, using the default template." )
186184 prompt_template = get_prompt_template (model_path , prompt_content , prompt_template_file , enable_think )
187- return (
188- DocumentedContextRagPromptTemplate .from_template (prompt_template )
189- if template_enhance_on
190- else prompt_template
191- )
185+ return prompt_template
192186 else :
193187 safe_root = "/templates"
194188 prompt_template_file = os .path .normpath (os .path .join (safe_root , prompt_template_file ))
195189 if not prompt_template_file .startswith (safe_root ):
196190 raise ValueError ("Invalid template path" )
197191 if not os .path .exists (prompt_template_file ):
198192 raise ValueError ("Template file not exists" )
199- if template_enhance_on :
200- return DocumentedContextRagPromptTemplate .from_file (prompt_template_file )
201- else :
202- return get_prompt_template (model_path , prompt_content , prompt_template_file , enable_think )
193+ return get_prompt_template (model_path , prompt_content , prompt_template_file , enable_think )
203194
204195 def set_prompt (self , prompt ):
205196 if "{context}" not in prompt :
0 commit comments