|
18 | 18 | from graphgen.operators import ( |
19 | 19 | build_kg, |
20 | 20 | chunk_documents, |
21 | | - generate_cot, |
22 | 21 | judge_statement, |
| 22 | + partition_kg, |
23 | 23 | quiz, |
24 | 24 | read_files, |
25 | 25 | search_all, |
26 | | - traverse_graph_for_aggregated, |
27 | | - traverse_graph_for_atomic, |
28 | | - traverse_graph_for_multi_hop, |
29 | 26 | ) |
30 | 27 | from graphgen.utils import ( |
31 | 28 | async_to_sync_method, |
@@ -237,54 +234,55 @@ async def quiz_and_judge(self, quiz_and_judge_config: Dict): |
237 | 234 |
|
238 | 235 | @async_to_sync_method |
239 | 236 | async def generate(self, partition_config: Dict, generate_config: Dict): |
| 237 | + pass |
240 | 238 | # Step 1: partition the graph |
241 | | - # TODO: implement graph partitioning, e.g. Partitioner().partition(self.graph_storage) |
242 | | - mode = generate_config["mode"] |
243 | | - if mode == "atomic": |
244 | | - results = await traverse_graph_for_atomic( |
245 | | - self.synthesizer_llm_client, |
246 | | - self.tokenizer_instance, |
247 | | - self.graph_storage, |
248 | | - partition_config["method_params"], |
249 | | - self.text_chunks_storage, |
250 | | - self.progress_bar, |
251 | | - ) |
252 | | - elif mode == "multi_hop": |
253 | | - results = await traverse_graph_for_multi_hop( |
254 | | - self.synthesizer_llm_client, |
255 | | - self.tokenizer_instance, |
256 | | - self.graph_storage, |
257 | | - partition_config["method_params"], |
258 | | - self.text_chunks_storage, |
259 | | - self.progress_bar, |
260 | | - ) |
261 | | - elif mode == "aggregated": |
262 | | - results = await traverse_graph_for_aggregated( |
263 | | - self.synthesizer_llm_client, |
264 | | - self.tokenizer_instance, |
265 | | - self.graph_storage, |
266 | | - partition_config["method_params"], |
267 | | - self.text_chunks_storage, |
268 | | - self.progress_bar, |
269 | | - ) |
270 | | - elif mode == "cot": |
271 | | - results = await generate_cot( |
272 | | - self.graph_storage, |
273 | | - self.synthesizer_llm_client, |
274 | | - method_params=partition_config["method_params"], |
275 | | - ) |
276 | | - else: |
277 | | - raise ValueError(f"Unknown generation mode: {mode}") |
| 239 | + # mode = generate_config["mode"] |
| 240 | + # batches = partition_kg(self.graph_storage, partition_config) |
| 241 | + # if mode == "atomic": |
| 242 | + # results = await traverse_graph_for_atomic( |
| 243 | + # self.synthesizer_llm_client, |
| 244 | + # self.tokenizer_instance, |
| 245 | + # self.graph_storage, |
| 246 | + # partition_config["method_params"], |
| 247 | + # self.text_chunks_storage, |
| 248 | + # self.progress_bar, |
| 249 | + # ) |
| 250 | + # elif mode == "multi_hop": |
| 251 | + # results = await traverse_graph_for_multi_hop( |
| 252 | + # self.synthesizer_llm_client, |
| 253 | + # self.tokenizer_instance, |
| 254 | + # self.graph_storage, |
| 255 | + # partition_config["method_params"], |
| 256 | + # self.text_chunks_storage, |
| 257 | + # self.progress_bar, |
| 258 | + # ) |
| 259 | + # elif mode == "aggregated": |
| 260 | + # results = await traverse_graph_for_aggregated( |
| 261 | + # self.synthesizer_llm_client, |
| 262 | + # self.tokenizer_instance, |
| 263 | + # self.graph_storage, |
| 264 | + # partition_config["method_params"], |
| 265 | + # self.text_chunks_storage, |
| 266 | + # self.progress_bar, |
| 267 | + # ) |
| 268 | + # elif mode == "cot": |
| 269 | + # results = await generate_cot( |
| 270 | + # self.graph_storage, |
| 271 | + # self.synthesizer_llm_client, |
| 272 | + # method_params=partition_config["method_params"], |
| 273 | + # ) |
| 274 | + # else: |
| 275 | + # raise ValueError(f"Unknown generation mode: {mode}") |
278 | 276 | # Step 2: generate QA pairs |
279 | 277 | # TODO |
280 | 278 |
|
281 | 279 | # Step 3: format |
282 | | - results = format_generation_results( |
283 | | - results, output_data_format=generate_config["data_format"] |
284 | | - ) |
285 | | - |
286 | | - await self.qa_storage.upsert(results) |
287 | | - await self.qa_storage.index_done_callback() |
| 280 | + # results = format_generation_results( |
| 281 | + # results, output_data_format=generate_config["data_format"] |
| 282 | + # ) |
| 283 | + # |
| 284 | + # await self.qa_storage.upsert(results) |
| 285 | + # await self.qa_storage.index_done_callback() |
288 | 286 |
|
289 | 287 | @async_to_sync_method |
290 | 288 | async def clear(self): |
|
0 commit comments