|
|
|
|
|
|
|
|
with_community = graphrag_conf.get("community", False) |
|
|
with_community = graphrag_conf.get("community", False) |
|
|
await run_graphrag(task, task_language, with_resolution, with_community, chat_model, embedding_model, progress_callback) |
|
|
await run_graphrag(task, task_language, with_resolution, with_community, chat_model, embedding_model, progress_callback) |
|
|
progress_callback(prog=1.0, msg="Knowledge Graph done ({:.2f}s)".format(timer() - start_ts)) |
|
|
progress_callback(prog=1.0, msg="Knowledge Graph done ({:.2f}s)".format(timer() - start_ts)) |
|
|
|
|
|
task_limiter = trio.CapacityLimiter(MAX_CONCURRENT_TASKS) |
|
|
return |
|
|
return |
|
|
else: |
|
|
else: |
|
|
# Standard chunking methods |
|
|
# Standard chunking methods |