RAGFlow.create_dataset(
    name: str,
    avatar: str = "",
    description: str = "",
    language: str = "English",
    permission: str = "me", 
    document_count: int = 0,
    chunk_count: int = 0,
    parse_method: str = "naive",
    parser_config: DataSet.ParserConfig = None
) -> DataSet
Creates a knowledge base (dataset).
str, RequiredThe unique name of the dataset to create. It must adhere to the following requirements:
strBase64 encoding of the avatar. Defaults to ""
strThe id of the tenant associated with the created dataset is used to identify different users. Defaults to None.
strThe description of the created dataset. Defaults to "".
strThe language setting of the created dataset. Defaults to "English". ????????????
Specify who can operate on the dataset. Defaults to "me".
intThe number of documents associated with the dataset. Defaults to 0.
intThe number of data chunks generated or processed by the created dataset. Defaults to 0.
strThe method used by the dataset to parse and process data. Defaults to "naive".
The parser configuration of the dataset. A ParserConfig object contains the following attributes:
chunk_token_count: Defaults to 128.layout_recognize: Defaults to True.delimiter: Defaults to '\n!?。;!?'.task_page_size: Defaults to 12.dataset object.Exceptionfrom ragflow import RAGFlow
rag_object = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag_object.create_dataset(name="kb_1")
RAGFlow.delete_datasets(ids: list[str] = None)
Deletes knowledge bases by name or ID.
The IDs of the knowledge bases to delete.
Exceptionrag.delete_datasets(ids=["id_1","id_2"])
RAGFlow.list_datasets(
    page: int = 1, 
    page_size: int = 1024, 
    orderby: str = "create_time", 
    desc: bool = True,
    id: str = None,
    name: str = None
) -> list[DataSet]
Retrieves a list of knowledge bases.
intThe current page number to retrieve from the paginated results. Defaults to 1.
intThe number of records on each page. Defaults to 1024.
strThe field by which the records should be sorted. This specifies the attribute or column used to order the results. Defaults to "create_time".
boolWhether the sorting should be in descending order. Defaults to True.
strThe id of the dataset to be got. Defaults to None.
strThe name of the dataset to be got. Defaults to None.
DataSet objects representing the retrieved knowledge bases.Exception.for ds in rag_object.list_datasets():
    print(ds)
dataset = rag_object.list_datasets(id = "id_1")
print(dataset[0])
DataSet.update(update_message: dict)
Updates the current knowledge base.
dict[str, str|int], Required"name": str The name of the knowledge base to update."tenant_id": str The "tenant_id you get after calling create_dataset()."embedding_model": str The embedding model for generating vector embeddings.
"chunk_count" is 0 before updating "embedding_model"."parser_method": str
"naive": General"manual: Manual"qa": Q&A"table": Table"paper": Paper"book": Book"laws": Laws"presentation": Presentation"picture": Picture"one":One"knowledge_graph": Knowledge Graph"email": EmailExceptionfrom ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets(name="kb_1")
ds.update({"embedding_model":"BAAI/bge-zh-v1.5", "parse_method":"manual"})
:::tip API GROUPING File management inside knowledge base :::
DataSet.upload_documents(document_list: list[dict])
list[dict]A list composed of dicts containing name and blob.
no return
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.create_dataset(name="kb_1")
ds.upload_documents([{name="1.txt", blob="123"}, ...] }
Document.update(update_message:dict)
dictonly name,parser_config,parser_method can be changed
no return
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds=rag.list_datasets(id='id')
ds=ds[0]
doc = ds.list_documents(id="wdfxb5t547d")
doc = doc[0]
doc.update([{"parser_method": "manual"...}])
Document.download() -> bytes
bytes of the document.
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds=rag.list_datasets(id="id")
ds=ds[0]
doc = ds.list_documents(id="wdfxb5t547d")
doc = doc[0]
open("~/ragflow.txt", "wb+").write(doc.download())
print(doc)
Dataset.list_documents(id:str =None, keywords: str=None, offset: int=0, limit:int = 1024,order_by:str = "create_time", desc: bool = True) -> list[Document]
strThe id of the document to be got
strList documents whose name has the given keywords. Defaults to None.
intThe beginning number of records for paging. Defaults to 0.
intRecords number to return, -1 means all of them. Records number to return, -1 means all of them.
strThe field by which the records should be sorted. This specifies the attribute or column used to order the results.
boolA boolean flag indicating whether the sorting should be in descending order.
list[Document]
A document object containing the following attributes:
strId of the retrieved document. Defaults to "".
strThumbnail image of the retrieved document. Defaults to "".
strKnowledge base ID related to the document. Defaults to "".
strMethod used to parse the document. Defaults to "".
ParserConfigConfiguration object for the parser. Defaults to None.
strSource type of the document. Defaults to "".
strType or category of the document. Defaults to "".
strCreator of the document. Defaults to "".
strstring
“
Name or title of the document. Defaults to "".
intSize of the document in bytes or some other unit. Defaults to 0.
intNumber of tokens in the document. Defaults to "".
intNumber of chunks the document is split into. Defaults to 0.
floatCurrent processing progress as a percentage. Defaults to 0.0.
strMessage indicating current progress status. Defaults to "".
datetimeStart time of the document processing. Defaults to None.
floatDuration of the processing in seconds or minutes. Defaults to 0.0.
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.create_dataset(name="kb_1")
filename1 = "~/ragflow.txt"
blob=open(filename1 , "rb").read()
list_files=[{"name":filename1,"blob":blob}]
ds.upload_documents(list_files)
for d in ds.list_documents(keywords="rag", offset=0, limit=12):
    print(d)
DataSet.delete_documents(ids: list[str] = None)
no return
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets(name="kb_1")
ds = ds[0]
ds.delete_documents(ids=["id_1","id_2"])
DataSet.async_parse_documents(document_ids:list[str]) -> None
DataSet.async_cancel_parse_documents(document_ids:list[str])-> None
list[str]The ids of the documents to be parsed ????????????????????????????????????????????????????
no return ????????????????????????????????????????????????????
#documents parse and cancel
rag = RAGFlow(API_KEY, HOST_ADDRESS)
ds = rag.create_dataset(name="God5")
documents = [
    {'name': 'test1.txt', 'blob': open('./test_data/test1.txt',"rb").read()},
    {'name': 'test2.txt', 'blob': open('./test_data/test2.txt',"rb").read()},
    {'name': 'test3.txt', 'blob': open('./test_data/test3.txt',"rb").read()}
]
ds.upload_documents(documents)
documents=ds.list_documents(keywords="test")
ids=[]
for document in documents:
    ids.append(document.id)
ds.async_parse_documents(ids)
print("Async bulk parsing initiated")
ds.async_cancel_parse_documents(ids)
print("Async bulk parsing cancelled")
Document.list_chunks(keywords: str = None, offset: int = 0, limit: int = -1, id : str = None) -> list[Chunk]
keywords: str
List chunks whose name has the given keywords
default: None
offset: int
The beginning number of records for paging
default: 1
limit: int
Records number to return
default: 30
id: str
The ID of the chunk to be retrieved
default: None
list[chunk]
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets("123")
ds = ds[0]
ds.async_parse_documents(["wdfxb5t547d"])
for c in doc.list_chunks(keywords="rag", offset=0, limit=12):
    print(c)
Document.add_chunk(content:str) -> Chunk
str, RequiredContains the main text or information of the chunk.
list[str]list the key terms or phrases that are significant or central to the chunk’s content.
chunk
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets(id="123")
ds = ds[0]
doc = ds.list_documents(id="wdfxb5t547d")
doc = doc[0]
chunk = doc.add_chunk(content="xxxxxxx")
Document.delete_chunks(chunk_ids: list[str])
list[str]The list of chunk_id
no return
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets(id="123")
ds = ds[0]
doc = ds.list_documents(id="wdfxb5t547d")
doc = doc[0]
chunk = doc.add_chunk(content="xxxxxxx")
doc.delete_chunks(["id_1","id_2"])
Chunk.update(update_message: dict)
content: str
Contains the main text or information of the chunk
important_keywords: list[str]
List the key terms or phrases that are significant or central to the chunk’s content
available: int
Indicating the availability status, 0 means unavailable and 1 means available
no return
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets(id="123")
ds = ds[0]
doc = ds.list_documents(id="wdfxb5t547d")
doc = doc[0]
chunk = doc.add_chunk(content="xxxxxxx")
chunk.update({"content":"sdfx...})
RAGFlow.retrieve(question:str="", datasets:list[str]=None, document=list[str]=None, offset:int=1, limit:int=30, similarity_threshold:float=0.2, vector_similarity_weight:float=0.3, top_k:int=1024,rerank_id:str=None,keyword:bool=False,higlight:bool=False) -> list[Chunk]
str, RequiredThe user query or query keywords. Defaults to "".
list[Dataset], RequiredThe scope of datasets.
list[Document]The scope of document. None means no limitation. Defaults to None.
intThe beginning point of retrieved records. Defaults to 0.
intThe maximum number of records needed to return. Defaults to 6.
floatThe minimum similarity score. Defaults to 0.2.
floatThe weight of vector cosine similarity, 1 - x is the term similarity weight. Defaults to 0.3.
intNumber of records engaged in vector cosine computaton. Defaults to 1024.
strID of the rerank model.  Defaults to None.
boolIndicating whether keyword-based matching is enabled (True) or disabled (False).
boolSpecifying whether to enable highlighting of matched terms in the results (True) or not (False).
list[Chunk]
from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
ds = rag.list_datasets(name="ragflow")
ds = ds[0]
name = 'ragflow_test.txt'
path = './test_data/ragflow_test.txt'
rag.create_document(ds, name=name, blob=open(path, "rb").read())
doc = ds.list_documents(name=name)
doc = doc[0]
ds.async_parse_documents([doc.id])
for c in rag.retrieve(question="What's ragflow?", 
             datasets=[ds.id], documents=[doc.id], 
             offset=1, limit=30, similarity_threshold=0.2, 
             vector_similarity_weight=0.3,
             top_k=1024
             ):
    print(c)
:::tip API GROUPING Chat APIs :::
RAGFlow.create_chat(
    name: str = "assistant", 
    avatar: str = "path", 
    knowledgebases: list[DataSet] = [], 
    llm: Chat.LLM = None, 
    prompt: Chat.Prompt = None
) -> Chat
Creates a chat assistant.
Chat object representing the chat assistant.ExceptionstrThe name of the chat assistant. Defaults to "assistant".
strBase64 encoding of the avatar. Defaults to "".
list[str]The associated knowledge bases. Defaults to ["kb1"].
LLMThe llm of the created chat. Defaults to None. When the value is None, a dictionary with the following values will be generated as the default.
strNone, the user’s default chat model will be returned.float0.1.float0.3float0.2.float0.7.int512.strInstructions for the LLM to follow.
"similarity_threshold": float A similarity score to evaluate distance between two lines of text. It’s weighted keywords similarity and vector cosine similarity. If the similarity between query and chunk is less than this threshold, the chunk will be filtered out. Defaults to 0.2."keywords_similarity_weight": float It’s weighted keywords similarity and vector cosine similarity or rerank score (0~1). Defaults to 0.7."top_n": int Not all the chunks whose similarity score is above the ‘similarity threshold’ will be feed to LLMs. LLM can only see these ‘Top N’ chunks. Defaults to 8."variables": list[dict[]] If you use dialog APIs, the variables might help you chat with your clients with different strategies. The variables are used to fill in the ‘System’ part in prompt in order to give LLM a hint. The ‘knowledge’ is a very special variable which will be filled-in with the retrieved chunks. All the variables in ‘System’ should be curly bracketed. Defaults to [{"key": "knowledge", "optional": True}]"rerank_model": str If it is not specified, vector cosine similarity will be used; otherwise, reranking score will be used. Defaults to ""."empty_response": str If nothing is retrieved in the knowledge base for the user’s question, this will be used as the response. To allow the LLM to improvise when nothing is retrieved, leave this blank. Defaults to None."opener": str The opening greeting for the user. Defaults to "Hi! I am your assistant, can I help you?"."show_quote: bool Indicates whether the source of text should be displayed Defaults to True."prompt": str The prompt content. Defaults to You are an intelligent assistant. Please summarize the content of the knowledge base to answer the question. Please list the data in the knowledge base and answer in detail. When all knowledge base content is irrelevant to the question, your answer must include the sentence "The answer you are looking for is not found in the knowledge base!" Answers need to consider chat history.
  Here is the knowledge base:
  {knowledge}
  The above is the knowledge base..from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
knowledge_base = rag.list_datasets(name="kb_1")
assistant = rag.create_chat("Miss R", knowledgebases=knowledge_base)
Chat.update(update_message: dict)
Updates the current chat assistant.
dict[str, Any], Required"name": str The name of the chat assistant to update."avatar": str Base64 encoding of the avatar. Defaults to """knowledgebases": list[str] Knowledge bases to update."llm": dict The LLM settings:
"model_name", str The chat model name."temperature", float Controls the randomness of the model’s predictions."top_p", float Also known as “nucleus sampling”, this parameter sets a threshold to select a smaller set of words to sample from."presence_penalty", float This discourages the model from repeating the same information by penalizing words that have appeared in the conversation."frequency penalty", float Similar to presence penalty, this reduces the model’s tendency to repeat the same words."max_token", int This sets the maximum length of the model’s output, measured in the number of tokens (words or pieces of words)."prompt" : Instructions for the LLM to follow.
"similarity_threshold": float A score to evaluate distance between two lines of text. It’s weighted keywords similarity and vector cosine similarity. If the similarity between query and chunk is less than this threshold, the chunk will be filtered out. Defaults to 0.2."keywords_similarity_weight": float It’s weighted keywords similarity and vector cosine similarity or rerank score (0~1). Defaults to 0.7."top_n": int Not all the chunks whose similarity score is above the ‘similarity threshold’ will be feed to LLMs. LLM can only see these ‘Top N’ chunks. Defaults to 8."variables": list[dict[]] If you use dialog APIs, the variables might help you chat with your clients with different strategies. The variables are used to fill in the ‘System’ part in prompt in order to give LLM a hint. The ‘knowledge’ is a very special variable which will be filled-in with the retrieved chunks. All the variables in ‘System’ should be curly bracketed. Defaults to [{"key": "knowledge", "optional": True}]"rerank_model": str If it is not specified, vector cosine similarity will be used; otherwise, reranking score will be used. Defaults to ""."empty_response": str If nothing is retrieved in the knowledge base for the user’s question, this will be used as the response. To allow the LLM to improvise when nothing is retrieved, leave this blank. Defaults to None."opener": str The opening greeting for the user. Defaults to "Hi! I am your assistant, can I help you?"."show_quote: bool Indicates whether the source of text should be displayed Defaults to True."prompt": str The prompt content. Defaults to You are an intelligent assistant. Please summarize the content of the knowledge base to answer the question. Please list the data in the knowledge base and answer in detail. When all knowledge base content is irrelevant to the question, your answer must include the sentence "The answer you are looking for is not found in the knowledge base!" Answers need to consider chat history.
Here is the knowledge base:
{knowledge}
The above is the knowledge base..Exceptionfrom ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
knowledge_base = rag.list_datasets(name="kb_1")
assistant = rag.create_chat("Miss R", knowledgebases=knowledge_base)
assistant.update({"name": "Stefan", "llm": {"temperature": 0.8}, "prompt": {"top_n": 8}})
Deletes specified chat assistants.
RAGFlow.delete_chats(ids: list[str] = None)
IDs of the chat assistants to delete. If not specified, all chat assistants will be deleted.
Exceptionfrom ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
rag.delete_chats(ids=["id_1","id_2"])
RAGFlow.list_chats(
    page: int = 1, 
    page_size: int = 1024, 
    orderby: str = "create_time", 
    desc: bool = True,
    id: str = None,
    name: str = None
) -> list[Chat]
Specifies the page on which the records will be displayed. Defaults to 1.
The number of records on each page. Defaults to 1024.
The attribute by which the results are sorted. Defaults to "create_time".
Indicates whether to sort the results in descending order. Defaults to True.
stringThe ID of the chat to retrieve. Defaults to None.
stringThe name of the chat to retrieve. Defaults to None.
Chat objects.Exception.from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
for assistant in rag.list_chats():
    print(assistant)
:::tip API GROUPING Chat-session APIs :::
Chat.create_session(name: str = "New session") -> Session
Creates a chat session.
The name of the chat session to create.
Session object containing the following attributes:
id: str The auto-generated unique identifier of the created session.name: str The name of the created session.message: list[Message] The messages of the created session assistant. Default: [{"role": "assistant", "content": "Hi! I am your assistant,can I help you?"}]chat_id: str The ID of the associated chat assistant.Exceptionfrom ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
assistant = rag.list_chats(name="Miss R")
assistant = assistant[0]
session = assistant.create_session()
Session.update(update_message: dict)
Updates the current session.
dict[str, Any], Required"name": str The name of the session to update.Exceptionfrom ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
assistant = rag.list_chats(name="Miss R")
assistant = assistant[0]
session = assistant.create_session("session_name")
session.update({"name": "updated_name"})
Session.ask(question: str, stream: bool = False) -> Optional[Message, iter[Message]]
The question to start an AI chat. Defaults to None.
Indicates whether to output responses in a streaming way. Defaults to False.
Optional[Message, iter[Message]]
stream is set to Falsestream is set to TruestrThe ID of the message. id is automatically generated.
strThe content of the message. Defaults to "Hi! I am your assistant, can I help you?".
list[Chunk]The auto-generated reference of the message. Each chunk object includes the following attributes:
strstrstrstrstrstrfloat
A general similarity score, usually a composite score derived from various similarity measures . This score represents the degree of similarity between two objects. The value ranges between 0 and 1, where a value closer to 1 indicates higher similarity.floatfloatlist[string]from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
assistant = rag.list_chats(name="Miss R")
assistant = assistant[0]
sess = assistant.create_session()    
print("\n==================== Miss R =====================\n")
print(assistant.get_prologue())
while True:
    question = input("\n==================== User =====================\n> ")
    print("\n==================== Miss R =====================\n")
    
    cont = ""
    for ans in sess.ask(question, stream=True):
        print(ans.content[len(cont):], end='', flush=True)
        cont = ans.content
Chat.list_sessions(
    page: int = 1, 
    page_size: int = 1024, 
    orderby: str = "create_time", 
    desc: bool = True,
    id: str = None,
    name: str = None
) -> list[Session]
Lists sessions associated with the current chat assistant.
Specifies the page on which records will be displayed. Defaults to 1.
The number of records on each page. Defaults to 1024.
The field by which the records should be sorted. This specifies the attribute or column used to sort the results. Defaults to "create_time".
Whether the sorting should be in descending order. Defaults to True.
The ID of the chat session to retrieve. Defaults to None.
The name of the chat to retrieve. Defaults to None.
Session objects associated with the current chat assistant.Exception.from ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
assistant = rag.list_chats(name="Miss R")
assistant = assistant[0]
for session in assistant.list_sessions():
    print(session)
Chat.delete_sessions(ids:list[str] = None)
Deletes specified sessions or all sessions associated with the current chat assistant.
IDs of the sessions to delete. If not specified, all sessions associated with the current chat assistant will be deleted.
Exceptionfrom ragflow import RAGFlow
rag = RAGFlow(api_key="<YOUR_API_KEY>", base_url="http://<YOUR_BASE_URL>:9380")
assistant = rag.list_chats(name="Miss R")
assistant = assistant[0]
assistant.delete_sessions(ids=["id_1","id_2"])