You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+13-14Lines changed: 13 additions & 14 deletions
Original file line number
Diff line number
Diff line change
@@ -8,30 +8,36 @@ The purpose of this package is to offer a convenient question-answering (RAG) sy
8
8
9
9
## Features
10
10
11
-
* Supported formats
11
+
* Supported document formats
12
12
* Build-in parsers:
13
13
* `.md` - Divides files based on logical components such as headings, subheadings, and code blocks. Supports additional features like cleaning image links, adding custom metadata, and more.
* Other common formats are supported by `Unstructured` pre-processor:
17
17
* List of formats see [here](https://unstructured-io.github.io/unstructured/core/partition.html).
18
18
19
-
* Support for table parsing via open-source gmft (https://github.com/conjuncts/gmft) or Azure Document Intelligence.
20
-
21
-
* Optional support for image parsing using Gemini API.
22
-
23
-
* Supports multiple collection of documents, and filtering the results by a collection.
19
+
* Allows interaction with embedded documents, internally supporting the following models and methods (including locally hosted):
20
+
* OpenAI models (ChatGPT 3.5/4 and Azure OpenAI).
21
+
* HuggingFace models.
22
+
* Llama cpp supported models - for full list see [here](https://github.com/ggerganov/llama.cpp#description).
24
23
25
-
*An ability to update the embeddings incrementally, without a need to re-index the entire document base.
24
+
*Interoperability with LiteLLM + Ollama via OpenAI API, supporting hundreds of different models (see [Model configuration for LiteLLM](sample_templates/llm/litellm.yaml))
26
25
27
26
* Generates dense embeddings from a folder of documents and stores them in a vector database ([ChromaDB](https://github.com/chroma-core/chroma)).
* Generates sparse embeddings using SPLADE (https://github.com/naver/splade) to enable hybrid search (sparse + dense).
34
34
35
+
* An ability to update the embeddings incrementally, without a need to re-index the entire document base.
36
+
37
+
* Support for table parsing via open-source gmft (https://github.com/conjuncts/gmft) or Azure Document Intelligence.
38
+
39
+
* Optional support for image parsing using Gemini API.
40
+
35
41
* Supports the "Retrieve and Re-rank" strategy for semantic search, see [here](https://www.sbert.net/examples/applications/retrieve_rerank/README.html).
36
42
* Besides the originally `ms-marco-MiniLM` cross-encoder, more modern `bge-reranker` is supported.
37
43
@@ -44,13 +50,6 @@ The purpose of this package is to offer a convenient question-answering (RAG) sy
44
50
45
51
* Supprts optional chat history with question contextualization
46
52
47
-
* Allows interaction with embedded documents, internally supporting the following models and methods (including locally hosted):
48
-
* OpenAI models (ChatGPT 3.5/4 and Azure OpenAI).
49
-
* HuggingFace models.
50
-
* Llama cpp supported models - for full list see [here](https://github.com/ggerganov/llama.cpp#description).
51
-
* AutoGPTQ models (temporarily disabled due to broken dependencies).
52
-
53
-
* Interoperability with LiteLLM + Ollama via OpenAI API, supporting hundreds of different models (see [Model configuration for LiteLLM](sample_templates/llm/litellm.yaml))
Copy file name to clipboardExpand all lines: docs/index.rst
+22-23Lines changed: 22 additions & 23 deletions
Original file line number
Diff line number
Diff line change
@@ -11,59 +11,58 @@ The purpose of this package is to offer a convenient question-answering system w
11
11
Features
12
12
--------
13
13
14
-
* Supported formats
14
+
* Supported document formats
15
15
* Build-in parsers:
16
16
* `.md` - Divides files based on logical components such as headings, subheadings, and code blocks. Supports additional features like cleaning image links, adding custom metadata, and more.
* Other common formats are supported by `Unstructured` pre-processor:
20
-
* List of formats https://unstructured-io.github.io/unstructured/core/partition.html
20
+
* List of formats see [here](https://unstructured-io.github.io/unstructured/core/partition.html).
21
21
22
-
* Support for table parsing via open-source gmft (https://github.com/conjuncts/gmft) or Azure Document Intelligence.
23
-
24
-
* Optional support for image parsing using Gemini API.
25
-
26
-
* Supports multiple collection of documents, and filtering the results by a collection.
27
-
28
-
* An ability to update the embeddings incrementally, without a need to re-index the entire document base.
22
+
* Allows interaction with embedded documents, internally supporting the following models and methods (including locally hosted):
23
+
* OpenAI models (ChatGPT 3.5/4 and Azure OpenAI).
24
+
* HuggingFace models.
25
+
* Llama cpp supported models - for full list see [here](https://github.com/ggerganov/llama.cpp#description).
29
26
30
-
* Generates dense embeddings from a folder of documents and stores them in a vector database (ChromaDB).
27
+
* Interoperability with LiteLLM + Ollama via OpenAI API, supporting hundreds of different models (see [Model configuration for LiteLLM](sample_templates/llm/litellm.yaml))
31
28
29
+
* Generates dense embeddings from a folder of documents and stores them in a vector database ([ChromaDB](https://github.com/chroma-core/chroma)).
* Generates sparse embeddings using SPLADE (https://github.com/naver/splade) to enable hybrid search (sparse + dense).
39
37
40
-
* Supports the "Retrieve and Re-rank" strategy for semantic search, see - https://www.sbert.net/examples/applications/retrieve_rerank/README.html.
38
+
* An ability to update the embeddings incrementally, without a need to re-index the entire document base.
39
+
40
+
* Support for table parsing via open-source gmft (https://github.com/conjuncts/gmft) or Azure Document Intelligence.
41
+
42
+
* Optional support for image parsing using Gemini API.
43
+
44
+
* Supports the "Retrieve and Re-rank" strategy for semantic search, see [here](https://www.sbert.net/examples/applications/retrieve_rerank/README.html).
41
45
* Besides the originally `ms-marco-MiniLM` cross-encoder, more modern `bge-reranker` is supported.
* Supports HyDE (Hypothetical Document Embeddings) - see [here](https://arxiv.org/pdf/2212.10496.pdf).
44
48
* WARNING: Enabling HyDE (via config OR webapp) can significantly alter the quality of the results. Please make sure to read the paper before enabling.
45
-
* Based on empirical observations, enabling HyDE significantly boosts quality of the output on a topics where user can't formulate the quesiton using domain specific language of the topic - e.g. when learning new topics.
49
+
* From my own experiments, enabling HyDE significantly boosts quality of the output on a topics where user can't formulate the quesiton using domain specific language of the topic - e.g. when learning new topics.
46
50
47
51
* Support for multi-querying, inspired by `RAG Fusion` - https://towardsdatascience.com/forget-rag-the-future-is-rag-fusion-1147298d8ad1
48
52
* When multi-querying is turned on (either config or webapp), the original query will be replaced by 3 variants of the same query, allowing to bridge the gap in the terminology and "offer different angles or perspectives" according to the article.
49
53
50
54
* Supprts optional chat history with question contextualization
51
55
52
-
* Allows interaction with embedded documents, internally supporting the following models and methods (including locally hosted):
53
-
* OpenAI models (ChatGPT 3.5/4 and Azure OpenAI).
54
-
* HuggingFace models.
55
-
* Llama cpp supported models - for full list see https://github.com/ggerganov/llama.cpp#description
56
-
* AutoGPTQ models (temporarily disabled due to broken dependencies).
57
-
58
-
* Interoperability with LiteLLM + Ollama via OpenAI API, supporting hundreds of different models (see [Model configuration for LiteLLM](sample_templates/llm/litellm.yaml))
59
56
60
57
* Other features
61
-
* Simple web interface.
58
+
* Simple CLI and web interfaces.
62
59
* Deep linking into document sections - jump to an individual PDF page or a header in a markdown file.
63
60
* Ability to save responses to an offline database for future analysis.
0 commit comments