The power of LLMs has revolutionized offline document interaction, allowing users to ask questions and receive answers without the need for an internet connection. By ensuring complete privacy and keeping data within your execution environment, LLMs enable you to explore your documents with confidence and peace of mind. Embrace this innovative technology, break free from online dependencies, and unlock the full potential of offline document interaction with the power of LLMs.
Ask questions to your documents without an internet connection, using the power of LLMs. 100% private, no data leaves your execution environment at any point. You can ingest documents and ask questions without an internet connection!
In order to set your environment up to run the code here, first install all requirements:
pip3 install -r requirements.txt
Then, download the LLM model and place it in a directory of your choice:
.env
file.Copy the example.env
template into .env
cp example.env .env
and edit the variables appropriately in the .env
file.
The power of LLMs has revolutionized offline document interaction, allowing users to ask questions and receive answers without the need for an internet connection. By ensuring complete privacy and keeping data within your execution environment, LLMs enable you to explore your documents with confidence and peace of mind. Embrace this innovative technology, break free from online dependencies, and unlock the full potential of offline document interaction with the power of LLMs.
The supported extensions are:
.csv
: CSV,.docx
: Word Document,.doc
: Word Document,.enex
: EverNote,.eml
: Email,.epub
: EPub,.html
: HTML File,.md
: Markdown,.msg
: Outlook Message,.odt
: Open Document Text,.pdf
: Portable Document Format (PDF),.pptx
: PowerPoint Document,.ppt
: PowerPoint Document,.txt
: Text file (UTF-8),Run the following command to ingest all the data.
python ingest.py
Output should look like this:
Creating new vectorstore
Loading documents from source_documents
Loading new documents: 100%|██████████████████████| 1/1 [00:01<00:00, 1.73s/it]
Loaded 1 new documents from source_documents
Split into 90 chunks of text (max. 500 tokens each)
Creating embeddings. May take some minutes...
Using embedded DuckDB with persistence: data will be stored in: db
Ingestion complete! You can now run privateGPT.py to query your documents
It will create a db
folder containing the local vectorstore. Will take 20-30 seconds per document, depending on the size of the document. You can ingest as many documents as you want, and all will be accumulated in the local embeddings database. If you want to start from an empty database, delete the db
folder.
Note: during the ingest process no data leaves your local environment. You could ingest without an internet connection, except for the first time you run the ingest script, when the embeddings model is down
pip3 install -r requirements.txt
python privateGPT.py
The web assistant should be able to provide quick and effective solutions to the user's queries, and help them navigate the website with ease.
The Web assistant is more then able to personalize the user's experience by understanding their preferences and behavior on the website.
The Web assistant can help users troubleshoot technical issues, such as broken links, page errors, and other technical glitches.
Please log in to gain access on Document Interaction: Ask Questions with LLMs in a 100% Private Environment file .