|
@@ -1,20 +0,0 @@
|
|
|
-# Run TAIDE RAG in VM
|
|
|
-
|
|
|
-## Prerequisites
|
|
|
-- Docker and Docker Compose
|
|
|
-- Ollama (for creating the taide-local model)
|
|
|
-- Download TAIDE .gguf file. https://huggingface.co/taide/TAIDE-LX-7B-Chat-4bit/tree/main?show_file_info=taide-7b-a.2-q4_k_m.gguf Update the file path in `Modelfile`
|
|
|
-
|
|
|
-## Setting up taide-local
|
|
|
-
|
|
|
-1. Install Ollama on your VM
|
|
|
-`curl -fsSL https://ollama.com/install.sh | sh`
|
|
|
-2. Create the taide-local model
|
|
|
-`ollama create taide-local -f Modelfile`
|
|
|
-
|
|
|
-## Running the Applciation
|
|
|
-1. Clone this repository.
|
|
|
-`git clone -b public https://github.com/yourusername/your-repo.git`
|
|
|
-2. Create a `.env` file in the project root with your API keys.
|
|
|
-3. Run `docker-compose up --build`
|
|
|
-4. The application will be available at `http://localhost:8000`.
|