This project is a locally-run chatbot built with Chainlit, LangChain, and the Ollama Llama 3.1 model. It supports real-time streaming responses and conversation history, providing an intelligent, context-aware conversation experience—all on your local machine. 🖥️
- Locally run: No external services required, everything runs on your machine.
- ⚡ Streaming responses: Token-by-token responses for faster, more interactive chats.
- 💬 Conversation history: The bot remembers your previous messages for more coherent replies.
-
Install dependencies:
poetry install
-
Start Ollama: Make sure you have the Ollama model ready to serve:
ollama pull llama3.1:8b ollama serve
-
Run the Chainlit app:
poetry run chainlit run app.py
-
Access the chatbot: Visit
http://localhost:8000
in your browser to start chatting with the bot. 🌍
- Chainlit for the user interface.
- LangChain to handle conversation logic and memory.
- Ollama Llama 3.1 as the language model backend.
This chatbot is ideal for testing and running advanced LLM models locally, offering real-time feedback and the ability to maintain conversation context. Perfect for experimenting with LLMs and local AI solutions without relying on cloud-based services.