This project is a conversational AI chatbot built using the Llama language model and Streamlit for the user interface. The chatbot can engage in meaningful conversations, provide responses to user queries, and adapt to different use cases such as customer support, educational tools, and more.
- Llama Language Model: Uses the powerful Llama model for natural language understanding and generating responses.
- Streamlit Interface: Provides an easy-to-use and interactive UI, making it simple to interact with the chatbot.
- Modular Architecture: Can be scaled or adapted for different use cases such as FAQ bots, personal assistants, or educational tools.
- Real-Time Interaction: Responds to user inputs in real-time with contextually relevant information.
- Python 3.9+
- Llama Model (ensure it's correctly installed)
- Streamlit
- Clone the repository:
git clone https://github.com/naveennk045/LlamaConvo cd llama-chatbot - Create a virtual environment:
python3 -m venv env source env/bin/activate # On Windows use `env\Scripts\activate`
- Install the required dependencies:
pip install -r requirements.txt Download the Llama model: Follow the instructions here to download the model and configure it for usage. - Run the chatbot:
streamlit run app.py
- Launch the chatbot by running the command streamlit run app.py.
- Open your browser and navigate to http://localhost:8501.
- Start interacting with the chatbot by entering your queries in the input box.
- Llama: For language understanding and response generation.
- Streamlit: For building a web-based user interface.
- Python: Core programming language for backend logic and integration.
Contributions are welcome! Please feel free to submit a Pull Request or open an Issue for any feature suggestions or bugs.
This project is licensed under the MIT License - see the LICENSE file for details.