This project is a demonstration of a Retrieval-Augmented Generation (RAG) application that works with unstructured data. It uses a Large Language Model (LLM) to answer questions based on the content of a provided web page.
The application fetches content from a specified URL, splits it into manageable chunks, and then uses a vector store to efficiently retrieve relevant information. When a user asks a question, the application retrieves the most relevant context from the vector store and uses it to generate a concise and accurate answer.
- LangChain: A framework for developing applications powered by language models.
- Google Gemini: The Large Language Model used for generating answers.
- ChromaDB: A vector store for efficient similarity search.
- Streamlit: For creating the user interface.
-
Clone the repository:
git clone https://github.com/your-username/RAG_App_Unstructed.git
-
Install the dependencies:
pip install -r requirements.txt
-
Set up your environment variables:
Create a
.envfile in the root directory and add your Google API key:GOOGLE_API_KEY="your-google-api-key" -
Run the application:
python app.py
Build an android app and deploy it