Apple iPhone Support RAG Chatbot
π§ Project Overview
This project showcases a Retrieval-Augmented Generation (RAG) chatbot capable of answering Apple iPhone support questions by referencing real documentation scraped from Appleβs official site.
The chatbot runs locally or with OpenAI models and supports streaming responses using Gradio. It demonstrates modern AI implementation practices including local inference with Ollama and hybrid deployment with GPT-4 for enterprise-grade reliability.
π§ Technologies Used
- LangChain β RAG pipeline orchestration with ConversationalRetrievalChain
- FAISS β Efficient vector similarity search with saved local index
- OpenAI Embeddings β For cloud-based text representation and semantic search
- OpenAI GPT4 β For LLM-based text generation
- Ollama β For local LLM inference using LLaMA 3 and nomic-embed-text
- Gradio β Interactive frontend for chatbot demo with model switching
- Playwright (Python) β For automated document scraping of Apple Support
π Features
- π Dynamic model selection between
llama3.2
(local but not shown in the current demo) andgpt-4
(cloud) - π§© Document chunking and embedding with semantic retrieval
- ποΈ Vector search via FAISS to enhance question answering
- π¬ Conversation memory with context persistence
- π Self-hosted or deployable via Docker
π― Why It Matters
This project is ideal for:
- Companies exploring AI-based customer support automation
- Teams evaluating RAG pipelines and vector databases
- Clients requiring private LLM deployment or hybrid cloud-local solutions
π Try It Yourself
You can take a look at the Demo HERE.

The entire project is containerized with Docker Compose, allowing easy deployment in local or cloud environments.
Want a similar AI assistant trained on your own knowledge base or product documentation?