🧠
Model

Llamaindex Agent

by swastikmaiti gh-model--swastikmaiti--llamaindex-agent
Nexus Index
0.0 Top 18%
P: Popularity 0
F: Freshness 0
C: Completeness 0
U: Utility 0
Tech Context
Vital Performance
0 DL / 30D
0.0%

An Llama Index based Agentic-RAG system to perform PDF Question-Answering. The Agent can choose from or to generate response. The LLM used is 3.8B. !alt text - ***Agentic-RAG:*** Llama Index - ***App:*** Gradio - ***LLM:*** Phi3 3.8B - ***Embedding:*** nomic-embed-text - ***Local LLM:*** Ollama - ***llamaindex_basic.ipynb:*** A simple introduction to Llama Index Agentic RAG concepts and terminologies. - ***agentic_rag_intro.ipynb:*** This notebook contains codes and step by step explanation o...

Tiny - Params
- Context
0 Downloads
Model Information Summary
Entity Passport
Registry ID gh-model--swastikmaiti--llamaindex-agent
Provider github
📜

Cite this model

Academic & Research Attribution

BibTeX
@misc{gh_model__swastikmaiti__llamaindex_agent,
  author = {swastikmaiti},
  title = {Llamaindex Agent Model},
  year = {2026},
  howpublished = {\url{https://github.com/swastikmaiti/LlamaIndex-Agent}},
  note = {Accessed via Free2AITools Knowledge Fortress}
}
APA Style
swastikmaiti. (2026). Llamaindex Agent [Model]. Free2AITools. https://github.com/swastikmaiti/LlamaIndex-Agent

đŸ”ŦTechnical Deep Dive

Full Specifications [+]

Quick Commands

🐙 Git Clone
git clone https://github.com/swastikmaiti/LlamaIndex-Agent

âš–ī¸ Nexus Index V16.5

0.0
TOP 18% SYSTEM IMPACT
Popularity (P) 0
Freshness (F) 0
Completeness (C) 0
Utility (U) 0

đŸ’Ŧ Index Insight

The Free2AITools Nexus Index for Llamaindex Agent aggregates Popularity (P:0), Freshness (F:0), and Completeness (C:0). The Utility score (U:0) represents deployment readiness and ecosystem adoption.

Free2AITools Nexus Index

Verification Authority

Unbiased Data Node Refresh: VFS Live
---

🚀 What's Next?

Technical Deep Dive

AGENTIC-RAG

An Llama Index based Agentic-RAG system to perform PDF Question-Answering. The Agent can choose from summarization query engine or vector query engine to generate response. The LLM used is phi3 3.8B.

alt text

Frameworks

  • Agentic-RAG: Llama Index
  • App: Gradio
  • LLM: Phi3 3.8B
  • Embedding: nomic-embed-text
  • Local LLM: Ollama

File Structure

  • llamaindex_basic.ipynb: A simple introduction to Llama Index Agentic RAG concepts and terminologies.
  • agentic_rag_intro.ipynb: This notebook contains codes and step by step explanation of how to build an Agentic-RAG with Llama Index.
  • agentic_rag_customization.ipynb Customizing the Agentic-RAG system to perform pdf Q/A with Phi3
  • utils.py Contains all the functions in one place.
  • app.py Creating Gradio application.

Introduction

RAG is a wonderful solution to make LLM even smarter with Memeory. However RAG is a single end2end pipeline. User will have various kind of queries which will require diffrent kind of processing with a specialized pipeline. This is where AGENTIC-RAG comes into action. A smart AGENT takes dicesion based on user queries and avaialble pipelines to fireup one or more of the pipelines to answer user queries.

Docker

For Docker Implementation of the Application Checkout the GitHub Repo. 🚛

Description

In this work we build a Agentic RAG with llamaindex. Retrieval Augmented Generation (RAG) is one of the most widespread usecase of LLM. In RAG there exist a single pipeline for the workflow. Hence all user queries are processed in exactly the same way. However there exist different types of user queries which may require different pipenine for processing. In this work we build two piplines to answer user queries with specific need. The pipelines are

  • Summarization pipeline
  • Question-Answering pipeline

Decription of files in sequence they were developed

The code description are provided within the files.

  • llamaindex_basic.ipynb: a brief intro on llamaindex framework
  • agentic_rag_intro.ipynb: a brief introduction to agentic rag development.
  • agentic_rag_customization.ipynb: the notebook for complete code on developing the agentic rag to answer user queries from a pdf file.
  • app.py: finally build a Application with Gradio. This is build on top of agentic_rag_customization.ipynb so all the necessary functions are present in utils.py.

How to RUN

  • All the work is developed in LINUX env so we need a LINUX system with atleast 8GB RAM.
  • Create a Virtual Env
  • Install libraries with make install
  • Download Ollama and start Ollama server with make ollama_download on a new CLI as this will block the CLI.
  • Pull models required for tasks with make models
  • To Start Graio App run python app.py

Acknowledgements

  • Thanks to DeepLearning.AI and LlamaIndex for the wonderful course
  • Thanks to Microsoft for open source Phi3

If you find the repo helpful, please drop a ⭐

📝 Limitations & Considerations

  • â€ĸ Benchmark scores may vary based on evaluation methodology and hardware configuration.
  • â€ĸ VRAM requirements are estimates; actual usage depends on quantization and batch size.
  • â€ĸ FNI scores are relative rankings and may change as new models are added.
  • ⚠ License Unknown: Verify licensing terms before commercial use.
  • â€ĸ Source: Unknown
🔄 Daily sync (03:00 UTC)

AI Summary: Based on GitHub metadata. Not a recommendation.

📊 FNI Methodology 📚 Knowledge Baseâ„šī¸ Verify with original source

đŸ›Ąī¸ Model Transparency Report

Verified data manifest for traceability and transparency.

100% Data Disclosure Active

🆔 Identity & Source

id
gh-model--swastikmaiti--llamaindex-agent
source
github
author
swastikmaiti
tags
agentic-ragagentic-workflowgradiollamaindexollamaphi3ragjupyter notebook

âš™ī¸ Technical Specs

architecture
null
params billions
null
context length
null
pipeline tag
feature-extraction

📊 Engagement & Metrics

likes
0
downloads
0

Free2AITools Constitutional Data Pipeline: Curated disclosure mode active. (V15.x Standard)