| Status: Ongoing Research & Implementation Repo | Stack: PyTorch, Hugging Face, LangChain, FAISS |
Modern NLP relies heavily on abstraction layers. To master the fundamentals of Large Language Models (LLMs), I implemented core deep learning architectures from scratch in PyTorch. This project serves as a testbed for benchmarking attention mechanisms, sequence modeling techniques, and retrieval strategies.
Implemented recurrent architectures to analyze the Vanishing Gradient Problem and long-term dependency retention.
Replicated the “Attention Is All You Need” architecture without using nn.Transformer.
Built a modular RAG pipeline to ground LLM responses in external data.
Retrieval: Implemented similarity search using FAISS (Facebook AI Similarity Search) to retrieve Top-K relevant contexts for the generator.
</div>