Machine translation
This project showcases the creation of a machine translation model that translates English to Hindi, leveraging Transformer architecture.
It explores essential components, including multi-head attention, cross-attention, masked attention, positional embedding, and feed-forward neural networks.
The article provides a comprehensive implementation in PyTorch along with optimization strategies to improve model performance.
View
Fine-Tuning LLM
We fine-tuned the Meta Llama-3.1–8B model to simplify Indian legal texts, including BNS sections and public & administrative laws.Using LoRA with Unsloth, we optimized training on e2e Networks' GPUs. To improve clarity, legal texts were rewritten in plain English with Gemini 2.0 Flash and structured in Alpaca-style instruction-response pairs.The model was deployed via vLLM, providing fast, scalable access through a REST API, making legal information more accessible to lawyers, students, and citizens.
View
BollywoodLyricsGPT
Developed a custom GPT-2 model utilizing Karpathy's implementation, specifically tailored for generating coherent Hindi lyrics.
The model was trained on a dataset of 10,000 Bollywood lyrics, featuring 124 million parameters and 300,000 tokens.
It incorporates 12 decoder layers for enhanced language understanding and employs the GPT-2 tokenizer, showcasing advanced NLP and deep learning techniques in PyTorch.
View
Cricket Analysis of IPL Data
This project involves a comprehensive analysis of IPL data, focusing on various aspects of batsmen and bowlers across different match situations.
By employing diverse visualization techniques, I aim to uncover valuable insights into player performance, helping to understand trends and patterns that influence game outcomes.
View