7+ years of software work experience with domain expertise in LLM foundation model training and inference, distributed cloud infrastructure for ML and big data pipelines. Demonstrated experience in technical leadership, leading cross-organizational initiatives, establishing technical standards and best practices.
Download my resume hereA collection of projects authored by Arjun, and likely shared out with the community as an open source project.
Express Wiki gives you a personal online wikipidea powered by AI so you can read and organize your wikipidea.
Google Chrome Extension that generates a caption for an image using deep learning network RCNN
A collection of efforts to which I contributed, but did not create. Contributing back to Open Source projects is a strong passion of mine, and requires a considerate approach to learn norms, standards and approach for each community for a successful merge!
Worked and experimented with CLR algorithm which is implemented in Keras that uses cyclical learning rates to train deep neural networks.
Won Aurora hackathon 2021 for developing novel prefetch algorithm for data read ahead for Aurora MySQL 8.0 engine.
Won prestigious kudos award 2020 for my contributions at CoreTech Adobe among team of 100+ builders.
Winner Adobe shark tank event for best product idea pitch amongst 100+ new hires.
Best B.Tech. Project Award received from director IIT Roorkee amongst 1816 students.
Secured All India Rank of 401 among ~1,50,000 participants for Kishore Vaigyanik Protsahan Yojana
All India Senior School Certificate Examination 2015 certificate received from education minister India.
Led cross-organizational effort across engineering, research, and infosec teams to architect and launch Palmyra self-evolving models that surpass frozen LLM perf ceiling on enterprise agentic workflows by 23% through continuous learning from model interactions. Led team of 7 engineers to develop Palmyra X5, achieving GPT 4.1 level perf at 4X lower cost and state-of-the-art latency (22s at 1M tokens).
Led development and launch for SageMaker HyperPod Distributed Training Recipes, comprising a team of 10, including senior ICs. Drove cross-team collaborations and integrations. These open-source recipes (80+ stars in GitHub) automate training by loading datasets, applying distributed training techniques, and managing checkpoints. Lead for SageMaker Distributed Model Parallel library, delivering up to 20% perf gains over PyTorch FSDP for large-scale LLM training. Improved parallelism of LLMs like Llama, Qwen & GLM by leading team of 5 ICs.
Developed machine learning ecosystem infrastructure Aurora ZeroETL and Bedrock and SageMaker integrations for Aurora, enabling native ML workflows for Aurora customers. Developed a novel prefetch algorithm for data read-ahead by incorporating clustering algorithms based on data access patterns, increasing SQL read throughput by 30%.
Developed Genshop, Adobe's core deep learning library for generative models, adopted across multiple products. Implemented CPU and GPU workflows with CUDA primitives for Adobe color, graphics & font engine.
Built a personalized document-highlighting system that used prior reader interactions to surface the most relevant content in new documents. Incorporated information retrieval and reinforcement learning techniques for keyword extraction.
Researched on action recognition with feature fusion which involved combining traditional computer vision techniques with deep learning to improve the accuracy of action recognition by 15%.