Profile banner
Aditya Shah

Aditya Shah

ML Research Engineer at Google focused on Gemini Foundation Models

Bay Area, California
Joined December 2025

Network

1.4K connections
🧠
AI/ML Researchers
💻
AI/ML Engineers
🦃
Virginia Tech CS Faculty
🤝
AI/Tech Recruiters
🚀
AI Startup Founders
⚖️
Immi Law
🏦
Capital One

Summary

Aditya Shah is a Machine Learning Research Engineer at Google, specializing in large language models, reinforcement learning, and post-training evaluations for the Gemini family of Foundation Models. His work involves collaborating with DeepMind and Google Cloud research to enhance underlying ML models for Multimodal Document Extraction. linkedin+3
Aditya possesses strong academic and research credentials with a Master of Science in Computer Science from Virginia Tech and a Bachelor of Engineering from Dwarkadas J. Sanghvi College of Engineering, both with high GPAs. He has published several papers in areas such as NLP, multimodal AI, and quantum computing, and holds an h-index of 6 and i10-index of 4, indicating a significant research impact. github+3
Aditya has diverse practical experience in various aspects of machine learning, including NLP, computer vision, and speech processing, gained through roles at Capital One, Saarthi.ai, Fynd, and QuickFits. He has developed and deployed ML models for financial risk identification, sarcasm detection, gender identification from audio, and visual apparel recommendation. github+1
Aditya has demonstrated leadership and initiative through his role as Co-Technical Head at ACM and as the founder of the 'Art of Quantum' blog, where he shares tutorials on Quantum Machine Learning. github+1

Work

Education

Writing

Gemini 2.5: Pushing the frontier with advanced reasoning, multimodality, long context, and next generation agentic capabilities

January 1, 2025

A research paper on the advancements and capabilities of the Gemini 2.5 model.

Favicon imagescholar.google.com

What’s in a cue?: Using natural language processing to quantify content characteristics of episodic future thinking in the context of overweight and obesity

January 1, 2025

A research paper exploring the use of NLP to quantify characteristics of episodic future thinking in relation to overweight and obesity.

Favicon imagescholar.google.com

Did you tell a deadly lie? evaluating large language models for health misinformation identification

January 1, 2024

Research evaluating large language models for their ability to identify health misinformation.

Favicon imagescholar.google.com

End-to-end multimodal fact-checking and explanation generation: A challenging dataset and models

January 1, 2023

Research on multimodal fact-checking and explanation generation, presenting a new challenging dataset and models.

Favicon imagescholar.google.com

Adept: Adapter-based efficient prompt tuning approach for language models

January 1, 2023

A research paper proposing an adapter-based efficient prompt tuning approach for language models.

Favicon imagescholar.google.com

Retrieval-based text selection for addressing class-imbalanced data in classification

January 1, 2023

A study on using retrieval-based text selection to handle class-imbalanced data in classification problems.

Favicon imagescholar.google.com

Leveraging Transformer Models and Elasticsearch to Help Prevent and Manage Diabetes through EFT Cues

January 1, 2023

Aditya Shah's Master's thesis, focusing on using Transformer models and Elasticsearch for diabetes prevention and management.

Favicon imagescholar.google.com

Filming multimodal sarcasm detection with attention

January 1, 2021

A paper exploring multimodal sarcasm detection using attention mechanisms.

Favicon imagescholar.google.com

How effective is incongruity? Implications for code-mixed sarcasm detection

January 1, 2021

A paper discussing the effectiveness of incongruity in code-mixed sarcasm detection.

Favicon imagescholar.google.com

Evolution of Neural Text Generation: Comparative Analysis

January 1, 2020

A comparative analysis of various neural text generation algorithms, showcasing the benefits of context-dependent models like ELMo, BERT, and GPT-2.

Favicon imageaditya-shahh.github.io

Leveraging quantum computing for supervised classification

January 1, 2020

Research on using quantum computing for supervised classification tasks.

Favicon imageaditya-shahh.github.io

Texture Synthesis and Style Transfer for Aesthetic Design Creation

January 1, 2019

A publication on texture synthesis and style transfer, likely related to his undergraduate thesis.

Favicon imageaditya-shahh.github.io

Art of Quantum blog

January 1, 2019

A blog founded by Aditya Shah to publish tutorials and explain key concepts of Quantum Machine Learning using Python and IBM's Qiskit framework.

Favicon imageaditya-shahh.github.io

Hobbies

Enjoys gourmet cooking, reading books on psychology and finance. youtube