I'm a Dual Degree student in Robotics at the Indian Institute of Technology Madras. I work with Prof. Anuj Tiwari at the DiRO Lab on Multi-agent Quadruped Robots.
I also collaborate with Prof. Arvind Easwaran at the CPS Research Group, NTU Singapore, where I focus on trustworthy learning agents for F1TENTH Autonomous Racing.
I worked as an AI Research Intern at Qneuro India Pvt Ltd, developing deep learning models for inferring EEG, PCG Signals.
I interned at HiRO Lab, IISC Bangalore, where I focused on learning for bimanual robot manipulation.
I was part of Team Anveshak, the student-run Mars Rover team at IIT Madras that competes in international rover competitions. I initially contributed as an Embedded Engineer before being promoted to Electronics and Software Lead.
I explore Reinforcement Learning, Deep Learning, Generative AI, LLMs and Robotics. My research interests lie in developing intelligent, model-based, heirarchical controllers for robotic applications. In parallel, I explore advancing reasoning capabilities in LLMs and fine-tuning them using RLHF and SFT.
Projects
Residual Reinforcement Learning for F1TENTH Racing
NTU Global Connect Fellow (2025)
Gokul M K, Eduardo de Conto, Subrat Prasad Panda, Arvind Easwaran
Improving the baseline tracking controller with residual Reinforcement Learning. Generalization across various racing tracks, static and dynamic obstacles.
Modulated Dynamical Systems for Coordinated Bimanual Manipulation
HiRO Lab, Robotics Summer Intern, RBCCPS IISC Bangalore
Gokul M K, Dr. Ravi Prakash
Implemented the research by LASA Lab, EPFL on modulated dynamical systems for Coordinated Bimanual Robotic Manipulation. Also worked on implementing Tossingbot by learning the residual physics for throwing.
Exploring Various Robotic Grasping Algorithms
e-Yantra Summer Intern, IIT Bombay
Gokul M K, Archit Jain, Jaison Jose, Ravikumar Chaurasia
Compared various learning-based, analytical grasping algorithms and benchmarked their perfomance in simulation and hardware. Formulated a light-weight grasping algorithm using Euclidean clustering.
Implicit Reinforcement Learning without Interaction at Scale
DA7400, Recent Advances in Reinforcement Learning Keerthivasan M, Gokul M K
Addressing sub-optimality and diversity challenges in Offline-RL trained with large datasets collected for long-horizon tasks. Employed a heirarchical framework and validated its performance.
Denoising and Deblurring MVTEC AD Dataset
EE5179, Deep Learning for Imaging Gokul M K, Keerthivasan M
Optimized RIDNet, a deep learning model to denoise and deblur MVTEC dataset. It achieved a PSNR of 32.8 on the training examples and 34.78 on the test examples. Benchmarked the results with other architectures.
Comparative Study of SMDP and Intra-Option Learning in the Taxi Domain
CS6700, Introduction to Reinforcement Learning Gokul M K, Keerthivasan M
Compared two option learning methods, SMDP and Intra Option Q learning in the taxi gym environment. Intra-Option made faster updates while an action is taken rather than waiting for the option to end.
Trajectory Continuous Optimal Planning for a Mobile Manipulator
ED5215, Introduction to Motion Planning Gokul M K, Keerthivasan M
Continuously tracing a trajectory using RRT* while minimizing the deviation in the end-effector pose through an Optimal control formulation. This shares similarities with 3D printing task by mobile manipulators.