Research Experiences

Multi-Agent Systems Research Group

Multi-Agent Systems Research Group

I'm a member of the University of Minnesota Multi-Agent Systems Research Group in the Next Generation Robotics Laboratory, where we study learning, cooperation, and adaptation in multi-agent robot systems.

Go2 picture

Reinforcement learning, simulation, and real robots
Researched reinforcement learning for quadruped locomotion with the Unitree Go2. Developed simulation environments and utilized reward design to achieve walking behavior. Built workflows connecting simulation with real-robot control for Sim-to-Real transfer.
Github for setting up an environment for communicating with a Go2 quadruped robot can be found here.
Github for training walking behavior for a Go2 in MuJoCo simulation using Stablebaselines3 PPO and SAC (and other policy gradient algorithms) can be found here.

Group picture

Applied robotics in agriculture
Participated in the Farm Robotics Challenge, to develop an award-winning system FarmGuard, which was designed to reduce deer intrusion in agricultural fields. Worked to combine hardware, computer vision, and path planning to effectively patrol and protect farmlands in collaboration with a local farmer.

TurtleBot3 robots used in UROP project

Fall Symposium Presentation
I presented this work at the University of Minnesota Undergraduate Research Symposium, highlighting the motivation, setup, and early platform results that enable future experiments in embodied evolution and reinforcement learning on physical TurtleBot3 robots. Additionally, you can read the abstract for this project below.

Social Learning and Behavioral Plasticity in Multi-Agent Robot Systems through Reinforcement Learning and Embodied Evolution
A key objective in artificial intelligence and robotics is to develop systems capable of autonomous decision-making and adaptation to dynamic environments. Evolutionary robotics, a sub-field of robotics, aims to achieve this by drawing inspiration from natural evolution and applying principles of gene selection and variation within robots to enable adaptation. Embodied evolution, a subset of evolutionary robotics, distinguishes itself from other evolutionary algorithms by distributing control among the robots rather than centrally, allowing for asynchronous, autonomous, and continuous evolution. Reinforcement learning is a method for agents to learn autonomously based on trial-and-error learning. Researchers suggest that combining reinforcement learning with embodied evolution may enhance robotic collectives' adaptability in dynamic settings, an idea inspired by studies of behavioral plasticity and social learning in animal populations. This research takes an essential step toward testing these ideas in physical systems rather than simulations, which are commonly used but fail to address many real-world complexities. Using TurtleBot3 robots equipped with Raspberry Pi 4 and 5 modules, this work involved constructing the robots, equipping them with hardware, and setting up ROS 2 as a foundational environment for implementing reinforcement learning and embodied evolution. While the theoretical algorithms remain to be tested, this groundwork is crucial for transitioning these concepts from simulation to reality. The emphasis on physical deployment ensures that the challenges and nuances of real-world applications, often overlooked in simulation, are addressed, paving the way for adaptive robotic systems with greater practical utility.