Image of Ayush Agrawal

Budding Researcher
Cognitive AI

CV

Hello I'm Ayush! A researcher passionate to equip embodied agents with human like learning capabilities. Currently, I am a 1st year CS PhD student at Northeastern University, advised by Prof. Lawson Wong. Prior to this, I spent time working on leveraging useful language based abstractions for open world navigation with Prof. David Hsu.

Education

  • PhD Computer Science 2025 - Present
    Northeastern University, Boston, MA, USA
  • B.E. in Electronics & Instrumentation, 2022
    BITS Pilani, Pilani

News

Publications

image of submitted paper
Sign Language: Towards Sign Understanding for Robot Autonomy
Ayush Agrawal*, Joel Loo*, Nicky Zimmerman*, David Hsu
Under Review|paper|code|dataset

We introduce navigational sign understanding – parsing locations and directions from signs to aid robot localization, navigation and scene understanding. We provide a new dataset and benchmark some baselines, add active sign localization, and show VLMs’ promise on a Spot robot demo for real-world feasibility



image of submitted paper
SignLoc: Robust Localization using Navigation Signs and Public Maps
Nicky Zimmerman, Joel Loo, Ayush Agrawal, David Hsu
Under Review|paper|code

We present SignLoc, a global localization method that uses navigation signs to localize robots directly on floor plans and OSM maps without prior mapping. Using a probabilistic Monte Carlo framework, SignLoc achieves robust localization after just one or two sign observations across campus, mall, and hospital environments.


image of submitted paper
Physical Reasoning and Object Planning for Household Embodied Agents
Ayush Agrawal, Raghav Prabhakar, Anirudh Goyal, Dianbo Liu
TMLR 2024|OpenReview|code|dataset

Demystifying the decision making process behind choosing an object for task completion, we develop a 3 step architecture and curate datasets to power future research in this domain. Further, we evaluate various LLM baselines and report the findings.


Image of Ayush Agrawal
Sequence Agnostic MultiON
Nandiraju Gireesh*, Ayush A*, Ahana Datta* ... Madhava Krishna
ICRA 2023 |arxiv|video|blog

You are already in a kitchen, and tasked to find a fridge. Would you search for it in the current area or other places in house? We train a RL policy based off semantic relationship between static objects to generate efficient long term goals to enable quick retreival of a list of objects.


Image of Ayush Agrawal
CLIPGraphs: Multimodal Graph Networks to Infer Object-Room Affinities
Ayush A*, Raghav Arora*, ... Madhava Krishna
RO-MAN 2023 |arxiv|code|page

Leveraging upon the knowledge that we humans have highly developed Object-Utility and Room-Utility relationships; we generate human commonsense aligned latent embeddings useful for varius Embodied AI tasks. We do this by developing a Graph Neural Network by processing Human Preference datasets and Foundation Model Features.