Chahat Deep Singh

Ph.D. Dissertation Defense

Minimal Perception: Enabling Robot Autonomy on Resource-Constrained Robots





Monday June 26th, 2023
3:00 p.m. EST (9:00 pm CET)
IRB 3137
Zoom Link

Tuesday June 27th, 2023 | 12:30 am IST

Add to Calender

Abstract: Mobile robots are widely used and crucial in diverse fields due to their autonomous task performance. They enhance efficiency, and safety, and enable novel applications like precision agriculture, environmental monitoring, disaster management, and inspection. Perception plays a vital role in their autonomous behavior for environmental understanding and interaction. Perception in robots refers to their ability to gather, process, and interpret environmental data, enabling autonomous interactions. It facilitates navigation, object identification, and real-time reactions. By integrating perception, robots achieve onboard autonomy, operating without constant human intervention, even in remote or hazardous areas. This enhances adaptability and scalability. This thesis explores the challenge of developing autonomous systems for smaller robots used in precise tasks like confined space inspections and robot pollination. These robots face limitations in real-time perception due to computing, power, and sensing constraints. To address this, we draw inspiration from small organisms such as insects and hummingbirds, known for their sophisticated perception, navigation, and survival abilities despite their minimalistic sensory and neural systems. This research aims to provide insights into designing compact, efficient, and minimal perception systems for tiny autonomous robots. Embracing this minimalism is paramount in unlocking the full potential of tiny robots and enhancing their perception systems. By streamlining and simplifying their design and functionality, these compact robots can maximize efficiency and overcome limitations imposed by size constraints. In this work, I propose a Minimal Perception framework that enables onboard autonomy in resource-constrained robots at scales (as small as a credit card) that were not possible before. Minimal perception refers to a simplified, efficient, and \textit{selective} approach from both hardware and software perspectives to gather and process sensory information. Adopting a task-centric perspective allows for further refinement of the minimalist perception framework for tiny robots. For instance, certain animals like jumping spiders, measuring just 1/2 inch in length, demonstrate minimal perception capabilities through sparse vision facilitated by multiple eyes, enabling them to efficiently perceive their surroundings and capture prey with remarkable agility. The contributions of this work can be summarized as follows:

  • Utilizing minimal quantities such as uncertainty in optical flow and its untapped potential to enable autonomous drone navigation, static and dynamic obstacle avoidance, and the ability to fly through unknown gaps.
  • By utilizing the principles of interactive perception, the framework proposes novel object segmentation in cluttered environments eliminating the reliance on neural network training for object recognition.
  • Introducing a generative simulator called WorldGen that has the power to generate countless cities and petabytes of high-quality annotated data, designed to minimize the demanding need for laborious 3D modeling and annotations, thus unlocking unprecedented possibilities for perception and autonomy tasks.
  • I propose a method to predict metric dense depth maps in never-seen or out-of-domain environments by fusing information from a traditional RGB camera and a sparse 64-pixel depth sensor.
  • The autonomous capabilities of the tiny robots are demonstrated on both aerial and ground robots: (a) autonomous car with a size smaller than a credit card (70mm), and (b) bee drone with a length of 120mm, showcasing navigation abilities, depth perception in all four main directions, and effective avoidance of both static and dynamic obstacles.

  • Bio: Chahat Deep Singh is a fifth-year Ph.D. candidate in the Perception and Robotics Group (PRG) with Professor Yiannis Aloimonos and Associate Research Scientist Cornelia Fermüller. He graduated with Master in Robotics at the University of Maryland in 2018. Later, he joined as a Ph.D. student in the Department of Computer Science. Singh’s research focuses on developing bio-inspired minimalist cognitive architectures to enable onboard autonomy on robots that are as small as a credit card. He was awarded Ann G. Wylie Fellowship for outstanding dissertation for the year 2022-2023, Future Faculty Fellowship 2022-2023 and UMD's Dean Fellowship in 2020. Recently, his work was featured in BBC, IEEE Spectrum, Voice of America, NVIDIA, Futurism and much more. He has been serving as the PRG Seminar Series organizer since 2018 and has served as Maryland Robotics Center Student Ambassador from 2021 to 2023. Chahat is also a reviewer for RA-L, T-ASE, CVPR, ICRA, IROS, ICCV, RSS among other top journals and conferences. Chahat will join UMD as a Postdoctoral Associate at Maryland Robotics Center in July 2023 under the supervision of Prof. Yiannis Aloimonos and Prof. Pratap Tokekar. For more, please visit here.
    Examining Committee:
    Dr. Yiannis Aloimonos (Chair)
    Dr. Inderjit Chopra (Department Representative)
    Dr. Guido de Croon (TU Delft)
    Dr. Christopher Metzler
    Dr. Nitin J. Sanket
    Dr. Cornelia Fermüller


    About Me



    Ph.D. Student
    Perception and Robotics Group
    Computer Science

    I am a third year Ph.D. student and a Dean's Fellow in the Perception & Robotics Group (PRG) at University of Maryland, College Park (UMD), advised by Prof. Yiannis Aloimonos and Dr. Cornelia Fermuller. PRG is associated with the University of Maryland Institute of Advanced Computer Science Studies (UMIACS) and Autonomy, Robotics and Cognition Lab (ARC).

    Interests: Active perception and deep learning applications to boost multi-robot interaction and navigation ability in aerial robots.

    Prior to pursuing Ph.D., I did my Masters in Robotics at UMD where I worked on active behaviour of aerial robots and published GapFlyt where we used motion cues of the aerial agent to detect an unknown-shaped gap using a monocular camera. Apart from research work, I love to teach! I, along with Nitin J. Sanket designed and taught an open-source graduate course: ENAE788M (Hands-On Autonomous Aerial Robotics) at UMD in Fall 2019. In my spare time, I love to capture nature on my camera, especially landscape and wildlife photographs; watch and play competitive video games — Counter Strike and Dota 2.

    Teaching

    Instructor

    ENAE788M: Hands On Autonomous Aerial Robotics

    Fall 2019

    This is an advanced graduate course that exposes the students with mathematical foundations of computer vision, planning and control for aerial robots. This course was designed and taught by me and Nitin J. Sanket. The course is designed to balance theory with an application on hardware.

    The entire course is open-source! The links to video lectures and projects are given below:

    Teaching Assistant

    CMSC733: Computer Processing of Pictorial Information

    (Instructor: Prof. Yiannis Aloimonos)

    Spring 2020 | Spring 2019

    CMSC 733 is an advanced graduate course on classical and deep learning approaches for geometric computer vision and computational photography which explores through image formation, visual features, image segmentation, recognition, motion estimation and 3D point clouds. We redesigned this course to showcase how to model classical 3D geometry problems using Deep Learning!

    The entire course is open-source! The link to projects and student outputs are given below:

    Teaching Assistant

    CMSC426: Computer Vision

    (Instructor: Prof. Yiannis Aloimonos)

    Fall 2020 | Fall 2019 | Fall 2018

    CMSC 426 is an introductory course on computer vision and computational photography that explores image formation, image features, image segmentation, image stitching, image recognition, motion estimation, and visual SLAM.

    The entire course is open-source! The link to projects and student outputs are given below:

    Research

    NudgeSeg: Zero-Shot Object Segmentation by Repeated Physical Interaction

    IROS 2021

    Chahat Deep Singh*, Nitin J. Sanket*, Cornelia Fermuller, Yiannis Aloimonos, IEEE International Conference on Intelligent Robots and Systems (IROS), 2021.
    * Equal Contribution

     UMD    Paper    Project Page    Code    UMD    Cite  

    EVPropNet: Detecting Drones By Finding Propellers For Mid-Air Landing And Following

    RSS 2021

    Nitin J. Sanket, Chahat Deep Singh, Chethan M. Parameshwara, Cornelia Fermuller, Guido C.H.E. de Croon, Yiannis Aloimonos, Robotics Science and Systems (RSS), 2021.

     UMD  

    MorphEyes: Variable Baseline Stereo For Quadrotor Navigation

    ICRA 2021

    Nitin J. Sanket, Chahat Deep Singh, Varun Asthana, Cornelia Fermuller, Yiannis Aloimonos, IEEE International Conference on Robotics and Automation (ICRA) , 2021.

     Paper    Project Page    Code    UMD    Cite  

    PRGFlow: Benchmarking SWAP-Aware Unified Deep Visual Inertial Odometry

    T-RO 2020

    Nitin J. Sanket, Chahat Deep Singh, Cornelia Fermuller, Yiannis Aloimonos, IEEE Transactions on Robotics (Under Review), 2020.

     Paper    Project Page    Code    UMD    Cite  

    0-MMS: Zero-Shot Multi-Motion Segmentation With A Monocular Event Camera

    ICRA 2021

    Chethan M. Parameshwara, Nitin J. Sanket, Chahat Deep Singh, Cornelia Fermuller, Yiannis Aloimonos, IEEE International Conference on Robotics and Automation (ICRA), 2021.

     Paper    Project Page    UMD    Cite  

    EVDodgeNet: Deep Dynamic Obstacle Dodging with Event Cameras

    ICRA 2020

    Nitin J. Sanket*, Chethan M. Parameshwara*, Chahat Deep Singh, Cornelia Fermuller, Davide Scaramuzza, Yiannis Aloimonos, IEEE International Confernce on Robotics and Automation, Paris, 2020.
    * Equal Contribution

     Paper    Project Page    Code    UMD    Cite  

    GapFlyt: Active Vision Based Minimalist Structure-less Gap Detection For Quadrotor Flight

    RA-L 2018 | IROS 2018

    Chahat Deep Singh*, Nitin J. Sanket*, Kanishka Ganguly, Cornelia Fermuller, Yiannis Aloimonos, IEEE Robotics and Automation Letters, 2018.
    * Equal Contribution

    Project Page  Paper  Video  Cite 

    Video Tutorials

    Portfolio

    My Works

    • All
    • App
    • Card
    • Web

    App 1

    App

    Web 3

    Web

    App 2

    App

    Card 2

    Card

    Web 2

    Web

    App 3

    App

    Card 1

    Card

    Card 3

    Card

    Web 3

    Web

    Services

    Peer Review


    • RA-L IEEE Robotics and Automation Letters
    • ICRA International Conference on Robotics and Automation
    • IROS International Conference on Intelligent Robots and Systems
    • CVPR Conference on Computer Vision and Pattern Recognition
    • ICCV International Conference on Computer Vision
    • RSS Robotics Science and Systems
    • TVCJ The Visual Computer Journal, Springer

    Contact

    Contact Me

    My Address

    A108 Adam Street, New York, NY 535022

    Social Profiles

    Email Me

    contact@example.com

    Call Me

    +1 5589 55488 55

    Loading
    Your message has been sent. Thank you!