I work on using human experiences to improve semantic understanding of the environment for mobile robots
(Rosa et al., 2018),(Rosa et al., 2018)
.
I study interpretable ways to learn sensor fusion strategies in deep VIO frameworks, for integrating novel sensor modalities such as 
millimeter wave radar and thermal imaging into a unified framework for first responders
(Chen et al., 2022), (Saputra et al., 2020), (Lu et al., 2020), (Chen et al., 2019)
.
  
  
    References
    
      2022
- 
  
    
  
  
  
    
    Learning selective sensor fusion for state estimation 
      
      
      Changhao
            Chen, Stefano
            Rosa, Chris Xiaoxuan
            Lu, and
        3 more authors
      
      
     
      IEEE Transactions on Neural Networks and Learning Systems,  2022
     
      
     
        we propose an end-to-end selective sensor fusion module that can be applied to modality pairs, such as monocular images and inertial measurements, depth images, and light detection and ranging (LIDAR) point clouds. 
 
 
 
2020
- 
  
    
  
  
  
    
    Deeptio: A deep thermal-inertial odometry with visual hallucination 
      
      
      Muhamad Risqi U
            Saputra, Pedro PB
            De Gusmao, Chris Xiaoxuan
            Lu, and
        7 more authors
      
      
     
      IEEE Robotics and Automation Letters,  2020
     
      
     
        We learn to hallucinate visual features from thermal images that can help first responders to navigate visually-denied scenarios. 
 
 
 
- 
  
    
  
  
  
    
    See through smoke: robust indoor mapping with low-cost mmwave radar 
      
      
      Chris Xiaoxuan
            Lu, Stefano
            Rosa, Peijun
            Zhao, and
        5 more authors
      
      
     
      In Proceedings of the 18th International Conference on Mobile Systems, Applications, and Services,  2020
     
      
     
        We show how to build dense occupancy grid maps of indoor environments from sparse, noisy mmWave measurements, with cross-modal training. 
 
 
 
2019
- 
  
    
  
  
  
    
    Selective sensor fusion for neural visual-inertial odometry 
      
      
      Changhao
            Chen, Stefano
            Rosa, Yishu
            Miao, and
        4 more authors
      
      
     
      In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition,  2019
     
      
     
        We show how data-learned sensor fusion strategies can improve accuracy and robustness in deep VIO when dealing with noisy/corrupted data, while adding interpretability. 
 
 
 
2018
- 
  
    
  
  
  
    
    Semantic place understanding for human–robot coexistence—toward intelligent workplaces 
      
      
      Stefano
            Rosa, Andrea
            Patane, Chris Xiaoxuan
            Lu, and
        1 more author
      
      
     
      IEEE Transactions on Human-Machine Systems,  2018
     
      
     
        Robots and users can work synergistically by mutually learning over time, and benefitting from each other by exploiting each other’s strengths. We show how detecting user activities can help robots to learn semantic understanding of the environment, while at the same time learning to better localise the user. 
 
 
 
- 
  
    
  
  
  
    
    CommonSense: Collaborative learning of scene semantics by robots and humans 
      
      
      Stefano
            Rosa, Andrea
            Patanè, Xiaoxuan
            Lu, and
        1 more author
      
      
     
      In Proceedings of the 1st International Workshop on Internet of People, Assistive Robots and Things,  2018