- Robot perception: seeing and understanding the physical world by multimodal robot sensors
- Robot vision vs. Computer vision
- Robot vision is embodied, active and environmentally situated.
- Embodied: Robots have physical bodies and experience the world directly. Their actions are part of a dynamic with the world and have immediate feedback on their own sensation.
- Active: Robots are active perceivers. It knows why it wishes to sense, and chooses what to perceive, and determines how, when and where to achieve that perception.
- Situated: Robots are situated in the world. They do not deal with abstract descriptions, but with the here and now of the world directly influencing the behavior of the system.
- Robot vision vs. Computer vision
- Modalities: neural network architectures designed for different sensory modalities
- Pixels (from RGB camera)
- Point cloud (from structure sensors)
- Time series (from F/T sensors)
- Tactile data (from the GelSights sensors)
- Representations: representation learning algorithms without strong supervision.
- A fundamental problem in robot perception is to learn the proper representations of the unstructured world.
- 밑 그림 참조
- Learn representations of the world with limited supervision: self-supervised learning. Supervision comes from the unlabeled data themselves
- Learn representations that fuse multiple sensory modalities together: Multimodal sensor fusion
- Tasks: state estimation tasks for robot navigation and manipulation
- State estimation methods: Bayes filtering
- What if models are hard to specify? Learning -> Embodied view of perception.
- Frontiers: embodied visual learning & synthetic data for visualAI
'Robotics & Perception > Basic' 카테고리의 다른 글
Imitation learning (0) | 2022.07.08 |
---|---|
[CS391R] Overview of Robot Decision Making (0) | 2022.05.15 |
[CS391R] Introduction of Robot Learning (0) | 2022.05.14 |
[Modern Robotics] Contents (0) | 2022.04.02 |
[Graphics] Obj (0) | 2022.02.02 |