Stanford Seminar - Get in Touch: Tactile Perception for Human-Robot Systems

Stanford Seminar - Get in Touch: Tactile Perception for Human-Robot Systems

Stanford Online via YouTube Direct link

Fluidic, multimodal tactile sensors

6 of 13

6 of 13

Fluidic, multimodal tactile sensors

Class Central Classrooms beta

YouTube videos curated by Class Central.

Classroom Contents

Stanford Seminar - Get in Touch: Tactile Perception for Human-Robot Systems

Automatically move to the next video in the Classroom when playback concludes

  1. 1 Introduction
  2. 2 Characterization of human strategies for grasp, manipulation, and haptic search
  3. 3 Biological mechanoreceptors
  4. 4 Microfluidic tactile sensor skins
  5. 5 Camera-based, elastomeric tactile sensors
  6. 6 Fluidic, multimodal tactile sensors
  7. 7 Haptic perception necessitates the abstraction of tactile sensor data generated by known actions.
  8. 8 Tactile perception of directionality
  9. 9 Tactile perception within granular media
  10. 10 Learning hard-to-code skills: Closing a ziplock bag
  11. 11 Preparations for reinforcement learning
  12. 12 Page flipping behaviors affect tactile sensor data
  13. 13 Exciting directions in robotics related to touch and social-physical interactions

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.