- ESP 32 ICONS AND ANIMATIONSPosted 11 hours ago
- Autonomous Fire Fighting Robot With Self Finding FlamesPosted 2 weeks ago
- Automatic tank filling at homePosted 2 weeks ago
- Automatic chess board designPosted 3 weeks ago
- Automated Hydroponic SystemPosted 3 weeks ago
- Arduino HangmanPosted 3 weeks ago
- All Band Receiver With Si4735 (FM / RDS, AM and SSB) With ArduinoPosted 3 weeks ago
- A Super Easy Security Camera With the ESP32 CAMPosted 3 weeks ago
- An ESP8266-based WiFi doorbell sends SMS messagesPosted 3 weeks ago
- 18650 Battery charger reverse polarity protectionPosted 4 weeks ago
Open Source Project Proposes Vision-Free Grasping With RFID and Touchscreens
Liatris is an open-source hardware and software project (led by roboticist Mark Silliman) that does away with vision completely. Instead, you can determine the identity and pose of slightly modified objects with just a touchscreen and an RFID reader. It’s simple, relatively inexpensive, and as long as you’re not trying to deal with anything new, it works impressively well.
To get around the perception problem, Liatris uses a few clever tricks. First, each object has an RFID tag attached to it with a unique identifier, so that the robot can wirelessly detect what it’s working with. Once the robot has scanned the RFID tag, it looks the identifier up in an open source, global database of objects and downloads a CAD model and a grasp pose that “defines the ideal pose for the gripper prior to grasping the object.”
So now that you know what the object is and how to grasp it, you just need to know exactly where it is and what orientation it’s in. You can’t get that sort of information very easily from an RFID tag, so this is where the touchscreen comes in: Each object is (slightly) modified with an isosceles triangle of conductive points on the base, giving the touchscreen an exact location for the object, as well as the orientation, courtesy the pointy end of the triangle. With this data, the robot can accurately visualize the CAD model of the object on the touchscreen, and as long as it knows exactly where the touchscreen is, it can then grasp the real object based solely on the model. The robot doesn’t have to “see” anything: you just need the touchscreen and a RFID reader, and a headless robot arm can grasp just about whatever you want it to.
Source: Open Source Project Proposes Vision-Free Grasping With RFID and Touchscreens – IEEE Spectrum
