<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Tactile Sensing on Fondsites</title><link>https://fondsites.com/tags/tactile-sensing/</link><description>Recent content in Tactile Sensing on Fondsites</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Thu, 07 May 2026 22:14:46 +0300</lastBuildDate><atom:link href="https://fondsites.com/tags/tactile-sensing/feed.xml" rel="self" type="application/rss+xml"/><item><title>Robot Hands and Dexterous Manipulation</title><link>https://fondsites.com/physical-ai-lab/guidebooks/robot-hands-and-manipulation/</link><pubDate>Thu, 07 May 2026 00:00:00 +0000</pubDate><guid>https://fondsites.com/physical-ai-lab/guidebooks/robot-hands-and-manipulation/</guid><description>&lt;p&gt;&lt;img
 src="https://fondsites.com/physical-ai-lab/images/guidebooks/robot-hands.avif"
 alt="A robot hand and several end-effectors arranged beside blocks, fabric, tools, tactile sensors, and force diagrams"
 loading="lazy"
 decoding="async"&gt;
&lt;/p&gt;
&lt;p&gt;Robot hands are where robotics stops looking like software and starts looking like the physical world fighting back.&lt;/p&gt;
&lt;p&gt;A text model can be wrong and produce a bad paragraph. A robot hand can be wrong and drop a glass, crush a tomato, miss a handle, tear a bag, jam a drawer, or push the object out of reach. Manipulation is hard because the robot has to perceive, touch, move, and adapt at the same time.&lt;/p&gt;</description></item></channel></rss>