<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Teleoperation on Fondsites</title><link>https://fondsites.com/tags/teleoperation/</link><description>Recent content in Teleoperation on Fondsites</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Mon, 11 May 2026 11:34:07 +0300</lastBuildDate><atom:link href="https://fondsites.com/tags/teleoperation/feed.xml" rel="self" type="application/rss+xml"/><item><title>Robot Teleoperation: The Human Still in the Loop</title><link>https://fondsites.com/physical-ai-lab/guidebooks/robot-teleoperation/</link><pubDate>Sun, 10 May 2026 00:00:00 +0000</pubDate><guid>https://fondsites.com/physical-ai-lab/guidebooks/robot-teleoperation/</guid><description>&lt;p&gt;Robot videos often imply a clean story: the robot sees the world, decides what to do, and acts. Real deployments are usually messier and more interesting. Many useful robots are not fully independent actors. They are supervised systems with people nearby, people online, or people available when the robot reaches the edge of its competence. Teleoperation is the name for the human side of that story.&lt;/p&gt;
&lt;p&gt;&lt;img
 src="https://fondsites.com/physical-ai-lab/images/guidebooks/robot-teleoperation-station.avif"
 alt="A robot teleoperation and supervision station with a wheeled robot in a lab, abstract camera feeds, joystick, emergency stop, blank checklist cards, and safety cones"
 loading="lazy"
 decoding="async"&gt;
&lt;/p&gt;</description></item><item><title>Robot Data Collection: How Physical AI Learns From Work</title><link>https://fondsites.com/physical-ai-lab/guidebooks/robot-data-collection/</link><pubDate>Mon, 11 May 2026 08:13:55 +0300</pubDate><guid>https://fondsites.com/physical-ai-lab/guidebooks/robot-data-collection/</guid><description>&lt;p&gt;&lt;img
 src="https://fondsites.com/physical-ai-lab/images/guidebooks/robot-data-collection.avif"
 alt="A robotics lab data collection station with a robot arm, mobile robot, sensors, test objects, logging workstation, calibration props, and safety tape"
 loading="lazy"
 decoding="async"&gt;
&lt;/p&gt;
&lt;p&gt;The robot&amp;rsquo;s most valuable part may not be the arm, the battery, the camera, or the model. It may be the record of what happened when the machine tried to do real work.&lt;/p&gt;
&lt;p&gt;Physical AI learns from contact with a world that does not simplify itself for software. The cup slips because the gripper pads are dusty. The cardboard box bows under pressure. The mobile base hesitates near a glass wall. A human blocks the aisle for a moment, then moves away. A cable appears on the floor after lunch. Each moment can become useful data, but only if the system records enough context to explain what happened.&lt;/p&gt;</description></item></channel></rss>