index 62df3d1..53f1531 100644
@@ -14,4 +14,4 @@ an alexa with arms — a voice-controlled robotic arm that handles physical task
the design space is more constrained than it sounds. modern hobby robotics (ROS, cheap servo arms, depth cameras) has made a basic manipulation platform buildable for a few hundred dollars. the hard part is the manipulation policy: getting reliable grasping on arbitrary household objects across lighting and position variation is an open ML research problem, but demos from RT-2, ACT, and similar work suggest it's becoming tractable. the simplest version would start with a fixed set of "known objects" the arm has been trained on, expanding the object vocabulary over time. voice interface via an LLM handles intent parsing, and a camera provides object detection and pose estimation.
-connects to [[sensor-capturer|sensor capturer]] as shared hardware-platform territory — both require embedded sensing and actuation. [[always-on-ai-assistant|always-on AI assistant]] is a natural pairing: if an AI is already listening and aware of context, attaching arms is the next layer. [[emg-bracelet|EMG bracelet]] and [[pupilometry-glasses|pupilometry glasses]] are adjacent wearable-control ideas. the deeper question this raises: what's the minimum manipulation capability that would make a home robot actually useful day-to-day? the answer is probably narrower than people assume.
\ No newline at end of file
+related: [[sensor-capturer|sensor capturer]], [[always-on-ai-assistant|always-on AI assistant]], [[emg-bracelet|EMG bracelet]], [[pupilometry-glasses|pupilometry glasses]]
\ No newline at end of file