<?xml version="1.0" encoding="utf-8" standalone="yes"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom">
  <channel>
    <title>Virtual-Fixtures on Markus Knauer</title>
    <link>http://localhost:1313/tags/virtual-fixtures/</link>
    <description>Recent content in Virtual-Fixtures on Markus Knauer</description>
    <generator>Hugo</generator>
    <language>en-us</language>
    <lastBuildDate>Fri, 24 Apr 2026 00:00:00 +0000</lastBuildDate>
    <atom:link href="http://localhost:1313/tags/virtual-fixtures/index.xml" rel="self" type="application/rss+xml" />
    <item>
      <title>**MOMO:** Seamless Physical, Verbal, and Graphical Robot Skill Adaptation</title>
      <link>http://localhost:1313/posts/momo-physical-verbal-graphical-skill-adaptation/</link>
      <pubDate>Fri, 24 Apr 2026 00:00:00 +0000</pubDate>
      <guid>http://localhost:1313/posts/momo-physical-verbal-graphical-skill-adaptation/</guid>
      <description>&lt;p&gt;What if non-expert operators could adapt an industrial robot&amp;rsquo;s skills as easily as touching, talking to, or clicking on it? Our framework &lt;strong&gt;MOMO&lt;/strong&gt; (&lt;strong&gt;Mo&lt;/strong&gt;tion &lt;strong&gt;Mo&lt;/strong&gt;dulation) unifies three complementary interaction modalities for seamless robot skill learning and adaptation.&lt;/p&gt;&#xA;&#xA;&#xA;&#xA;&#xA;&#xA;&#xA;&lt;picture class=&#34;&#34;&gt;&#xA;  &lt;source srcset=&#34;http://localhost:1313/images/momo_overview.webp&#34; type=&#34;image/webp&#34;&gt;&#xA;  &lt;img src=&#34;http://localhost:1313/images/momo_overview.png&#34; alt=&#34;Overview of the MOMO framework unifying physical, verbal, and graphical modalities for robot skill adaptation&#34; loading=&#34;lazy&#34;&gt;&#xA;&lt;/picture&gt;&#xA;&lt;p&gt;MOMO integrates five components around a central motion modulation module:&lt;/p&gt;&#xA;&lt;ul&gt;&#xA;&lt;li&gt;&lt;strong&gt;Kinesthetic touch&lt;/strong&gt; for precise spatial corrections, using energy-tank-based human intention detection that automatically inserts via-points into the underlying motion model.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Natural language&lt;/strong&gt; for high-level semantic modifications through a tool-based LLM architecture that selects and parameterizes pre-validated functions — never generating executable code.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;A graphical web interface&lt;/strong&gt; (Vue.js/Three.js) for visualizing geometric relations, inspecting parameters, and editing via-points via drag-and-drop on a real-time digital twin.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Probabilistic Virtual Fixtures&lt;/strong&gt; for guided demonstration recording.&lt;/li&gt;&#xA;&lt;li&gt;&lt;strong&gt;Ergodic control&lt;/strong&gt; for surface finishing tasks.&lt;/li&gt;&#xA;&lt;/ul&gt;&#xA;&lt;p&gt;A key result: the tool-based LLM architecture &lt;strong&gt;generalizes beyond KMP-based skills to ergodic control&lt;/strong&gt;, enabling the same chat interface to drive voice-commanded surface finishing. Users freely switch between modalities — voice for obstacle avoidance, kinesthetic for fine corrections, graphical for verification.&lt;/p&gt;</description>
    </item>
  </channel>
</rss>
