{"id":2035,"date":"2025-06-04T15:32:16","date_gmt":"2025-06-04T07:32:16","guid":{"rendered":"http:\/\/1628.webi.svipwebs.com\/?p=2035"},"modified":"2025-06-04T16:00:02","modified_gmt":"2025-06-04T08:00:02","slug":"one-shot-learning-driven-autonomous-robotic-assembly-via-human-robot-symbiotic-interaction","status":"publish","type":"post","link":"https:\/\/www.rzautoassembly.com\/pl\/one-shot-learning-driven-autonomous-robotic-assembly-via-human-robot-symbiotic-interaction\/","title":{"rendered":"One-shot learning-driven autonomous robotic assembly via human-robot symbiotic interaction"},"content":{"rendered":"<div id=\"ez-toc-container\" class=\"ez-toc-v2_0_73 counter-hierarchy ez-toc-counter ez-toc-grey ez-toc-container-direction\">\n<div class=\"ez-toc-title-container\">\n<p class=\"ez-toc-title\" style=\"cursor:inherit\">Table of Contents<\/p>\n<span class=\"ez-toc-title-toggle\"><a href=\"#\" class=\"ez-toc-pull-right ez-toc-btn ez-toc-btn-xs ez-toc-btn-default ez-toc-toggle\" aria-label=\"Toggle Table of Content\"><span class=\"ez-toc-js-icon-con\"><span class=\"\"><span class=\"eztoc-hide\" style=\"display:none;\">Toggle<\/span><span class=\"ez-toc-icon-toggle-span\"><svg style=\"fill: #999;color:#999\" xmlns=\"http:\/\/www.w3.org\/2000\/svg\" class=\"list-377408\" width=\"20px\" height=\"20px\" viewbox=\"0 0 24 24\" fill=\"none\"><path d=\"M6 6H4v2h2V6zm14 0H8v2h12V6zM4 11h2v2H4v-2zm16 0H8v2h12v-2zM4 16h2v2H4v-2zm16 0H8v2h12v-2z\" fill=\"currentColor\"><\/path><\/svg><svg style=\"fill: #999;color:#999\" class=\"arrow-unsorted-368013\" xmlns=\"http:\/\/www.w3.org\/2000\/svg\" width=\"10px\" height=\"10px\" viewbox=\"0 0 24 24\" version=\"1.2\" baseprofile=\"tiny\"><path d=\"M18.2 9.3l-6.2-6.3-6.2 6.3c-.2.2-.3.4-.3.7s.1.5.3.7c.2.2.4.3.7.3h11c.3 0 .5-.1.7-.3.2-.2.3-.5.3-.7s-.1-.5-.3-.7zM5.8 14.7l6.2 6.3 6.2-6.3c.2-.2.3-.5.3-.7s-.1-.5-.3-.7c-.2-.2-.4-.3-.7-.3h-11c-.3 0-.5.1-.7.3-.2.2-.3.5-.3.7s.1.5.3.7z\"\/><\/svg><\/span><\/span><\/span><\/a><\/span><\/div>\n<nav><ul class='ez-toc-list ez-toc-list-level-1' ><li class='ez-toc-page-1 ez-toc-heading-level-1'><a class=\"ez-toc-link ez-toc-heading-1\" href=\"https:\/\/www.rzautoassembly.com\/pl\/one-shot-learning-driven-autonomous-robotic-assembly-via-human-robot-symbiotic-interaction\/#One-shot_learning-driven_autonomous_robotic_assembly_via_human-robot_symbiotic_interaction\" title=\"One-shot learning-driven autonomous robotic assembly via human-robot symbiotic interaction\">One-shot learning-driven autonomous robotic assembly via human-robot symbiotic interaction<\/a><ul class='ez-toc-list-level-3' ><li class='ez-toc-heading-level-3'><ul class='ez-toc-list-level-3' ><li class='ez-toc-heading-level-3'><a class=\"ez-toc-link ez-toc-heading-2\" href=\"https:\/\/www.rzautoassembly.com\/pl\/one-shot-learning-driven-autonomous-robotic-assembly-via-human-robot-symbiotic-interaction\/#Abstract\" title=\"Abstract\">Abstract<\/a><\/li><li class='ez-toc-page-1 ez-toc-heading-level-3'><a class=\"ez-toc-link ez-toc-heading-3\" href=\"https:\/\/www.rzautoassembly.com\/pl\/one-shot-learning-driven-autonomous-robotic-assembly-via-human-robot-symbiotic-interaction\/#Introduction\" title=\"Introduction\">Introduction<\/a><\/li><li class='ez-toc-page-1 ez-toc-heading-level-3'><a class=\"ez-toc-link ez-toc-heading-4\" href=\"https:\/\/www.rzautoassembly.com\/pl\/one-shot-learning-driven-autonomous-robotic-assembly-via-human-robot-symbiotic-interaction\/#Conclusion\" title=\"Conclusion\">Conclusion<\/a><\/li><\/ul><\/li><\/ul><\/li><\/ul><\/nav><\/div>\n<h1 style=\"text-align: center;\"><span class=\"ez-toc-section\" id=\"One-shot_learning-driven_autonomous_robotic_assembly_via_human-robot_symbiotic_interaction\"><\/span><span style=\"font-family: 'times new roman', times, serif;\"><strong><b>One-shot learning-driven autonomous robotic assembly via human-robot symbiotic interaction<\/b><\/strong><\/span><span class=\"ez-toc-section-end\"><\/span><\/h1>\n<p><img fetchpriority=\"high\" decoding=\"async\" class=\"size-medium wp-image-2084 aligncenter\" src=\"https:\/\/www.rzautoassembly.com\/wp-content\/smush-webp\/2025\/06\/HMS-\u7f51\u7edc-2025-\u5e74\u5ea6\u62a5\u544a-211-300x287.png.webp\" alt=\"\" width=\"300\" height=\"287\" srcset=\"https:\/\/www.rzautoassembly.com\/wp-content\/smush-webp\/2025\/06\/HMS-\u7f51\u7edc-2025-\u5e74\u5ea6\u62a5\u544a-211-300x287.png.webp 300w, https:\/\/www.rzautoassembly.com\/wp-content\/smush-webp\/2025\/06\/HMS-\u7f51\u7edc-2025-\u5e74\u5ea6\u62a5\u544a-211-1024x981.png.webp 1024w, https:\/\/www.rzautoassembly.com\/wp-content\/smush-webp\/2025\/06\/HMS-\u7f51\u7edc-2025-\u5e74\u5ea6\u62a5\u544a-211-768x736.png.webp 768w, https:\/\/www.rzautoassembly.com\/wp-content\/smush-webp\/2025\/06\/HMS-\u7f51\u7edc-2025-\u5e74\u5ea6\u62a5\u544a-211.png.webp 1536w\" sizes=\"(max-width: 300px) 100vw, 300px\" \/><\/p>\n<h3><span class=\"ez-toc-section\" id=\"Abstract\"><\/span><strong><b>Abstract<\/b><\/strong><span class=\"ez-toc-section-end\"><\/span><\/h3>\n<p>In the realm of multi-procedure robotic assembly, where robots must sequentially assemble components, traditional programming proves labor-intensive, and end-to-end learning struggles with vast task spaces. This paper introduces a one-shot learning-from-demonstration (LfD) framework leveraging third-person visual observations to minimize human intervention and enhance adaptability. We propose an\u00a0<strong><b>object-centric representation<\/b><\/strong>\u00a0to preprocess human assembly demonstrations via RGB-D cameras, a\u00a0<strong><b>kinetic energy-based changepoint detection algorithm<\/b><\/strong>\u00a0to segment procedures and decode human intent, and a\u00a0<strong><b>demo-trajectory adaptation-enhanced dynamical movement primitive (DA-DMP)<\/b><\/strong>\u00a0to improve motion skill generalization. Validated on a robotic assembly platform, the system achieves accurate sequence learning from a single demonstration, efficient motion planning, and a 93.3% success rate, advancing trustworthy human\u2013machine symbiotic manufacturing aligned with human-centered automation.<\/p>\n<h3><span class=\"ez-toc-section\" id=\"Introduction\"><\/span><strong><b>Introduction<\/b><\/strong><span class=\"ez-toc-section-end\"><\/span><\/h3>\n<p>Amid Industry 5.0\u2019s human-centric manufacturing paradigm, industrial robots must adapt intuitively to dynamic environments, particularly in assembly tasks that account for 50% of manufacturing time and 30% of costs. While manual programming offers reliability, it relies on expert knowledge, and end-to-end learning demands extensive data and training. Learning from Demonstration (LfD) bridges this gap, enabling skill transfer via imitation. Among LfD paradigms,\u00a0<strong><b>passive visual observation<\/b><\/strong>\u00a0stands out for multi-procedure tasks, requiring no physical guidance or teleoperation\u2014ideal for natural human\u2013robot symbiosis.<\/p>\n<p>This paper presents a one-shot LfD approach for long-horizon assembly tasks using third-person visual data. By extracting object-centric representations, automating procedure segmentation, and enhancing motion primitives, the framework empowers robots to learn complex assembly sequences from a single demonstration. Unlike prior work requiring multiple trials or hybrid teaching methods, our approach minimizes human effort and facilitates rapid deployment in real-world manufacturing, as validated in a seven-part shaft-gear assembly case study.<\/p>\n<h3><span class=\"ez-toc-section\" id=\"Conclusion\"><\/span><strong><b>Conclusion<\/b><\/strong><span class=\"ez-toc-section-end\"><\/span><\/h3>\n<p>The proposed one-shot LfD framework redefines multi-procedure robotic assembly by integrating visual perception, automatic task segmentation, and adaptive motion planning. By enabling robots to learn from a single human demonstration, the system reduces programming overhead, enhances adaptability to new scenarios, and achieves a 93.3% success rate in real-world testing. This advances human\u2013machine symbiosis in manufacturing, aligning with Industry 5.0\u2019s vision of intuitive, efficient production.<\/p>\n<p><strong><b>For broader industrial applicability<\/b><\/strong>, the framework\u2019s object-centric design and adaptive motion primitives could extend to\u00a0<strong><b>bathroom components assembly machines<\/b><\/strong>, where multi-procedure tasks (e.g., installing faucets, aligning pipes, and securing fixtures) require precise, sequential operations. By applying our one-shot LfD approach, such machines could rapidly learn to assemble diverse bathroom components\u2014from sinks to shower systems\u2014with minimal human intervention, enhancing flexibility in customized manufacturing while maintaining high precision and throughput. This underscores the methodology\u2019s potential to transform not only automotive or mechanical assembly but also specialized industries requiring intricate, multi-step robotic manipulation.<\/p>","protected":false},"excerpt":{"rendered":"<p>One-shot learning-driven autonomous robotic assembly via human-robot symbiotic interaction Abstract In the realm of multi-procedure robotic assembly, where robots must sequentially assemble components, traditional programming proves labor-intensive, and end-to-end learning struggles with vast task spaces. This paper introduces a one-shot learning-from-demonstration (LfD) framework leveraging third-person visual observations to minimize human intervention and enhance adaptability. We [\u2026]<\/p>","protected":false},"author":2,"featured_media":2083,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[1,124],"tags":[],"class_list":["post-2035","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-news","category-technology"],"acf":[],"_links":{"self":[{"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/posts\/2035","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/comments?post=2035"}],"version-history":[{"count":0,"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/posts\/2035\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/media\/2083"}],"wp:attachment":[{"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/media?parent=2035"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/categories?post=2035"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.rzautoassembly.com\/pl\/wp-json\/wp\/v2\/tags?post=2035"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}