diff --git a/_talks/bohg_23.md b/_talks/bohg_23.md index 54d9de0..1a735b3 100644 --- a/_talks/bohg_23.md +++ b/_talks/bohg_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Large Language Models for Solving Long-Horizon Manipulation Problems" abstract: "My long-term research goal is enable real robots to manipulate any kind of object such that they can perform many different tasks in a wide variety of application scenarios such as in our homes, in hospitals, warehouses, or factories. Many of these tasks will require long-horizon reasoning and sequencing of skills to achieve a goal state. In this talk, I will present our work on enabling long-horizon reasoning on real robots for a variety of different long-horizon tasks that can be solved by sequencing a large variety of composable skill primitives. I will specifically focus on the different ways Large Language Models (LLMs) can help with solving these long-horizon tasks. The first part of my talk will be on TidyBot, a robot for personalised household clean-up. One of the key challenges in robotic household cleanup is deciding where each item goes. People's preferences can vary greatly depending on personal taste or cultural background. One person might want shirts in the drawer, another might want them on the shelf. How can we infer these user preferences from only a handful of examples in a generalizable way? Our key insight: Summarization with LLMs is an effective way to achieve generalization in robotics. Given the generalised rules, I will then show how TidyBot then solves the long-horizon task of cleaning up a home. In the second part of my talk, I will focus on more complex long-horizon manipulation tasks that exhibit geometric dependencies between different skills in a sequence. In these tasks, the way a robot performs a certain skill will determine whether a follow-up skill in the sequence can be executed at all. I will present an approach called text2motion that utilises LLMs for task planning without the need for defining complex symbolic domains. And I will show how we can verify whether the plan that the LLM came up with is actually feasible. The basis for this verification is a library of learned skills and an approach for sequencing these skills to resolve geometric dependencies prevalent in long-horizon tasks." -youtube-code: "TBD" +# youtube-code: "TBD" --- diff --git a/_talks/chaudhari_23.md b/_talks/chaudhari_23.md index b225820..4b9f6d4 100644 --- a/_talks/chaudhari_23.md +++ b/_talks/chaudhari_23.md @@ -14,5 +14,5 @@ Does the data induce capacity control in deep learning? Rubing Yang, Jialin Mao, Deep Reference Priors: What is the best way to pretrain a model? Yansong Gao, Rahul Ramesh, and Pratik Chaudhari. [ICML '22] https://arxiv.org/abs/2202.00187 A picture of the space of typical learnable tasks. Rahul Ramesh, Jialin Mao, Itay Griniasty, Rubing Yang, Han Kheng Teoh, Mark Transtrum, James P. Sethna, and Pratik Chaudhari. [ICML ’23]. https://arxiv.org/abs/2210.17011 The Training Process of Many Deep Networks Explores the Same Low-Dimensional Manifold. Jialin Mao, Itay Griniasty, Han Kheng Teoh, Rahul Ramesh, Rubing Yang, Mark K. Transtrum, James P. Sethna, Pratik Chaudhari. 2023. https://arxiv.org/abs/2305.01604" -youtube-code: "TBD" +youtube-code: "ZD2cL-QoI5g" --- diff --git a/_talks/fidler_23.md b/_talks/fidler_23.md index 5b1c90a..0324306 100644 --- a/_talks/fidler_23.md +++ b/_talks/fidler_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "A.I. for 3D Content Creation" abstract: "3D content is key in several domains such as architecture, film, gaming, robotics, and lies in the heart of the metaverse applications. However, creating 3D content can be very time consuming -- the artists need to sculpt high quality 3d assets, compose them into large worlds, and bring these worlds to life by writing behaviour models that drive the agents around in the world. In this talk, I'll present some of our ongoing efforts on creating virtual worlds with A.I., with the focus on street level simulation for autonomous driving." -youtube-code: "TBD" +# youtube-code: "TBD" --- diff --git a/_talks/ichter_23.md b/_talks/ichter_23.md index 6c8345e..40181c2 100644 --- a/_talks/ichter_23.md +++ b/_talks/ichter_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Connecting Robotics and Foundation Models" abstract: "Foundation models can encode a wealth of semantic knowledge about the world, but can be limited by their lack of interactive, real-world experience. This poses a challenge for leveraging them in robotics, which requires interactive decision making and reasoning for a given embodiment. This talk will discuss several research directions towards addressing these challenges, from grounding them in their environment (SayCan, InnerMonologue, Grounded Decoding, NL-Maps), to directly outputting grounded code (Code as Policies), and finally training them with embodied robotics data (PaLM-E, RT-1)." -youtube-code: "TBD" +youtube-code: "wnPNRh-jtZM" --- diff --git a/_talks/morozovsky_23.md b/_talks/morozovsky_23.md index c4850a2..1cf3927 100644 --- a/_talks/morozovsky_23.md +++ b/_talks/morozovsky_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Meet Astro: Amazon Consumer Robotics’ first robot for homes and small-to-medium businesses" abstract: "Astro is Amazon’s first robot designed for homes and small-to-medium businesses. In this talk, we will show Astro’s use cases, including home monitoring, Human Robot Interaction, and Virtual Security Guard, and share how we designed Astro for customers. Making the first prototype work is a very different problem from designing and testing robots for mass production. Design for manufacturability, repairability, and sustainability are important tenets for making robots that are easy to assemble, test, and fix. Reliability is another critical concern for mass-produced robots, we’ll describe some of the some of the extensive testing we performed to make sure that every Astro robot works for customers for years. The Mobility and Perception capabilities of Astro are what make it useful and capable in unstructured and variable environments like homes. Our software team overcame challenges to deliver these capabilities with consumer-grade sensors and compute. We’ll conclude with some of Amazon’s programs to engage with the academic community. Note: No youtube video" -youtube-code: "TBD" +# youtube-code: "TBD" --- diff --git a/_talks/russo_23.md b/_talks/russo_23.md index b6371b5..13113d4 100644 --- a/_talks/russo_23.md +++ b/_talks/russo_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Soft Material Robotics and Next-Generation Surgical Robots" abstract: "Minimally invasive surgical (MIS) procedures pose significant challenges for robots, which need to safely navigate through and manipulate delicate anatomy while performing complex tasks to treat tumors in remote areas. Soft robots hold considerable potential in MIS given their compliant nature, inherent safety, and high dexterity. Yet, a significant breakthrough of soft robots in surgery is impeded by current limitations in the design, manufacturing, and integration of soft materials that combine actuation, sensing, and control. Scientific understanding of medical and surgical robotics is entering an exciting new era where early approaches relying on rigid materials, standard manufacturing, and conventional kinematics are giving way to Soft Material Robotics. Our research at the Material Robotics Lab at Boston University is focused on the design, mechanics, and manufacturing of novel multi-scale and multi-material biomedical robotic systems. This talk will illustrate our work towards achieving safe navigation, distal actuation, integrated sensing, and effective force transmission in MIS by highlighting different classes of soft surgical robots, i.e., soft continuum robots, soft-foldable robots, and soft reactive skins with applications in lung cancer, colorectal cancer, and brain cancer surgery." -youtube-code: "TBD" +youtube-code: "iw629hk4EV8" --- diff --git a/_talks/student_bobu_spring_23.md b/_talks/student_bobu_spring_23.md index 13f9e8e..54cad41 100644 --- a/_talks/student_bobu_spring_23.md +++ b/_talks/student_bobu_spring_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Aligning Robot and Human Representations" abstract: "To perform tasks that humans want in the world, robots rely on a representation of salient task features; for example, to hand me a cup of coffee, the robot considers features like efficiency and cup orientation in its behavior. Prior methods try to learn both a representation and a downstream task jointly from data sets of human behavior, but this unfortunately picks up on spurious correlations and results in behaviors that do not generalize. In my view, what’s holding us back from successful human-robot interaction is that human and robot representations are often misaligned: for example, our lab’s assistive robot moved a cup inches away from my face -- which is technically collision-free behavior -- because it lacked an understanding of personal space. Instead of treating people as static data sources, my key insight is that robots must engage with humans in an interactive process for finding a shared representation for more efficient, transparent, and seamless downstream learning. In this talk, I focus on a divide and conquer approach: explicitly focus human input on teaching robots good representations before using them for learning downstream tasks. This means that instead of relying on inputs designed to teach the representation implicitly, we have the opportunity to design human input that is explicitly targeted at teaching the representation and can do so efficiently. I introduce a new type of representation-specific input that lets the human teach new features, I enable robots to reason about the uncertainty in their current representation and automatically detect misalignment, and I propose a novel human behavior model to learn robust behaviors on top of human-aligned representations. By explicitly tackling representation alignment, I believe we can ultimately achieve seamless interaction with humans where each agent truly grasps why the other behaves the way they do." -youtube-code: "TBD" +youtube-code: "GnbGDe2ZuZ0" --- diff --git a/_talks/student_richards_spring_23.md b/_talks/student_richards_spring_23.md index 5899b0d..0c55d85 100644 --- a/_talks/student_richards_spring_23.md +++ b/_talks/student_richards_spring_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Control-Oriented Learning for Dynamical Systems" abstract: "Robots are inherently nonlinear dynamical systems, for which synthesizing a stabilizing feedback controller with a known system model is already a difficult task. When learning a nonlinear model and controller from data, naive regression can produce a closed-loop model that is poorly conditioned for stable operation over long time horizons. In this talk, I will present our work on control-oriented learning, wherein the model learning problem is augmented to be cognizant of the desire for a stable closed-loop system. I will discuss how principles from control theory inform such augmentation to produce performant closed-loop models in a data efficient manner. This will involve ideas from contraction theory, constrained optimization, structured learning, adaptive control, and meta-learning." -youtube-code: "TBD" +youtube-code: "h0KMxcEAKHs" --- diff --git a/_talks/wu_23.md b/_talks/wu_23.md index 248daff..ff42458 100644 --- a/_talks/wu_23.md +++ b/_talks/wu_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Intelligent Coordination for Sustainable Roadways – If Autonomous Vehicles are the Answer, then What is the Question?" abstract: "For all its hype, autonomous vehicles have yet to make our roadways more sustainable: safer, cheaper, cleaner. This talk suggests that key to unlocking sustainable roadways is to shift the focus from autonomy-driven design to use-driven design. Based on recent work, the talk focuses on three critical priorities––safety, cost, and environment––each leveraging the 'autonomy' capability of coordinating vehicles. But fully autonomous agents are not the only entities that can coordinate. A paragon of safety is air traffic control, in which expert operators remotely coordinate aircraft. The work brings these ideas to the dense traffic on roadways and analyzes the scalability of operators. Another much cheaper way to coordinate is to give a smartphone app to drivers. The work characterizes how well lower-tech systems can still achieve autonomous capabilities. For cleaner roadways, dozens of articles have considered coordinating vehicles to reduce emissions. This work models whether doing so would move the needle on climate change mitigation goals. To study these multi-agent coordination problems, the work leverages queueing theory, Lyapunov stability analysis, transfer learning, and multi-task reinforcement learning. The talk will also substantiate issues of robustness that arise when applying learning-based techniques and a new line of work designed to address them. Overall, the results indicate promise for intelligent coordination to enable sustainable roadways." -youtube-code: "TBD" +youtube-code: "6FyS2WWw3J0" --- diff --git a/_talks/zakhypov_23.md b/_talks/zakhypov_23.md index efb4821..9057dd1 100644 --- a/_talks/zakhypov_23.md +++ b/_talks/zakhypov_23.md @@ -7,5 +7,5 @@ location: Skilling Auditorium location-url: "https://campus-map.stanford.edu/?id=04-550&lat=37.42697371527761&lng=-122.17280664808126&zoom=18&srch=undefined" title: "Multimaterial Design for Multifunctional Miniature Robots" abstract: "Small-scale animals like trap-jaw ants exhibit remarkable behaviors, not just through communication, but also via their adaptable jaw-jump and leg-jump mechanisms that enable them to thrive in diverse environments. These creatures have successfully tackled the challenges of miniaturization, multifunctionality, and multiplicity, which are critical factors in the development of small-scale robotic systems. By creating these abilities in mesoscale robots, we can unlock a vast array of applications. For instance, we could build artificial multi-locomotion swarms to explore and monitor diverse physical environments with high task efficiency or design compact and distributed haptic actuators to simulate compelling human touch interactions in virtual environments with high fidelity and minimal encumbrance. However, conventional design methods for creating miniature yet multifunctional robots are limited due to constraints in downsizing classical electric motors, transmission gears, and mechanisms. Additionally, increasing the number of components requires meticulous manual assembly processes. In this talk, I will delve into how multimaterial layer composition and folding (origami robotics) and 3D printing can enable miniature, multifunctional, and mass-manufacturable robots. I will provide insights into a systematic design methodology that breaks down mesoscale robot design in terms of mechanisms, geometry, materials, and fabrication, highlighting their relation and challenges. I will demonstrate unique robotic platforms built on this paradigm, including Tribots, 10-gram palm-sized multi-locomotion origami robots that jump, roll, and crawl to traverse uneven terrains and manipulate objects collectively, as well as shape-morphing grippers and structures. These robots use functional materials like shape memory alloy and fluids to achieve tunable power, compact actuators, and mechanisms. Additionally, I will present my latest research on monolithically 3D-printed, soft finger and wrist-worn haptic displays called FingerPrint and Hoxels. FingerPrint produces 4-DoF motion on the finger pad and phalanges with tunable forces and torques for skin shear, pressure, and vibrotactile interaction and can be mass-printed requiring " -youtube-code: "TBD" +youtube-code: "NFGTqhcxVdo" ---