MIT News - Computer Science and Artificial Intelligence Laboratory 09月12日
机器人通过内部传感器感知物体
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

研究人员开发了一种技术,使机器人仅通过内部传感器就能感知物体的重量、软度或内容物。通过拿起和轻轻摇晃物体,机器人可以在几秒钟内准确地猜测物体的质量等参数。这种方法无需外部测量工具或摄像头,特别适用于摄像头效果不佳的应用场景,如黑暗地下室中的物体分类或地震后部分坍塌建筑的废墟清理。该技术利用了机器人的本体感觉,通过关节编码器收集信号,并结合模拟机器人及其运动的模型和物体动力学模型,快速识别物体的特性。

📦 机器人通过内部传感器感知物体:研究人员开发了一种技术,使机器人仅通过内部传感器就能感知物体的重量、软度或内容物。通过拿起和轻轻摇晃物体,机器人可以在几秒钟内准确地猜测物体的质量等参数。这种方法无需外部测量工具或摄像头,特别适用于摄像头效果不佳的应用场景,如黑暗地下室中的物体分类或地震后部分坍塌建筑的废墟清理。

🔗 本体感觉的应用:该技术利用了机器人的本体感觉,通过关节编码器收集信号,来监测机器人的运动。例如,当机器人抬起一个哑铃时,它可以感知到哑铃的重量,即使哑铃在它的手中。同样地,机器人可以通过其手臂中的多个关节“感觉”到物体的重量。

🌐 模拟过程的关键作用:该技术的关键在于模拟过程,它结合了机器人和物体的模型,以快速识别机器人与物体交互时物体的特性。通过模拟机器人及其运动的模型和物体动力学模型,算法可以预测物体属性(如质量或软度)的微小变化如何影响机器人的最终关节位置。

🔍 数据高效的算法:与依赖计算机视觉或外部传感器的方法不同,该算法不需要大量的训练数据集,因此在面对未知环境或新物体时不易失败。这种数据高效的算法可以在几秒钟内完成计算,并且只需要一个真实的机器人运动轨迹即可进行计算。

🚀 未来应用前景:研究人员希望将该方法与计算机视觉结合,创建更强大的多模态传感技术。他们还希望探索更复杂的机器人系统(如软机器人)和更复杂的物体(如晃动的液体或沙粒)的应用,以改进机器人的学习能力,使未来的机器人能够快速开发新的操作技能并适应环境变化。

A human clearing junk out of an attic can often guess the contents of a box simply by picking it up and giving it a shake, without the need to see what’s inside. Researchers from MIT, Amazon Robotics, and the University of British Columbia have taught robots to do something similar.

They developed a technique that enables robots to use only internal sensors to learn about an object’s weight, softness, or contents by picking it up and gently shaking it. With their method, which does not require external measurement tools or cameras, the robot can accurately guess parameters like an object’s mass in a matter of seconds.

This low-cost technique could be especially useful in applications where cameras might be less effective, such as sorting objects in a dark basement or clearing rubble inside a building that partially collapsed after an earthquake.

Key to their approach is a simulation process that incorporates models of the robot and the object to rapidly identify characteristics of that object as the robot interacts with it. 

The researchers’ technique is as good at guessing an object’s mass as some more complex and expensive methods that incorporate computer vision. In addition, their data-efficient approach is robust enough to handle many types of unseen scenarios.

“This idea is general, and I believe we are just scratching the surface of what a robot can learn in this way. My dream would be to have robots go out into the world, touch things and move things in their environments, and figure out the properties of everything they interact with on their own,” says Peter Yichen Chen, an MIT postdoc and lead author of a paper on this technique.

His coauthors include fellow MIT postdoc Chao Liu; Pingchuan Ma PhD ’25; Jack Eastman MEng ’24; Dylan Randle and Yuri Ivanov of Amazon Robotics; MIT professors of electrical engineering and computer science Daniela Rus, who leads MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL); and Wojciech Matusik, who leads the Computational Design and Fabrication Group within CSAIL. The research will be presented at the International Conference on Robotics and Automation.

Sensing signals

The researchers’ method leverages proprioception, which is a human or robot’s ability to sense its movement or position in space.

For instance, a human who lifts a dumbbell at the gym can sense the weight of that dumbbell in their wrist and bicep, even though they are holding the dumbbell in their hand. In the same way, a robot can “feel” the heaviness of an object through the multiple joints in its arm.

“A human doesn’t have super-accurate measurements of the joint angles in our fingers or the precise amount of torque we are applying to an object, but a robot does. We take advantage of these abilities,” Liu says.

As the robot lifts an object, the researchers’ system gathers signals from the robot’s joint encoders, which are sensors that detect the rotational position and speed of its joints during movement. 

Most robots have joint encoders within the motors that drive their moveable parts, Liu adds. This makes their technique more cost-effective than some approaches because it doesn’t need extra components like tactile sensors or vision-tracking systems.

To estimate an object’s properties during robot-object interactions, their system relies on two models: one that simulates the robot and its motion and one that simulates the dynamics of the object.

“Having an accurate digital twin of the real-world is really important for the success of our method,” Chen adds.

Their algorithm “watches” the robot and object move during a physical interaction and uses joint encoder data to work backward and identify the properties of the object.

For instance, a heavier object will move slower than a light one if the robot applies the same amount of force.

Differentiable simulations

They utilize a technique called differentiable simulation, which allows the algorithm to predict how small changes in an object’s properties, like mass or softness, impact the robot’s ending joint position. The researchers built their simulations using NVIDIA’s Warp library, an open-source developer tool that supports differentiable simulations.

Once the differentiable simulation matches up with the robot’s real movements, the system has identified the correct property. The algorithm can do this in a matter of seconds and only needs to see one real-world trajectory of the robot in motion to perform the calculations.

“Technically, as long as you know the model of the object and how the robot can apply force to that object, you should be able to figure out the parameter you want to identify,” Liu says.

The researchers used their method to learn the mass and softness of an object, but their technique could also determine properties like moment of inertia or the viscosity of a fluid inside a container.

Plus, because their algorithm does not need an extensive dataset for training like some methods that rely on computer vision or external sensors, it would not be as susceptible to failure when faced with unseen environments or new objects.

In the future, the researchers want to try combining their method with computer vision to create a multimodal sensing technique that is even more powerful.

“This work is not trying to replace computer vision. Both methods have their pros and cons. But here we have shown that without a camera we can already figure out some of these properties,” Chen says.

They also want to explore applications with more complicated robotic systems, like soft robots, and more complex objects, including sloshing liquids or granular media like sand.

In the long run, they hope to apply this technique to improve robot learning, enabling future robots to quickly develop new manipulation skills and adapt to changes in their environments.

“Determining the physical properties of objects from data has long been a challenge in robotics, particularly when only limited or noisy measurements are available. This work is significant because it shows that robots can accurately infer properties like mass and softness using only their internal joint sensors, without relying on external cameras or specialized measurement tools,” says Miles Macklin, senior director of simulation technology at NVIDIA, who was not involved with this research.

This work is funded, in part, by Amazon and the GIST-CSAIL Research Program.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

机器人 内部传感器 本体感觉 计算机视觉 模拟过程
相关文章