The robot helps to put a jacket on a human.
The robot helps to put a jacket on a human.
Source: MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL)
16.07.2021 •

Robot helps people with limited mobility get dressed

Scientists at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have developed an algorithm to help a robot find efficient motion plans to ensure physical safety of its human counterpart.

Basic safety needs in the paleolithic era have largely evolved with the onset of the industrial and cognitive revolutions. We interact a little less with raw materials, and interface a little more with machines.

Robots don’t have the same hardwired behavioral awareness and control, so secure collaboration with humans requires methodical planning and coordination. You can likely assume your friend can fill up your morning coffee cup without spilling on you, but for a robot, this seemingly simple task requires careful observation and comprehension of human behavior.

Scientists from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have created a new algorithm to help a robot find efficient motion plans to ensure physical safety of its human counterpart. In this case, the bot helped put a jacket on a human, which could potentially prove to be a powerful tool in expanding assistance for those with disabilities or limited mobility.

“Developing algorithms to prevent physical harm without unnecessarily impacting the task efficiency is a critical challenge,” says MIT PhD student Shen Li, a lead author on a new paper about the research. “By allowing robots to make non-harmful impact with humans, our method can find efficient robot trajectories to dress the human with a safety guarantee.”

Human modeling, safety, and efficiency

Proper human modeling -- how the human moves, reacts, and responds -- is necessary to enable successful robot motion planning in human-robot interactive tasks. A robot can achieve fluent interaction if the human model is perfect, but in many cases, there’s no flawless blueprint.

A robot shipped to a person at home, for example, would have a very narrow, “default” model of how a human could interact with it during an assisted dressing task. It wouldn’t account for the vast variability in human reactions, dependent on a myriad of variables such as personality and habits. A screaming toddler would react differently to putting on a coat or shirt than a frail elderly person, or those with disabilities who might have rapid fatigue or decreased dexterity.

If that robot is tasked with dressing, and plans a trajectory solely based on that default model, the robot could clumsily bump into the human, resulting in an uncomfortable experience or even possible injury. However, if it’s too conservative in ensuring safety, it might pessimistically assume that all space nearby is unsafe, and then fail to move, something known as the "freezing robot" problem.

To provide a theoretical guarantee of human safety, the team's algorithm reasons about the uncertainty in the human model. Instead of having a single, default model where the robot only understands one potential reaction, the team gave the machine an understanding of many possible models, to more closely mimic how a human can understand other humans. As the robot gathers more data, it will reduce uncertainty and refine those models.

To resolve the freezing robot problem, the team redefined safety for human-aware motion planners as either collision avoidance or safe impact in the event of a collision. Often, especially in robot-assisted tasks of activities of daily living, collisions cannot be fully avoided. This allowed the robot to make non-harmful contact with the human to make progress, so long as the robot's impact on the human is low. With this two-pronged definition of safety, the robot could safely complete the dressing task in a shorter period of time.

For example, let’s say there are two possible models of how a human could react to dressing. “Model One” is that the human will move up during dressing, and “Model Two” is that the human will move down during dressing. With the team’s algorithm, when the robot is planning its motion, instead of selecting one model, it will try to ensure safety for both models. No matter if the person is moving up or down, the trajectory found by the robot will be safe.

To paint a more holistic picture of these interactions, future efforts will focus on investigating the subjective feelings of safety in addition to the physical during the robot-assisted dressing task.

“This multifaceted approach combines set theory, human-aware safety constraints, human motion prediction, and feedback control for safe human-robot interaction,” says Zackory Erickson, Assistant Professor in The Robotics Institute at Carnegie Mellon University (Fall 2021). “This research could potentially be applied to a wide variety of assistive robotics scenarios, towards the ultimate goal of enabling robots to provide safer physical assistance to people with disabilities.”

Subscribe to our newsletter

Related articles

A contact aware robot design

A contact aware robot design

Researchers have developed a new method to computationally optimize the shape and control of a robotic manipulator for a specific task.

Self-learning robots go full steam ahead

Self-learning robots go full steam ahead

Researchers have shown that a group of small autonomous, self-learning robots can adapt easily to changing circumstances. They connected the simple robots in a line, after which each individual robot taught itself to move forward as quickly as possible.

Smart system helps robots navigate emergency rooms

Smart system helps robots navigate emergency rooms

Scientists have developed a more accurate navigation system that allows robots to better negotiate busy clinical environments in general and emergency departments more specifically.

Algorithm designs soft robots that sense

Algorithm designs soft robots that sense

Deep learning technique optimizes the arrangement of sensors on a robot’s body to ensure efficient operation.

Key task in computer vision gets a boost

Key task in computer vision gets a boost

A researcher has demonstrated a technique that reduces the computing time for non-rigid point set registration relative to other approaches.

Customized “brains” for robots

Customized “brains” for robots

Researchers have developed an automated way to design customized hardware, or “brains,” that speeds up a robot’s operation.

COVID-19: Robot takes patients’ vital signs

COVID-19: Robot takes patients’ vital signs

Researchers hope to reduce the risk to healthcare workers posed by Covid-19 by using robots to remotely measure patients’ vital signs.

Your new lab partner: the robot scientist

Your new lab partner: the robot scientist

Researchers have built an intelligent mobile robot scientist that can work 24-7, carrying out experiments by itself.

Fighting hand tremors with AI and robots

Fighting hand tremors with AI and robots

Researchers have tapped AI techniques to build an algorithmic model that will make the robots more accurate, faster, and safer when battling hand tremors.

Popular articles

Subscribe to Newsletter