The team in Jeffrey C. Trinkles lab work with a next-generation lightweight...
The team in Jeffrey C. Trinkle's lab work with a next-generation lightweight robotic arm designed specifically for academic and industrial research. A key feature of the arm, according to Jinda Cui, Ph.D. student and research assistant, is that it has "seven degrees of freedom," or seven independently controlled joints that allow for superior reachability in the three-dimensional workspace.
Source: Christa Neu, Lehigh University

Computer simulation will accelerate development of 'smart robots'

In new paper, experts working at the intersection of robotics, machine learning, and physics-based simulation share how computer simulation could accelerate the development of "smart robots" which "might interact with humans"

Jeffrey C. Trinkle has always had a keen interest in robot hands. And, though it may be a long way off, Trinkle, who has studied robotics for more than thirty years, says he's most compelled by the prospect of robots performing "dexterous manipulation" at the level of a human "or beyond."

"I've always felt that for robots to be really useful they have to pick stuff up, they have to be able to manipulate it and put things together and fix things, to help you off the floor and all that," he says, adding: "It takes so many technical areas together to look at a problem like that that a lot of people just don't bother with it."

But Trinkle, a faculty member and Chair of the Computer Science and Engineering Department at Lehigh University, does more than bother with it. He and his team are working on projects that involve just such technical challenges. One such project is a collaboration to develop a new approach to the design and construction of soft robots—the future of human-machine collaboration, says Trinkle—inspired by the movement of natural muscles in soft animal structures. Think: giraffe tongues, octopus tentacles and elephant trunks. Computer simulation is key to developing this new approach.

Trinkle, along with colleagues at other institutions, has co-authored a "Perspective" paper that appears in the Proceedings of the National Academy of Sciences (PNAS) arguing that "...well-validated computer simulation can provide a virtual proving ground that in many cases is instrumental in understanding safely, faster, at lower costs, and more thoroughly how the robots of the future should be designed and controlled for safe operation and improved performance."

"Against this backdrop," continue the authors, "we discuss how simulation can help in robotics, barriers that currently prevent its broad adoption, and potential steps that can eliminate some of these barriers"

The article summarizes the points of view expressed during a 2018 National Science Foundation/Department of Defense/National Institute for Standards and Technology workshop dedicated to the topic. The meeting brought together participants from a range of organizations, disciplines, and fields. The expertise represented comes from an intersection of robotics, machine learning, and physics-based simulation, says the authors.

The authors conclude: "The pressing question is how to jump start a robotics-simulation cross-pollination process that would speedily transition the effort from the 'academic debate' phase into a 'simulation-enabled building and demonstration of technology' phase. This transition can be catalyzed by a sustained decade-long financial commitment, which would ensure funding for cross-disciplinary efforts that promote collaboration, competition, and compilation of open repositories of validated models and source code. As witnessed in the aerospace and automotive industries, the paradigm shift to digital, while manifestly impactful, took decades to coalesce. Learning from this experience, the hope is that simulation will lead to breakthroughs in the design of smart robots in a matter of years rather than decades."

Recommended article

Training a robot's neural network: like a baby learning to crawl

Trinkle's soft robots project is a collaboration with Yale University, the University of Washington and Brown University, funded by a National Science Foundation Emerging Frontiers in Research and Innovation grant. His role is to use mathematical models, along with computer science techniques like search algorithms, to develop the computer system that "tells" the robot how to move. He offers the Roomba, the robotic vacuum that moves across the floor autonomously, as an accessible example. "Roombas use the same kind of technologies we are using, just in a very simplified setting," says Trinkle. "Roombas know that there's a wall. They have an internal map that ensures they can move around without bumping into things too much."

Using a simulation tool, Trinkle applies mathematical models to instruct the simulated appendage to curl around a simulated object such as a circle, representing what would be a disk in three-dimensions.

To design the "brain" or "map" that will ultimately instruct a three-dimensional robot how to move, Trinkle uses techniques employed in building artificial neural networks, a type of machine learning that is modeled on the human brain. These neural networks are trained through data in a process similar to human learning. The data "trains" the system through a process akin to "trial and error." In this case, the network is trained with data generated by the computer simulation of an abstracted elephant trunk.

To get the abstracted trunk to curl around the circle and, ultimately, move the circle to another part of the screen, involves multiple steps and a lot of trial and error as the system gets trained.

"[The system] doesn't know anything, so when it tightens up the fibers on one side or the other, it doesn't know in advance how it's going to move, but the neural network is going to figure out that if it does certain things to its muscle fibers, it's going to move a certain way," says Trinkle. "Over a long time horizon, the simulation will figure some things out. It's going to try to end up pushing the disk in the right location." 

He likens the process to an infant learning to crawl. "If a baby is trying to learn how to crawl, it's going to do some things that won't work, and eventually the infant figures it out," says Trinkle. "At some point, all of a sudden, the baby solves the problems and now it's crawling because its neural network has been trained from its experience."

In this research, computer simulation is the training ground for robot systems the teams will build.

Recommended article

Making robots "more intelligent, more autonomous, more adaptive"

Trinkle and his students conduct experiments to extract "ground truth," or directly observed information to feed into their computer program. They are currently working with a next-generation lightweight robotic arm designed specifically for academic and industrial research. A key feature of the arm, according to Jinda Cui, Ph.D. student and research assistant in Trinkle's lab, is that it has "seven degrees of freedom," or seven independently controlled joints that allow for superior reachability in the three-dimensional workspace.

"This robot has one redundant joint that can move without impacting the end-effector," says Cui.

This is significant because it means, for example, that the robot's gripping mechanism, or "hand," can remain in the same pose even while the other "arm" joints are moving. This is especially useful when the environment is cluttered and the robot needs to make a lot of adjustments, explains Cui.

Still, the more joints it has, the more challenging a robot is to control. Another benefit of the advanced nature of the robot is that the team can access the low-level system that directly controls the joint movements. "Through the internet or other communication protocols, we can communicate with the controller directly and tell the joints what to do," says Cui. "We can modify the joint speed individually, or even how much current, or how much torque we want to give a particular joint."

Trinkle and his team have also set up a motion capture camera system, similar to what is used to map the movement of actors in computer-generated imagery in film. Their system is designed to gather data about object positioning that would eventually inform their computer program. "If this robot is trying to manipulate something, like pick your phone up, the vision system should be able to track where your phone is and then the robot can adjust if the phone is not where it expected it to be," says Trinkle.

In the future, instead of an external motion capture camera system, robots will have built-in sensors that track object motion.

The team's robot comes with a three-dimensional camera already installed. For their experiments, Trinkle and Cui plan to add other detectors, such as tactile sensors which would be important for programming the robot to note changing contact conditions, like an object's slipperiness, and then be able to adjust for it. "Our work," says Cui, "is to make it more intelligent, more autonomous, more adaptive."

Recommended article

Closing the "reality gap" between simulation and actual robot

Trinkle notes that training a robot's control system with simulations has its shortcomings. When attempted in three-dimensional space, there is often a "reality gap." The artificial neural network, trained from simulation data, once applied to a physical robot may try to perform the same task as in the simulation and fail. He says that occurs because the model that was learned from the simulation data was biased toward the way the simulation works.

The challenge for robotics researchers like Trinkle is to develop a solid simulation and then do some testing in the physical world, knowing that some retraining will be required. Hopefully, says Trinkle, researchers are just making small adjustments so that it works in the physical world without having to start the whole process over.

In other words, in developing this new approach to teaching soft robots, Trinkle will try to train the "baby" 90 percent of the way to crawling in simulation, and then get it the rest of the way there by experimenting on a physical robot.

The team plans to build three full robot systems as testing prototypes, from modular motor units including a soft robotic hand that can grasp a wide range of object sizes and shapes; a trunk-like structure with a static base that can grasp and manipulate forward and backward, up and down and left and right; and a worm-like robot that can move freely over terrain with large obstacles. These robots will likely involve some sort of silicone skin to create a more continuous surface for contact.

"The hope is that simulation will lead to breakthroughs in the design of smart robots in a matter of years rather than decades."

As for robots someday being able to grasp, Trinkle notes: "It was a hot topic while I was in grad school and then it got cold for 15 years and then it got hot again. Perhaps because so many other problems were solved, and the remaining ones were just as hard as the grasping problem, so now there are a lot more people working on grasping again. And now that AI and neural networks have gotten so big, people are trying to apply those techniques in all sorts of different ways to grasping because it's still the hard problem that it was 30 years ago."

Though he believes there is a long way to go before scientists resolve the technical challenges of getting robots to grasp as well as humans or better, Trinkle acknowledges that those are not the only challenges that need to be overcome. "There are social issues such as: Will people want to be close to robots? Will they become friends?" he ponders. There are ethical challenges to wrestle with as well. He points to autonomous cars as an example. In a difficult circumstance, he asks, should the car be programmed to save the driver and passengers, or save a passing pedestrian who might be impacted?

"There are so many different kinds of problems that people could study," Trinkle adds. "When it comes to robots, who knows if society will be able to accept them and how far we will be able to advance the technology."

Subscribe to our newsletter

Related articles

Using AI to predict 3D printing processes

Using AI to predict 3D printing processes

Engineers use Frontera supercomputer to develop physics-informed neural networks for additive manufacturing.

'Liquid' machine learning system adapts to changing conditions

'Liquid' machine learning system adapts to changing conditions

A machine learning system learns on the job. By continuously adapting to new data inputs, this “liquid network” could aid decision-making in medical diagnosis.

Fighting hand tremors with AI and robots

Fighting hand tremors with AI and robots

Researchers have tapped AI techniques to build an algorithmic model that will make the robots more accurate, faster, and safer when battling hand tremors.

Simulator helps robots sharpen their cutting skills

Simulator helps robots sharpen their cutting skills

A simulation engine predicts the forces acting on a knife as it cuts through soft materials, a capability that could have applications for safer surgical robotics.

Liquid metal sensors and AI used for prosthetics

Liquid metal sensors and AI used for prosthetics

For the first time, researchers incorporated stretchable tactile sensors using liquid metal on the fingertips of a prosthetic hand.

A contact aware robot design

A contact aware robot design

Researchers have developed a new method to computationally optimize the shape and control of a robotic manipulator for a specific task.

Artificial intelligence for emergency management

Artificial intelligence for emergency management

A consortium aims to develop a platform that will serve as the basis for novel services and test the use of new artificial intelligence tools.

Deep learning predicts viral infections

Deep learning predicts viral infections

Using fluoresence images from live cells, researchers have trained an artificial neural network to reliably recognize cells that are infected by adenoviruses or herpes viruses.

When the robot smiles back

When the robot smiles back

Researchers use AI to teach robots to make appropriate reactive human facial expressions, an ability that could build trust between humans and their robotic co-workers and care-givers.

Popular articles

Photo

The “RoboWig” untangle your hair

Nurses typically spend 18 to 40 percent of their time performing direct patient care tasks, oftentimes for many patients and with little time to spare. Personal care robots that brush your hair could provide substantial help and relief.

Subscribe to Newsletter