Page 250 of 542
1 248 249 250 251 252 542

Towards an interactive cyber-physical human platform to generate contact-rich whole-body motions

Humans naturally perform numerous complex tasks. These include sitting down, picking something up from a table, and pushing a cart. These activities involve various movements and require multiple contacts, which makes it difficult to program robots to perform them.

What are co-bots?

Co-bot stands for “collaborative robots”. These robots are designed to work alongside humans, by performing repetitive or heavy tasks, which would greatly ease the burden on the human worker while he or she can focus on tasks that require higher skills.

A critical item for making this possible is to implement certain safety mechanisms on these robots, in order to prevent harm to humans. This can be achieved by certain sensors placed on the robots body, in order to prevent impacts to human workers. These sensors include force, torque and ultrasonic sensors. It is also necessary to cover robots body with soft material.

Working alongside humans means that there will be more unexpected circumstances in robots environment, in comparison to unchanging environments of industrial robots. Therefore, these robots must operate with considerably more complex visual recognition and AI abilities. These robots can also be trained for new tasks by literally guiding them physically in addition to classical programming, which is very intuitive and efficient.

The market of collaborative robots is ever growing, since their beginning about 1-2 decades ago.  

What are co-bots?

Co-bot stands for “collaborative robots”. These robots are designed to work alongside humans, by performing repetitive or heavy tasks, which would greatly ease the burden on the human worker while he or she can focus on tasks that require higher skills.

A critical item for making this possible is to implement certain safety mechanisms on these robots, in order to prevent harm to humans. This can be achieved by certain sensors placed on the robots body, in order to prevent impacts to human workers. These sensors include force, torque and ultrasonic sensors. It is also necessary to cover robots body with soft material.

Working alongside humans means that there will be more unexpected circumstances in robots environment, in comparison to unchanging environments of industrial robots. Therefore, these robots must operate with considerably more complex visual recognition and AI abilities. These robots can also be trained for new tasks by literally guiding them physically in addition to classical programming, which is very intuitive and efficient.

The market of collaborative robots is ever growing, since their beginning about 1-2 decades ago.  

Engineers devise a modular system to produce efficient, scalable aquabots

Underwater structures that can change their shapes dynamically, the way fish do, push through water much more efficiently than conventional rigid hulls. But constructing deformable devices that can change the curve of their body shapes while maintaining a smooth profile is a long and difficult process. MIT's RoboTuna, for example, was composed of about 3,000 different parts and took about two years to design and build.

Microelectronics give researchers a remote control for biological robots

A photograph of an eBiobot prototype, lit with blue microLEDs. Remotely controlled miniature biological robots have many potential applications in medicine, sensing and environmental monitoring. Image courtesy of Yongdeok Kim

By Liz Ahlberg Touchstone

First, they walked. Then, they saw the light. Now, miniature biological robots have gained a new trick: remote control.

The hybrid “eBiobots” are the first to combine soft materials, living muscle and microelectronics, said researchers at the University of Illinois Urbana-Champaign, Northwestern University and collaborating institutions. They described their centimeter-scale biological machines in the journal Science Robotics.

“Integrating microelectronics allows the merger of the biological world and the electronics world, both with many advantages of their own, to now produce these electronic biobots and machines that could be useful for many medical, sensing and environmental applications in the future,” said study co-leader Rashid Bashir, an Illinois professor of bioengineering and dean of the Grainger College of Engineering.

Rashid Bashir. Photo by L. Brian Stauffer

Bashir’s group has pioneered the development of biobots, small biological robots powered by mouse muscle tissue grown on a soft 3D-printed polymer skeleton. They demonstrated walking biobots in 2012 and light-activated biobots in 2016. The light activation gave the researchers some control, but practical applications were limited by the question of how to deliver the light pulses to the biobots outside of a lab setting.

The answer to that question came from Northwestern University professor John A. Rogers, a pioneer in flexible bioelectronics, whose team helped integrate tiny wireless microelectronics and battery-free micro-LEDs. This allowed the researchers to remotely control the eBiobots.

“This unusual combination of technology and biology opens up vast opportunities in creating self-healing, learning, evolving, communicating and self-organizing engineered systems. We feel that it’s a very fertile ground for future research with specific potential applications in biomedicine and environmental monitoring,” said Rogers, a professor of materials science and engineering, biomedical engineering and neurological surgery at Northwestern University and director of the Querrey Simpson Institute for Bioelectronics.

Remote control steering allows the eBiobots to maneuver around obstacles, as shown in this composite image of a bipedal robot traversing a maze. Image courtesy of Yongdeok Kim

To give the biobots the freedom of movement required for practical applications, the researchers set out to eliminate bulky batteries and tethering wires. The eBiobots use a receiver coil to harvest power and provide a regulated output voltage to power the micro-LEDs, said co-first author Zhengwei Li, an assistant professor of biomedical engineering at the University of Houston.

The researchers can send a wireless signal to the eBiobots that prompts the LEDs to pulse. The LEDs stimulate the light-sensitive engineered muscle to contract, moving the polymer legs so that the machines “walk.” The micro-LEDs are so targeted that they can activate specific portions of muscle, making the eBiobot turn in a desired direction. See a video on YouTube.

The researchers used computational modeling to optimize the eBiobot design and component integration for robustness, speed and maneuverability. Illinois professor of mechanical sciences and engineering Mattia Gazzola led the simulation and design of the eBiobots. The iterative design and additive 3D printing of the scaffolds allowed for rapid cycles of experiments and performance improvement, said Gazzola and co-first author Xiaotian Zhang, a postdoctoral researcher in Gazzola’s lab.

The eBiobots are the first wireless bio-hybrid machines, combining biological tissue, microelectronics and 3D-printed soft polymers. Image courtesy of Yongdeok Kim

The design allows for possible future integration of additional microelectronics, such as chemical and biological sensors, or 3D-printed scaffold parts for functions like pushing or transporting things that the biobots encounter, said co-first author Youngdeok Kim, who completed the work as a graduate student at Illinois.

The integration of electronic sensors or biological neurons would allow the eBiobots to sense and respond to toxins in the environment, biomarkers for disease and more possibilities, the researchers said.

“In developing a first-ever hybrid bioelectronic robot, we are opening the door for a new paradigm of applications for health care innovation, such as in-situ biopsies and analysis, minimum invasive surgery or even cancer detection within the human body,” Li said.

The National Science Foundation and the National Institutes of Health supported this work.


Piloting drones reliably with mobile communication technology

Drones are operating increasingly in areas out of sight of the person controlling them. However, conventional remote controls have a limited range, which makes them unsuitable for these flights. On the other hand, simple mobile network-based systems have so far been unable to guarantee a reliable connection when mobile network loads are high or where there is a lack of network coverage. Researchers at the Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institut, HHI have joined forces with partners in the SUCOM project to develop a new mobile network system that can be used to control drones even over long distances and over difficult terrain.

A lizard-inspired robot to explore the surface of Mars

Technological advances have opened exciting possibilities for space exploration, which could potentially lead to new discoveries about the celestial bodies in our galaxy. Robots have proved to be particularly promising tools to explore other planets, particularly Mars, a terrestrial planet in the solar system that is known to host some similar elements to those found on Earth.

Robot Talk Episode 35 – Interview with Emily S. Cross

Claire chatted to Professor Emily S. Cross from the University of Glasgow and Western Sydney University all about neuroscience, social learning, and human-robot interaction.

Emily S. Cross is a Professor of Social Robotics at the University of Glasgow, and a Professor of Human Neuroscience at the MARCS Institute at Western Sydney University. Using interactive learning tasks, brain scanning, and dance, acrobatics and robots, she and her Social Brain in Action Laboratory team explore how we learn by watching others throughout the lifespan, how action experts’ brains enable them to perform physical skills so exquisitely, and the social influences that shape human-robot interaction.

Page 250 of 542
1 248 249 250 251 252 542