In a landmark study, Associate Professor Yoshihiro Nakata of the University of Electro-Communications, Japan, in collaboration with researchers from Osaka University, conducted research on edible robotics. Published in PLOS ONE, this study is the first to explore the experience of consuming a moving edible robot.
Inspired by the remarkable adaptability observed in biological organisms like the octopus, a breakthrough has been achieved in the field of soft robotics. A research team, led by Professor Jiyun Kim in the Department of Materials Science and Engineering at UNIST has successfully developed an encodable multifunctional material that can dynamically tune its shape and mechanical properties in real time.
Research involving Oregon State University has shown that a "swarm" of more than 100 autonomous ground and aerial robots can be supervised by one person without subjecting the individual to an undue workload.
A team of roboticists at New York University, working with a colleague from AI at Meta, has developed a robot that is capable of picking up designated objects in an unfamiliar room and placing them in a new designated location. In their paper posted on the arXiv preprint server, the team describes how the robot was programmed and how well it performed when tested in multiple real-word environments.
The New York City subway has pulled its controversial security robot out of service after little more than five months patrolling the busy Times Square station.
In the coming decades, NASA plans to send human crews back to the moon, build a space station in lunar orbit, establish a permanent base on the lunar surface, and—hopefully—send astronauts to Mars.
In the future, soft robots will be able to perform tasks that cannot be done by conventional robots. These soft robots could be used in terrain that is difficult to access and in environments where they are exposed to chemicals or radiation that would harm electronically controlled robots made of metal. This requires such soft robots to be controllable without any electronics, which is still a challenge in development.
Recent technological advances, such as increasingly sophisticated drones and cameras, have opened exciting new possibilities for cinematography. Most notably, film directors can now shoot scenes from a wide range of angles that were previously inaccessible and in far higher resolution.
Figure AI Inc., a startup developing humanlike robots, is in talks to raise as much as $500 million in a funding round led by Microsoft Corp. and OpenAI, according to a person with knowledge of the matter.
Our built environment is aging and failing faster than we can maintain it. Recent building collapses and structural failures of roads and bridges are indicators of a problem that's likely to get worse, according to experts, because it's just not possible to inspect every crack, creak and crumble to parse dangerous signs of failure from normal wear and tear.
It abseils from a height and swings around obstacles: Robot Avocado will one day maneuver through the canopy of the rainforest and collect data for researchers about this hard-to-reach habitat.
Researchers at ETH Zurich have recently developed artificial muscles for robot motion. Their solution offers several advantages over previous technologies: It can be used wherever robots need to be soft rather than rigid or where they need more sensitivity when interacting with their environment.
Researchers have developed a robotic sensor that incorporates artificial intelligence techniques to read braille at speeds roughly double that of most human readers.
Social robots can help us with many things: teaching, learning, caring. Because they're designed to interact with humans, they're designed to make us comfortable—and that includes the way they talk. But how should they talk? Some research suggests that people like robots to use a familiar accent or dialect, while other research suggests the opposite.
Against a backdrop of looming skyscrapers, a robot dinosaur raises its feathery wings, sending its prey fleeing in terror.