Connect with us

Robotics

Robots Take Part in Military Drills

Published

 on

Robots Take Part in Military Drills

One area that will be completely changed due to artificial intelligence (AI) and robotics is the military and defense. The U.S. Army Research Laboratory has been working with universities to develop possible human-robot teams. One of their well known projects, sponsored by the Office of Secretary of Defense’s Autonomy Research Pilot Initiative, is called the Autonomous Squad Member (ASM). The ASM is a small ground robot that is used within an infantry squad, and it is able to communicate and interact with members of the squad.

There are new developments with the research, which has actually been a 10-year project aimed at developing autonomous robots and discovering their potential. Army researchers used ground robots to perform military-style exercises, mimicking real soldiers. The exercises were performed at a robotics testing site in Pennsylvania.

The RoMan

One of the new developments from the research is the RoMan, or Robotic Manipulator. It is a tracked robot that can use its robotic arms and hands to maneuver difficult environments. It is capable of moving heavy objects, and it is used to clear paths and remove debris for military vehicles. There was an immense amount of effort put into developing and programming these robots to be capable of operating in such complex environments.

The exercise with the RoMan was part of several events showcasing the achievements of ten years of research led by different scientists and engineers from the U.S. Army Combat Capabilities Development Command’s Army Research Laboratory. They worked with the NASA/Jet Propulsion Laboratory, University of Washington, University of Pennsylvania, Carnegie Mellon University, and General Dynamics Land Systems.

The work is part of ARL’s Robotics Collaborative Technology Alliance. One of the main aspects was state-of-the-art basic and applied research with ground robotics technologies. The major goal is to develop autonomous robots that can be part of a human-robot team. This research, along with the RCTA program, is important in developing future combat ground vehicles.

A Robot Capable of Reasoning on the Battlefield

The robot military-style exercises put into action the research which has been built around developing a robot that can reason on the field. The goal is to have it capable of interacting with unknown objects and their physical properties. If successful, the robot will be able to reason and figure out the best way to interact with certain objects, leading to the completion of specific tasks.

CCDC ARL’s Dr. Chad Kessens is a Robotic Manipulation researcher. He spoke about the new technology.

“Given a task like ‘clear a path’, the robot needs to identify potentially relevant objects, figure out how objects can be grasped by determining where and with what hand shape, and decide what type of interaction to use, whether that’s lifting, moving, pushing or pulling to achieve its task,” he said.

The RoMan performed successfully during the most recent exercise. It cleared out multi-object debris, and it dragged heavy objects such as tree limbs. It also was successful in opening a container and removing a bag from it.

Controlled Through Verbal Commands

The robot can also interact with the team members through speech. The soldiers give verbal commands to the RoMan, and it can pick up on natural human language depending on the scenario.

“Planning and learning and their integration cut across all these problems. The ability of the robot to improve its performance over time and to adapt to new scenarios by building models on-the-fly while incorporating the power of model-based reasoning will be important to achieving the kinds of unstructured tasks we want to be able to do without putting Soldiers in harm’s way,” Kessens said.

The Army Research Laboratory (ARL) is the U.S. Army’s corporate research laboratory. They have various components including Computational and Information Sciences (CISD), Human Research and Engineering (HRED), Sensors and Electron Devices (SEDD), Survivability/Lethality Devices (SEDD), Vehicle Technology (VTD), and Weapons and Materials Research (WMRD). They are one of the major players in developing artificial intelligence and robotics to be used in combat.

 

Spread the love

Deep Learning Specialization on Coursera

Robotics

Study Suggests Robots Are More Persuasive When They Pretend To Be Human

mm

Published

on

Study Suggests Robots Are More Persuasive When They Pretend To Be Human

Advances in artificial intelligence have created bots and machines that can potentially pass as humans if they interact with people exclusively through a digital medium. Recently, a team of computer science researchers have studied how robots/machines and humans interact when the humans believe that the robots are also human. As reported by ScienceDaily, the results of the study found that people find robots/chatbots more persuasive when they believe the bots are human.

Talal Rahwan, the associate professor of Computer Science at NYU Abu Dhabi, has recently led a study that examined how robots and humans interact with each other. The results of the experiment were published in Nature Machine Intelligence in a report called Transparency-Efficiency Tradeoff in Human-Machine Cooperation. During the course of the study, test subjects were instructed to play a cooperative game with a partner, and the partner may be either a human or a bot.

The game was a twist on the classic Prisoner’s Dilemma, where participants must decide whether or not to cooperate or betray the other on every round. In a prisoner’s dilemma, one side may choose to defect and betray their partner to achieve a benefit at cost to the other player, and only by cooperating can both sides assure themselves of gain.

The researchers manipulated their test subjects by providing them with either correct or incorrect information about the identity of their partner. Some of the participants were told that they were playing with a bot, even though their partner was actually human. Other participants were in the inverse situation. Over the course of the experiment, the research team was able to quantify if people treated partners differently when they were told their partners were bots. The researchers tracked the degree to which any prejudice against the bots existed, and how these attitudes impacted interactions with bots who identified themselves.

The results of the experiment demonstrated that bots were more effective at engendering cooperation from their partners when the human believed that the bot was also a human. However, when it was revealed that the bot was a bot, cooperation levels dropped. Rahwan explained that while many scientists and ethicists agree that AI should be transparent regarding how decisions are made, it’s less clear that they should also be transparent about their nature when communicating with others.

Last year, Google Duplex made a splash when a stage demo showed that it was capable of making phone calls and booking appointments on behalf of its use, generating human-like speech so sophisticated that many people would have mistaken it for a real person had they not been told they were speaking to a bot. Since the debut of Google Duplex, many AI and robot ethicists voiced their concerns over the technology, prompting Google to say that it would have the agent identify itself as a bot in the future. Currently, Google Duplex is only being used in a very limited capacity. It will soon see use in New Zealand, but only to check for the operating hours of businesses. Ethicists are still worried about the degree to which the technology could be misused.

Rahawan argues that the recent study demonstrates that we should consider what costs we are willing to pay in return for transparency:

“Is it ethical to develop such a system? Should we prohibit bots from passing as humans, and force them to be transparent about who they are? If the answer is ‘Yes’, then our findings highlight the need to set standards for the efficiency cost that we are willing to pay in return for such transparency.”

Spread the love

Deep Learning Specialization on Coursera
Continue Reading

Robotics

Flexible Robot “Grows” Like a Plant

Published

on

Flexible Robot “Grows” Like a Plant

Engineers from MIT have designed a robot that can extend a chain-like appendage. This makes the robot extremely flexible, and it can configure itself in multiple different ways. At the same time, it is strong enough to support heavy weight or apply torque, making it capable of assembling parts in small spaces. After completing its tasks, the robot is able to retract the appendage, and it can extend it again with a different length and shape. 

This newly developed robot can make a difference in areas like warehouses, where most of the robots are not able to put themselves in narrow spaces. The new plant-like robot can be used to grab products at the back of a shelf, and it can even move around a car’s engine parts to unscrew an oil cap. 

The design was inspired by plants and the way they grow. In that process, nutrients are transported to the plant’s tip as a fluid. Once they reach the tip, they are converted into solid material that produces, a little at a time, a supportive stem. 

The plant-like robot has a “growing point” or gearbox, which draws a loose chain of interlocking blocks into the box. Once there, gears lock the chain units together and release the chain, unit by unit, until it forms a rigid appendage. 

Team of Engineers

The new robot was presented this week at the IEEE International Conference on Intelligent Robots and Systems (IROS) in Macau. In the future, the engineers would like to add on grippers, cameras, and sensors that could be mounted onto the gearbox. This would allow the robot to tighten a loose screw after making its way through an aircraft’s propulsion system. It could also retrieve a product without disturbing anything in the near surroundings. 

Harry Asada is a professor of mechanical engineering at MIT.

“Think about changing the oil in your car,”  Asada says. “After you open the engine roof, you have to be flexible enough to make sharp turns, left and right, to get to the oil filter, and then you have to be strong enough to twist the oil filter cap to remove it.”

Tongxi Yan is a former graduate student in Asada’s lab, and he led the work.

“Now we have a robot that can potentially accomplish such tasks,” he says. “It can grow, retract, and grow again to a different shape, to adapt to its environment.”

The team of engineers also consisted of MIT graduate student Emily Kamienski and visiting scholar Seiichi Teshigawara.

Plant-Like Robot

After defining the different aspects of plant growth, the team looked to implement it into a robot. 

“The realization of the robot is totally different from a real plant, but it exhibits the same kind of functionality, at a certain abstract level,” Asada says.

The gearbox was designed to represent the robot’s “growing tip,” which is the equivalent of a bud of a plant. That is where most nutrients flow up to the site, and the tip builds a rigid stem. The box consists of a system of gears and motors, and they pull up a fluidized material. For this robot, it is a sequence of 3-D printed plastic units that are connected with each other. 

The robot is capable of being programmed to choose which units to lock together and which to leave unlocked. This allows it to form specific shapes and “grow” in specific directions.

“It can be locked in different places to be curved in different ways, and have a wide range of motions,” Yan says.

The chain is able to support a one-pound weight when locked and rigid. If a gripper were to be attached, the researchers believe it would be able to grow long enough to maneuver through a narrow space, and perform tasks such as unscrewing a cap.

 

Spread the love

Deep Learning Specialization on Coursera
Continue Reading

Robotics

Researchers Develop Resilient RoboBee with Soft Muscles

Published

on

Researchers Develop Resilient RoboBee with Soft Muscles

Researchers at the Harvard Microrobotics Laboratory at the Harvard John A. Paulson School of Engineering and Applied Science (SEAS), along with the Wyss Institute for Biologically Inspired Engineering, have developed a RoboBee powered by soft artificial muscles. The microrobot is capable of crashing into walls, falling on the ground, and colliding with other RoboBees without suffering damage. In what is a big moment for robotics, the RoboBee is the first microrobot powered by soft actuators that is able to achieve controlled flight. 

Yufeng Chen is first author of the paper and a former graduate student and postdoctoral fellow at SEAS.

“There has been a big push in the field of microrobotics to make mobile robots out of soft actuators because they are so resilient,” said Chen. “However, many people in the field have been skeptical that they could be used for flying robots because the power density of those actuators simply hasn’t been high enough and they are notoriously difficult to control. Our actuator has high enough power density and controllability to achieve hovering flight.”

The research was published in Nature.

Issues Encountered

One of the problems that the researchers dealt with was power density. They looked to the electrically-driven soft actuators that were developed in the lab of David Clarke, the Extended Tarr Family Professor of Materials. The soft actuators are created by using dielectric elastomers, which are soft materials that have strong insulating properties. When an electric field is applied, the dielectric elastomers deform. 

After improving the electrode conductivity, the actuator was able to be operated at 500 Hertz. This is similar to previously used rigid actuators in robots. 

One of the other issues with soft actuators is that the system often becomes unstable. To get past this, the researchers developed a lightweight airframe. It consisted of a piece of vertical constraining thread in order to prevent the actuator from buckling. 

Flight Capability

Within the small scale robots, the soft actuators are able to be swapped out and assembled easily. The researchers developed multiple different models of the soft-powered RoboBee in order to showcase the various flight capabilities. 

One of the models has two wings, and it can take off from the ground. However, this model has no further control. A four-wing, two actuator model is capable of flying in a crowded environment. Within a single flight, the RoboBee is able to avoid multiple collisions.

Elizabeth Farrell Helbling is a former graduate student at SEAS, and she co-authored the paper. 

“One advantage of small-scale, low-mass robots is their resilience to external impacts,” she said. “The soft actuator provides an additional benefit because it can absorb impact better than traditional actuation strategies. This would come in handy in potential applications such as flying through rubble for search and rescue missions.”

Another model is the eight-wing, four-actuator RoboBee. It is capable of performing controlled hovering flight, which is the first time it has been demonstrated by a soft-powered flying microrobot. 

What’s Next?

The researchers are now looking to increase the efficiency of the soft-powered RoboBee. It still has a long way to go before catching up to traditional flying robots. 

Robert Wood is a Charles River Professor of Engineering and Applied Sciences in SEAS. He is also a core faculty member of the Wyss Institute for Biologically Inspired Engineering and senior author of the paper. 

“Soft actuators with muscle-like properties and electrical activation represent a grand challenge in robotics,” says Professor Wood.  “If we could engineer high-performance artificial muscles, the sky is the limit for what robots we could build.”

 

Spread the love

Deep Learning Specialization on Coursera
Continue Reading