International Space Station astronauts remotely control planetary rover from space
July 31, 2013
NASA has begun studying how remotely-operated vehicles may one day help astronauts explore other worlds.
NASA tested the Surface Telerobotics exploration concept, in which an astronaut in an orbiting spacecraft remotely operates a robot on a planetary surface. In the future, astronauts orbiting other planetary bodies, such as Mars, asteroids or the moon, could use this approach to perform work on the surface using robotic avatars.
“The initial test was notable for achieving a number of firsts for NASA and the field of human-robotic exploration,” said Terry Fong, Human Exploration Telerobotics project manager and director of the Intelligent Robotics Group at NASA’s Ames Research Center, Moffett Field, Calif., which designed and manages the tests. “Specifically, this project represents the first fully-interactive remote operation of a planetary rover by an astronaut in space.”
During the June 17 test, Expedition 36 Flight Engineer Chris Cassidy of NASA remotely operated the K10 planetary rover in the Roverscape — an outdoor robotic test area the size of two football fields located at NASA Ames — hundreds of miles below on Earth’s surface from his post aboard the International Space Station (ISS). For more than three hours, Cassidy used the robot to perform a survey of the Roverscape’s rocky, lunar-like terrain.
The July 26 test picked up where Cassidy left off. Fellow Expedition 36 Flight Engineer Luca Parmitano of the European Space Agency remotely-controlled the rover and began deploying a simulated Kapton film-based radio antenna.
These tests represent the first time NASA’s open-source Robot Application Programming Interface Delegate (RAPID) robot data messaging system was used to control a robot from space. RAPID originally was developed by NASA’s Human-Robotic Systems project and is a set of software data structures and routines that simplify the process of communicating information between different robots and their command and control systems. RAPID has been used with a wide variety of systems including rovers, walking robots, free-flying robots and robotic cranes.
The test also is the first time the NASA Ensemble-based software — jointly developed at Ames and NASA’s Jet Propulsion Laboratory in Pasadena, Calif. — was used in space for telerobotics. Ensemble is an open architecture for the development, integration and deployment of mission operations software.
“Whereas it is common practice in undersea exploration to use a joystick and have direct control of remote submarines, the K10 robots are more intelligent,” said Fong. “Astronauts interact with the robots at a higher level, telling them where to go, and then the robot itself independently and intelligently figures out how to safely get there.”
The primary objective of the Surface Telerobotics testing is to collect engineering data from astronauts aboard the space station, the K10 robot and data communication links. This will allow engineers to characterize the system and validate previous ground tests.
“During future missions beyond low-Earth orbit, some work will not be feasible for humans to do manually,” said Fong. “Robots will complement human explorers, allowing astronauts to perform work via remote control from a space station, spacecraft or other habitat.”
The primary goal of the Human Exploration Telerobotics project is to understand how human and robot activities, such as Surface Telerobotics, can be coordinated to improve crew safety, enhance science activities and increase mission success while also reducing cost, risk and consumables, such as fuel and oxygen, during future exploration missions.
The K10 robot is a four-wheel drive, four-wheel steer robot that stands about 4.5 feet tall, weighs about 220 pounds and can travel about three feet per second (a little slower than the average person’s walking pace). For the Surface Telerobotics tests, K10 is equipped with multiple cameras and a 3-D scanning laser system to perform survey work, as well as a mechanism to deploy the simulated radio antenna.
This year’s Surface Telerobotics tests simulate a possible future mission involving astronauts aboard NASA’s Orion spacecraft traveling to the L2 Earth-moon Lagrange point. The L2 point is where the combined gravity of the Earth and moon allows a spacecraft to easily maintain a stationary orbit and is located 40,000 miles above the far side of the moon. From L2, astronauts would remotely operate a robot to perform surface science work, such as deploying a radio telescope. This mission concept was developed by the Lunar University Network for Astrophysics Research (LUNAR), which is based at the University of Colorado, Boulder (CU).
“Deploying a radio telescope on the farside of the moon would allow us to make observations of the early universe free from the radio noise of Earth,” said Jack Burns, a professor at CU, director of LUNAR and co-investigator at NASA’s Lunar Science Institute. “The Surface Telerobotics test represents a next step in new modes of exploration that will bring together humans and robots, as well as science and exploration. Such telerobotics technology will be needed for exploration of the moon, asteroids and eventually the surface of Mars.”
Students from several universities assisted with the development of Surface Telerobotics. Industrial design students from the Academy of Art University in San Francisco collaborated with NASA engineers to create the user interface for remotely operating the K10 rover. Undergraduates from CU and the University of Idaho helped design the Kapton film deployer, which is mounted on K10.
“These surface telerobotics tests, in collaboration with astronauts aboard the ISS, offer exciting opportunities for our students to have hands-on engineering and mission operations experiences with realistic simulations of future human-robot missions to planetary bodies,” said Burns. ”Such experiences inspire our students to careers in the aerospace sciences. These students are destined for bright futures as part of NASA’s exploration of the solar system.”
“This work really tests the notion that robots can project human presence to other planetary surfaces,” said Fong. “Ultimately, this will allow us to discover and explore dangerous and remote places, whether they’re at the bottom of the ocean or at the far reaches of our solar system.”