Researchers at Keio University's Haptics Research Center have developed a 'real-time-avatar-robotic arm' that transmits sound, vision, and highly sensitive feelings of touch to remotely located users. This innovative touch sensitive robotic technology that was reported in IEEE Transactions on Industrial Electronics and demonstrated in Tokyo, is expected to find applications in areas such as industrial manufacturing, harvesting farm produce, and nursing care.
There is demand for robotic technology to overcome the daunting challenges of the 21st century such as providing care for the elderly in rapidly aging industrialized nations, supporting labor intensive agriculture, and responding to extreme emergencies where humans cannot intervene directly, such as nuclear power station disasters.
With this background, a growing number of researchers are focussing on the potential of 'haptics'— man-machine communication based on touch—to solve these and related problems. In its simplest form, haptics enables users to feel the sense of touch via vibrations of forced motion. Such technology employs touch sensors that can be difficult to calibrate and often malfunction in extreme environments such as heat and radiation. Furthermore, conventional haptics technology is based on vibrations and is pseudo-tactile. So, although it can be used for games and entertainment, its range of industrial applications is very limited.
Takahiro Nozaki and colleagues of the Faculty of Science and Technology and Haptics Research Center at Keio University developed a haptic-based avatar-robot with a General Purpose Arm (GPA) that transmits sound, vision, movement, and importantly, highly sensitive sense of touch (force tactile transmission), to a remotely located user in real time. "This 'real-haptics' is an integral part of the Internet of Actions (IoA) technology, having applications in manufacturing, agriculture, medicine, and nursing care," says Nozaki.
This is the world's first high precision tactile force transmission technology that remembers human movements, edits them, and reproduces them. Also, this arm does not employ conventional touch sensors, thereby making it cheaper, more compact, and robust with respect to malfunction and noise.
The core technology behind this avatar-robot is based on high precision motors integrated in the avatar arm and algorithms to drive them. High precision control of force and position is critical for transmitting a sense of touch without using touch sensors.
Nozaki and colleagues have launched 'Motion Lib' to commercialize their 'real-haptics technology'. The main product is an integrated chip called the 'ABC-CORE' IC force/tactile controller. This IC chip controls the force adjustment of DC/AC servomotors and forces tactile transmission with two motors synchronized in motion. Importantly, since the load force applied to the motor is calculated by an algorithm in the chip, it is not necessary to install force or torque sensors.
High precision robotic arms are widely used in industry, for repetitive actions in automobile assembly lines, for example. However, such robotic arms only repeat a preprogramed series of commands, grabbing well-defined, solid components used for constructing cars.
The challenge is to be able to recognize the shape, material composition—soft or hard— and position of an object, and manipulate it according to real-time instructions from a user located at a distance from the arm, where the arm acts as a real-time avatar.
Nozaki has set up a consortium with 30 companies to undertake proof of concept projects for the commercialization of this technology as an integral part of the Internet of Actions (IoA). The assist-avatar robotic GPA is being tested for use in supporting farmers to pick fruit and other agricultural applications.
Source and top image: Keio University
Learn more at the next leading event on the topic: IDTechEx Show! Europe 2018 on 11 - 12 Apr 2018 at the Estrel Convention Center, Berlin, Germany hosted by IDTechEx.