Researchers at Keio Universitys Haptics Research Center have developed a ‘real-time-avatar-robotic arm that transmits sound, vision, and highly sensitive feelings of touch to remotely located users. This innovative touch sensitive robotic technology was reported in the October 2017 issue of IEEE Transactions on Industrial Electronics and demonstrated at CEATEC (October 2017, Tokyo). The technology is expected to find applications in areas such as industrial manufacturing, harvesting farm produce, and nursing care.
Contributed by | Keio University
There is demand for robotic technology to overcome the daunting challenges of the 21st century such as providing care for the elderly in rapidly aging industrialized nations, supporting labor intensive agriculture, and responding to extreme emergencies where humans cannot intervene directly, such as nuclear power station disasters.
With this background, a growing number of researchers are focussing on the potential of ‘haptics’— man-machine communication based on touch—to solve these and related problems. In its simplest form, haptics enables users to feel the sense of touch via vibrations of forced motion. Such technology employs touch sensors that can be difficult to calibrate and often malfunction in extreme environments such as heat and radiation. Furthermore, conventional haptics technology is based on vibrations and is pseudo-tactile. So, although it can be used for games and entertainment, its range of industrial applications is very limited.
Takahiro Nozaki and colleagues of the Faculty of Science and Technology and Haptics Research Center at Keio University developed a haptic-based avatar-robot with a General Purpose Arm (GPA) that transmits sound, vision, movement, and importantly, highly sensitive sense of touch (force tactile transmission), to a remotely located user in real time. “This ‘real-haptics’ is an integral part of the Internet of Actions (IoA) technology, having applications in manufacturing, agriculture, medicine, and nursing care,” says Nozaki.
This is the world’s first high precision tactile force transmission technology that remembers human movements, edits them, and reproduces them. Also, this arm does not employ conventional touch sensors, thereby making it cheaper, more compact, and robust with respect to malfunction and noise. The core technology behind this avatar-robot is based on high precision motors integrated in the avatar arm and algorithms to drive them. High precision control of force and position is critical for transmitting a sense of touch without using touch sensors.
Nozaki and colleagues have launched ‘Motion Lib’ to commercialize their ‘real-haptics technology’. The main product is an integrated chip called the ‘ABC-CORE’ IC force/tactile controller. This IC chip controls the force adjustment of DC/AC servomotors and forces tactile transmission with two motors synchronized in motion. Importantly, since the load force applied to the motor is calculated by an algorithm in the chip, it is not necessary to install force or torque sensors.
Background
- High precision robotic arms are widely used in industry, for repetitive actions in automobile assembly lines, for example. However, such robotic arms only repeat a preprogramed series of commands, grabbing well-defined, solid components used for constructing cars.
- The challenge is to be able to recognize the shape, material composition—soft or hard— and position of an object, and manipulate it according to real-time instructions from a user located at a distance from the arm, where the arm acts as a real-time avatar.
- The critical technical breakthroughs in motor control and robotics for the robotic-avatar developed by Nozaki and co-workers were first reported by Keio University’s Kouhei Ohnishi, in 1983 in a paper titled, “Torque –speed regulation of DC motor based on load torque estimation method” (lPEC― Tokyo'83, page 1209).
- Ohnishi continued to develop his ideas in his 1993 paper on ‘sensorless torque control’ (IEEE Transactions on Industrial Electronics, 40, 259, (1993)).
- This report was followed by his proposals for ‘motion control in mechatronics’ (IEEE Transactions on Mechatronics, 1, 56, (1996)).
- Then, in 2004, Ohnishi addressed the issue of ‘good sense of remote objects’ in AMC 2004– Kawasaki, Japan.
Future work
Nozaki has set up a consortium with 30 companies to undertake proof of concept projects for the commercialization of this technology as an integral part of the Internet of Actions (IoA). The assist-avatar robotic GPA is being tested for use in supporting farmers to pick fruit and other agricultural applications.
Touch sensitive avatar-robotic arm based on real-time haptics developed by Takahiro Nozaki and colleagues at the Keio University Haptics Research Center.
Takahiro Nozaki
Assistant Professor, Faculty of Science and Technology, Keio University
Reference
1. Satoshi Fukushima, Hiromu Sekiguchi, Yuki Saito, Wataru Iida, Takahiro Nozaki, and Kouhei Ohnishi. Artificial Replacement of Human Sensation Using Haptic Transplant Technology. IEEE Transactions on Industrial Electronics, Online 2 October 2017. DOI: 10.1109/TIE.2017.2758757
Related information
1. Motion Lib, Inc.
http://motionlib.com/ (Japanese)
2. Haptics Research Center, Keio University
http://haptics-c.keio.ac.jp/ (Japanese)
The content & opinions in this article are the author’s and do not necessarily represent the views of RoboticsTomorrow
Comments (0)
This post does not have any comments. Be the first to leave a comment below.