Socionext and Tohoku University Significantly Accelerate Deep Learning-based SLAM Processing
YOKOHAMA, Japan, Oct. 12, 2021 /PRNewswire/ -- Socionext Inc. has collaborated with a research group from Tohoku University, led by Professor Takayuki Okatani, and developed a new method that can reduce processing time required for SLAM (Simultaneous Localization and Mapping), which is essential for devices that perform autonomous control, down to 1/60 of the time needed by conventional technologies. The new method enables advanced SLAM processing even on SoCs for edge devices with limited CPU performance and power consumption, and is expected to be applied to self-driving cars, AGVs (Automatic Guided Vehicle), robots, drones, and other devices that perform autonomous control, as well as devices such as AR (augmented reality) glasses.
The research work has been accepted as an oral presentation paper for ICCV, one of the most prestigious conferences in the field of computer vision, held online from October 11 through 17.
SLAM can be classified into two main types according to the sensing method used to acquire 3D information around an object. One is LiDAR (Light Detection and Ranging), which measures distance using a laser. The other is Visual SLAM, which uses camera images. Visual SLAM is evolving significantly, because the cameras used are relatively inexpensive and many applications can be expected by combining the method with various control processes using image recognition in addition to simultaneous localization.
Image recognition technology has made remarkable progress in recent years with the introduction of deep learning, and the use of deep learning has become an important factor in the evolution of Visual SLAM. However, the huge amount of computation required for bundle adjustment (BA), which optimizes the unknown 3D information of landmarks based on the feature points and camera orientation, is the bottleneck of conventionally proposed methods. This can make practical processing difficult for edge-type SoC devices with limited CPU processing power. (Fig. 1)
As a solution to this challenge, the research team proposed a method of approximate computation by inference using "Graph Network (GN)", a kind of graph neural network. The new method consists of the inference of updated information from keyframe and landmark information input by GN blocks (Fig. 2), and convergence to the final value by a multi-stack GN structure (Fig. 3). These have enabled inference processing that requires less computation compared to the standard bundle adjustment using the conventional Levenberg-Marquardt method.
The research team implemented Visual SLAM bundle adjustment using this new inference method and compared it with "g2o", a method widely used today. PC simulation results confirmed that the new method can reduce the processing time to 1/60 compared to g2o (Fig. 4).
Socionext will accumulate experiences from this research work and establish Visual SLAM technology as one of the offerings for the company's custom SoC solutions. The company will propose innovative performance improvements for customer systems in fields that require image recognition, such as industrial equipment and mobility. In addition, the company will continue research and development on processing efficiency improvement by the new inference method, and expand its usage to new customer applications outside of image recognition.
All company or product names mentioned herein are trademarks or registered trademarks of their respective owners. Information provided in this press release is accurate at time of publication and is subject to change without advance notice.
View original content to download multimedia:https://www.prnewswire.com/news-releases/socionext-and-tohoku-university-significantly-accelerate-deep-learning-based-slam-processing-301397924.html
SOURCE Socionext Inc.
Keynote Presentation - Open to all Badge Holders
We Only have 5 Senses, but Things have Sensors Galore
Brunch Served in IoT Evolution Expo #TECHSUPERSHOW Expo Hall