3D Mapping Could Be The Code That Self-Driving Cars Are Looking For

As LiDAR advances and self-driving cars take a step nearer to reality, maybe this tech from Japan can act as a breakthrough? In ancient Egypt the Rosetta Stone translated key words from one language to another and the same principle may well underpin AI’s understanding of terrain. Thinking of 3D objects as language translators in real time, as symmetrical strings of code, not just roadside objects, could help vehicles truly make sense of their complex urban surroundings one day. Here’s the word from Osaka University.
While a vast amount of point cloud data has been accumulated for Japanese public works, it has little use in the unprocessed state. Researchers from Japan have now proposed a deep learning algorithm that uses high-precision 3D maps to automatically generate training data for constructing a road feature identification model from the point cloud data. With a precision higher than the existing methods, the new model could become indispensable in autonomous driving and urban management.
 
In Japan, a substantial amount of point cloud data–a set of data points in space–has been measured and accumulated for public works using mobile mapping systems and terrestrial laser scanners. However, this vast amount of data is of limited use in an unprocessed and unstructured state. Fortunately, it can be structured by automatically extracting a feature using a “plan of completion drawing” that shows the completed geometry of a construction object.
Earlier this year, researchers from Japan, led by Professor Ryuichi Imai of Hosei University, Japan proposed another method for extracting road features using high-precision 3D (HD) map data. However, the applicability of their approach is limited to the developed sections of road maps. While the issue can be solved with deep learning-based identification, they require a large amount of manually-prepared, high-quality training data.
Recently, Prof. Imai and his collaborators, Kenji Nakamura of Osaka University of Economics, Yoshinori Tsukada of Setsunan University, Noriko Aso of Dynamic Map Platform, and Jin Yamamoto of Hosei University developed an algorithm to automate the process of training data generation and constructed a road feature identification model from point cloud data extracted automatically from HD maps.
“Currently, people need to visually check the point cloud data to identify road features as computers cannot recognize them. But with our proposed method, the feature extraction can be done automatically, including the features at undeveloped road map sections,” explains Prof. Imai. Their work was presented at the Joint 12th International Conference on Soft Computing and Intelligent Systems and 23rd International Symposium on Advanced Intelligent Systems on 29th November 2022.
ROAD FEATURE LEARNING CURVE
In their study, the researchers first separated the ground surface from the point cloud data using the CloudCompare software. Next, they generated area data from the HD map and extracted component points of features. While these points were assigned as either road signs or traffic lights, other labels were provided for the remaining data. Then, the area data corresponding to the component points was extended to generate the training data. Using this, the researchers further generated the point cloud projection images. Lastly, they used the training data to construct the identification model using a YOLOv3 object-detection algorithm. The model could detect road features based on clustering points other than those identified for the ground surface using CloudCompare.
Having established the computational framework, the researchers performed demonstration experiments in the Shizuoka Prefecture on a road with 65 road signs, 46 traffic lights, and noise features over a distance of 1.5 kilometers. They used 258 road signs and 168 traffic lights to train their identification model and used 36 and 24 images, respectively, to calculate the algorithm determination accuracy.
The researchers found that the precision, recall, and F-measure were 0.84, 0.75, and 0.79, respectively, for the road signs and 1.00, 0.75, and 0.86, respectively for the traffic lights, indicating zero false determinations. The precision of the proposed model was shown to be higher than the existing models.
Prof. Imai concludes by highlighting the future implications of the work. “A product model constructed from point cloud data will enable the realization of a digital twin environment for urban space with regularly updated road maps. It will be indispensable for managing and reducing traffic restrictions and road closures during road inspections. The technology is expected to reduce time costs for people using roads, cities, and other infrastructures in their daily lives.”

About alastair walker 19320 Articles
20 years experience as a journalist and magazine editor. I'm your contact for press releases, events, news and commercial opportunities at Insurance-Edge.Net

Be the first to comment

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.