Photos are two-dimensional (2D), but autonomous vehicles and other technologies have to navigate the three-dimensional (3D) world. Researchers have developed a new method to help artificial intelligence (AI) extract 3D information from 2D images, making cameras more useful tools for these emerging technologies.
“Existing techniques for extracting 3D information from 2D images are good, but not good enough,” says Tianfu Wu, co-author of a paper on the work and an associate professor of electrical and computer engineering at North Carolina State University. “Our new method, called MonoXiver, can be used in conjunction with existing techniques – and makes them significantly more accurate.”
The work is particularly useful for applications such as autonomous vehicles. That’s because cameras are less expensive than other tools used to navigate 3D spaces, such as LIDAR, which relies on lasers to measure distance. Because cameras are more affordable than these other technologies, designers of autonomous vehicles can install multiple cameras, building redundancy into the system. But that’s only useful if the AI in the autonomous vehicle can extract 3D navigational information from the 2D images taken by a camera. This is where MonoXiver comes in.
Existing techniques that extract 3D data from 2D images – such as developed by Wu and his collaborators – make use of “bounding boxes.” Specifically, these techniques train AI to scan a 2D image and place 3D bounding boxes around objects in the 2D image, such as each car on a street. These boxes are cuboids, which have eight points – think of the corners on a shoebox. The bounding boxes help the AI estimate the dimensions of the objects in an image, and where each object is in relation to other objects. In other words, the bounding boxes can help the AI determine how big a car is, and where it is in relation to the other cars on the road.
However, the bounding boxes of existing programs are imperfect, and often fail to include parts of a vehicle or other object that appears in a 2D image.
The new MonoXiver method uses each bounding box as a starting point, or anchor, and has the AI perform a second analysis of the area surrounding each bounding box. This second analysis results in the program producing many additional bounding boxes surrounding the anchor.
To determine which of these secondary boxes has best captured any “missing” parts of the object, the AI does two comparisons. One comparison looks at the “geometry” of each secondary box to see if it contains shapes that are consistent with the shapes in the anchor box. The other comparison looks at the “appearance” of each secondary box to see if it contains colors or other visual characteristics that are similar to the visual characteristics of what is within the anchor box.
“One significant advance here is that MonoXiver allows us to run this top-down sampling technique – creating and analyzing the secondary bounding boxes – very efficiently,” Wu says.
To measure the accuracy of the MonoXiver method, the researchers tested it using two datasets of 2D images: the well-established KITTI dataset and the more challenging, large-scale Waymo dataset.
“We used the MonoXiver method in conjunction with MonoCon and two other existing programs that are designed to extract 3D data from 2D images, and MonoXiver significantly improved the performance of all three programs,” Wu says. “We got the best performance when using MonoXiver in conjunction with MonoCon.
“It’s also important to note that this improvement comes with relatively minor computational overhead,” Wu says. “For example, MonoCon, by itself, can run at 55 frames per second. That slows down to 40 frames per second when you incorporate the MonoXiver method – which is still fast enough for practical utility.
“We are excited about this work, and will continue to evaluate and fine-tune it for use in autonomous vehicles and other applications,” Wu says.
The paper, “,” will be presented Oct. 4 at the International Conference on Computer Vision in Paris, France. First author of the paper is Xianpeng Liu, a Ph.D. student at NC State. The paper was co-authored by Kelvin Cheng, a former Ph.D. student at NC State; Ce Zheng, of the University of Central Florida; Nan Xue, of Ant Group; and Guo-Jun Qi, of the OPPO Seattle Research Center and Westlake University.
The work was done with support from the U.S. Army Research Office under grants W911NF1810295 and W911NF2210010; and by the ³Ô¹ÏÍøÕ¾ Science Foundation, under grants 1909644, 1822477, 2024688 and 2013451.