SPRACX1A April   2021  – April 2021 TDA4VM , TDA4VM-Q1

 

  1.   Trademarks
  2. 1Introduction
  3. 2The Visual Localization Problem
    1. 2.1 Key Point Extraction and Descriptor Computation
    2. 2.2 Feature Matching and Pose Estimation
  4. 3Visual Localization on TDA4VM
  5. 4Example Visual Localization Application
    1. 4.1 Optimized Building Blocks for Your Own Visual Localization Pipeline
  6. 5References
  7. 6Revision History

The Visual Localization Problem

In the simplest sense, visual localization, as the name implies, is the problem of determining the location of a vehicle or robot by matching key-points in a stored map with key-points extracted from images captured by a camera mounted on a vehicle/robot. A key-point is a unique or distinctive point in space from which a descriptor can be extracted. A descriptor is a set of values (a vector) that holds information about a key-point, that will help distinguish said key-point from others. The method that is used to compute these features is described in the next section.

The first step in localization is extracting key-points from the image. Then, the extracted key-points, which are on a 2D image plane, need to matched with a 3D sparse map held in memory. To create the 3D sparse map, features need to be extracted and stored together with their corresponding locations in some arbitrary but known coordinate system. This task is typically achieved by driving a vehicle equipped with a high precision differential GPS and camera along all the paths that make up the map. In order to ensure the features are not biased by the time of day, or day of the year, information is gathered throughout the year to refine the map. Then, when the position of the vehicle/robot needs to be estimated, key points extracted from an image are matched with key points in the sparse 3D map, and using the point correspondence the pose of the vehicle/robot is estimated. This process is described in more detail in the next section.

The entire localization process is shown at a high level in Figure 2-1.

GUID-20210312-CA0I-GQN1-DJ0F-7HX6KJK9V5CZ-low.svg Figure 2-1 High-Level Block Diagram on Visual Localization

In the next two sections, the implementation of the steps that make up visual localization, namely key point extraction, descriptor computation, feature matching and pose estimation, are described in more detail.