Skip to content

Latest commit

 

History

History
62 lines (40 loc) · 3.05 KB

notes.md

File metadata and controls

62 lines (40 loc) · 3.05 KB

Recent Work

Methodology

From the research thus far it seems the simplist approach is to track vanishing points in the image and relate those to the relative rotation of the camera.

The intrinsic matrix K projects a point in camera coordinates onto the image plane. It requires three parameters:

  • focal length of camera
  • pixel width and height, the physical dimension of each pixel
  • the image center
  1. Solve for K using three orthogonal vanishing points
  2. Get rotation directly from vanishing points once K is known

Consider Convolutional Neural Networks (CNN)

A perspective transformation can be used to model yaw and pitch misalignments as the camera moves. The 3D perspective mapping can be modelled as a modified rotation matrix as seen in this paper as Equation (5). We can assume there is a perspective change across each consecutive frame, and that the roll is always 0 deg. because the camera is fixed to the vehicle. The proposed architecture would utilize a CNN to estimate the rotation transformation parameters:

  • This problem could be formulated as a classification task (0 - 360 deg.)
  • Separate layers for pitch and yaw angles?
  • Transfer learning with ResNet50 as shown in this article
  • Fine tune on labeled training data

This paper might be a helpful aid.