The Kanade-Lucas-Tomasi (KLT) Feature Tracker algorithm estimates the 2D translation and scale changes of an image template between original template coordinates and a given reference image using the Inverse Compositional algorithm. For more information, see [1].
Inputs are an array of template bounding boxes, a translation and scale changes predictions array and a reference image. Additionally, a template image input is used to update template patches (see details below).
Outputs are the translation and scale changes estimation array from the input bounding box coordinates to the reference image coordinates and the template bounding box coordinates array in the reference image.
Tracking Result
Note
Video output requires HTML5-capable browser that supports H.264 mp4 video decoding.
Implementation
Each template bounding box defines a template image patch stored internally with the function descriptor. These template patches are tracked in reference images based on predicted translation and scale changes. An estimated translation and scale change from the original bounding box coordinates to reference image coordinates is computed. Each such estimation includes a tracking validity flag (tracking success or failure) and whether a template update is required, based on user-defined threshold parameters.
C API functions
For list of limitations, constraints and backends that implements the algorithm, consult reference documentation of the following functions:
(optional) Define a function to update the input bounding boxes and predictions for the next tracking given the output bounding boxes and estimations from the previous tracking. This custom update is similar to the default update, added here as an example on how to define a custom update.
# This function reads the results of the KLT tracker (output bounding boxes and estimations) and updates the input bounding boxes and predictions for the next iteration.
# If the template status is update needed, it updates the input bounding box with the corresponding output, making its prediction the identity matrix (fixing the bounding box).
if outBoxes_[i].template_status == vpi.KLTTemplateStatus.UPDATE_NEEDED:
inBoxes_[i] = outBoxes_[i]
inPreds_[i] = np.eye(3)
else:
# If the update is not needed, just update the input prediction by the corresponding output estimation.
Convert an input frame from OpenCV to a gray-scale OpenCV frame and a VPI image with the image data. Both are used in processing the input and output video frames.
def convertFrameImage(inputFrame):
if inputFrame.ndim == 3 and inputFrame.shape[2] == 3:
Define the input bounding boxes as a VPI Array of type vpi.Type.KLT_TRACKED_BOUNDING_BOX. Its capacity is the total number of bounding boxes to be tracked in all video frames. This is done to guarantee the maximum storage in case all bounding boxes throughout the whole video are tracked.
Read the first input frame from the input video and convert it to gray-scale OpenCV frame and VPI image.
validFrame, cvFrame = inVideo.read()
ifnot validFrame:
print("Error reading first input frame", file=sys.stderr)
exit(1)
# Convert OpenCV frame to gray returning also the VPI image
cvGray, imgTemplate = convertFrameImage(cvFrame)
Create the VPI KLTFeatureTracker object that will contain all the information needed by the algorithm. One such information is the input predictions, that can be retrieved from it via in_predictions() for other processing. The constructor receives the image template, the input bounding boxes, and the backend to execute the algorithm. It is assumed that all input frames have the same size, thus the image template may be the first frame.
At each valid frame, first check if the current frame is in allBoxes, a dictionary mapping frame indices to a list of input bounding boxes to start tracking at that frame. If yes, add these bounding boxes to the klt object.
while validFrame:
if curFrame in allBoxes:
klt.add_boxes(allBoxes[curFrame])
Read the next input frame from the input video and convert it to gray-scale OpenCV frame and VPI image reference.
curFrame += 1
validFrame, cvFrame = inVideo.read()
ifnot validFrame:
break
cvGray, imgReference = convertFrameImage(cvFrame)
Then execute the algorithm on the input image frame using the CUDA backend, defined in the klt creation. The update function passed is the one defined in the beginning. When this argument is not present, the klt call runs the default update. The output bounding boxes are returned with the tracked information.
outBoxes = klt(imgReference, update=customUpdate)
Initialization phase
Include the header that defines the needed functions and structures.
Declares functions that implement the KLT Feature Tracker algorithm.
Define the input frames and input bounding boxes. Refer to VPIBoundingBox documentation for instructions on how to properly fill each bounding box given an axis-aligned bounding box, the reference frames, the input boxes and input predictions.
Create the bounding box array with tracking information. For new bounding boxes, trackingStatus must be 0, indicating that bounding box tracking is valid. templateStatus must be 1, indicating that the template corresponding to this bounding box must be updated.
Create the bounding box transformation prediction array, initially filled with identity transforms, since the template matches exactly the bounding box contents in the template image.
Create the payload that will contain all temporary buffers needed for processing. It is assumed that all input frames have the same size, so the first frame dimensions and type are used to create the payload.
Create the output tracked bounding box array. It will contain the estimated current frame's bounding box based on previous frame and the template information gathered so far. It also contains the bounding box current tracking status.
Create the output estimated transforms. It will contain the transform that makes the bounding box template match the corresponding bounding box on the current (reference) frame.
Start of the processing loop from the second frame. The previous frame is where the algorithm fetches the tracked templates from, the current frame is where these templates are matched against.
for (int idframe = 1; idframe < frame_count; ++idframe)
Submit the algorithm. The first time it's run, it will go through all input bounding boxes, crop them from the template frame and store them in the payload. Subsequent runs will either repeat the cropping and storing process for new bounding boxes added (doesn't happen in this example, but happens in the sample application), or perform the template matching on the reference frame.
Lock the input arrays so that their state for the next iteration can be updated. Since they are actually wrappers, the wrapped data will be updated directly. In order to do that, the corresponding VPI array must be locked for writing.
Update bounding box statuses. If tracking was lost (trackingStatus==1), the input bounding box must also be marked as such, so subsequent KLT iterations ignore it. If the template needs to be updated (templateStatus==1), the next iteration will do the updating, or else it will perform the template matching.
If template for this bounding box must be updated in next KLT iteration, the user must re-define the bounding box. There are several ways to do it. One can use a feature detector such as Harris keypoint detector to help fetch a brand-new bounding box, use updated_bbox[b] and either refine it through other means to avoid accumulating tracking errors, or simply use it as-is, which is a less robust approach, but still yields decent results. This example chooses this last, simpler approach.
if (updated_bbox[b].templateStatus)
{
tracked_bboxes[b] = updated_bbox[b];
Also reset the corresponding input predicted transforms, setting it to identity, as it's now assumed that the input bounding box matches exactly the object being tracked.
Destroy a stream instance and deallocate all HW resources.
For more information, see KLT Feature Tracker in the "C API Reference" section of VPI - Vision Programming Interface.
References
Simon Baker, Iain Matthews, "Lucas-Kanade 20 Years On: A Unified Framework".
International Journal of Computer Vision, February 2004, Volume 56, issue 3, pp 221-255.