Notes on Policy Gradient Methods

Notes on Policy Gradient Methods

Some notes on Chapter 13 - Policy Gradient Methods from Sutton & Barto’s book Reinforcement Learning: An Introduction in 2017.

1. The Framework of the RL Book

The book Reinforcement Learning: An Introduction provides a clear and simple account of the key ideas and algorithms of reinforcement learning. The discussion ranges from the history of the field’s intellectual foundations to the most recent developments and applications. A rough roadmap is as follow:

  • Tabular Solution Methods
    • A Single State Problem Example: Multi-armed Bandits
    • General Problem Formulation: Finite Markov Decision Processes
    • Solving Finite MDP Using:
      • Dynamic Programming
      • Monte Carlo Methods
      • Temporal-Difference Learning
    • Combining:
      • MC Methods & TD Learning
      • DP & TD Learning
  • Approximate Solution Methods
    • On-policy:
      • Value-based Methods
      • Policy-based Methods
    • Off-policy Methods
    • Eligibility Traces
    • Policy Gradient Methods

2. Notes on Policy Gradient Methods

2.1 What’s Policy Gradient?

Almost all the methods in the book have been action-value methods until chapter 13. In this chapter, a parameterized policy is learned, with or without learning a value function.

Policy Gradient is:

  • Methods that learn a parameterized policy that can select actions without consulting a value function.
  • The parameters are learned based on the gradient of some scalar performance measure J(θ) with respect to the policy parameter.
  • The general scheme of an update in policy gradient methods looks like this. It seeks to maximize performance.
θt+1=θt+αJ(θt)^\theta_{t+1} = \theta_t + \alpha \hat{\nabla J(\theta_t)}

There might or might not be a learned value function. If both the approximations of policy and value function are learned, then it’s called an actor-critic method. (‘critic’ is usually a state-value function)

EE5731 - Panoramic Image Stitching

EE5731 - Panoramic Image Stitching

SIFT is an algorithm from the paper Distinctive Image Features from Scale-Invariant Keypoints, published by David G. Lowe in 2004. The solution for one of its application, image stitching, is proposed in Automatic Panoramic Image Stitching using Invariant Features, by Matthew Brown and David G. Lowe in 2017. The continuous assessment 1 of Visual Computing is based on these two papers.

EE5731 - Viola-Jones Face Detection Algorithm

EE5731 - Viola-Jones Face Detection Algorithm

Viola-Jones algorithm is proposed on ICCV 2001 by, of course, Paul Viola and Michael Jones. Details and explanations of this algorithm can be found everywhere, and the module EE5731 also gives you step-by-step instructions. Even so, I would love to review the algorithm again for two reasons: It’s a good start for the millennials to have a view of CV 20 years ago. Also, there are some fun facts about the algorithm that are not covered in the lessons.

1. Viola-Jones Face Detection Algorithm

Face detection or object detection is a binary classification problem different from face recognition. Face detection can consider a substantial part of face recognition operations. Opposite to many of the existing algorithms using one single strong classifier, the Viola-Jones algorithm uses a set of weak classifiers, constructed by thresholding each of the Haar-like features. They can then be ranked and organized into a cascade.

The 3 main contributions in Viola-Jones face detection algorithm are integral image, AdaBoost-based learning and cascade classifier structure. The first step is to apply Haar-like feature extraction, which is an accelerated variant of image convolution using integral image. The lengthy features are then being selected by a learning algorithm based on AdaBoost. The cascade classifiers can drop the negative samples quickly and expedite the testing process enormously. Let’s have a quick go through towards the important points of the paper.

1.1 Feature Extraction: Haar-like Features

We can format this step as:

Input: a 24 × 24 image with zero mean and unit variance
Output: a 162336 × 1 scalar vector with its feature index ranging from 1 to 162336

This paper is a good example of combining classical CV with ML techniques. In the feature extraction step, CNN is not used. Instead, the intuitive Haar-like features are adopted.

Figure 1

The sum of the pixels which lie within the black regions is subtracted from the sum of pixels in the white regions. The mask set consists of 2 two-rectangle edge features, 2 three-rectangle line features, and a four-rectangle feature, which can be seen as a diagonal line feature. They are called Haar-like since the elements are all +1/-1, just like the square-shaped functions in Haar wavelet.

They are effective for tasks like face detection. Regions like eye sockets and cheeks usually have a distinctly bright and dark boundary, while nose and mouse can be considered as lines with a specific width.

Figure 2

The masks are applied over the sliding windows with a size of 24x24. The window is called a detector in the paper and 24x24 is the base solution of the detector. We’ll discuss how to get the detectors later. The 5 masks, starting from size 2x1, 3x1, 1x2, 1x3, 2x2, to the size of a detector, will slide over the whole window. For example, a 2x1 edge mask will move row by row, then extend to 4x1 and move all over the window again … until 24x1; then it starts over from 2x2 to 24x2, 2x3 to 24x3, 2x4 to 24x4 … till it became 24x24.

It’s time to answer why the length of the output descriptor is 162,336. A 24x24 image has 43,200, 27,600, 43,200, 27,600 and 20,736 features of category (a), (b), (c), (d) and (e) respectively, hence 162336 features in all.

Here is a MATLAB script for the calculation of the total number of features. Note that some would say the total number is 180,625. That’s the result when you use two four-rectangle features instead of one.

frameSize = 24;
features = 5;
% All five feature types:
feature = [[2,1]; [1,2]; [3,1]; [1,3]; [2,2]];
count = 0;
% Each feature:
for ii = 1:features
    sizeX = feature(ii,1);
    sizeY = feature(ii,2);
    % Each position:
    for x = 0:frameSize-sizeX
        for y = 0:frameSize-sizeY
            % Each size fitting within the frameSize:
            for width = sizeX:sizeX:frameSize-x
                for height = sizeY:sizeY:frameSize-y
                    count=count+1;
                end
            end
        end
    end
end
display(count)

Now we understand how to define and extract our features. But some detailed operations are still not covered, which made me confused when I first learned this part of the algorithm.

⚠️ Nitty-gritty alert!

Getting Detectors
So where do the detectors come from? A video I found gives the perfect demonstration.
The size of a box that might have a face inside varies a lot between different photos. (The size of the original photos in the discussion is 384x288.) So basically we can select the size of the box empirically (no smaller than 24x24) and then resize it to a 24x24 detector.
The paper also proposed a standardized scheme. We can scan the image with a 24x24 fixed size detector. Instead of enlarging the detector, the image is down-sampled by a factor of 1.25 before being scanned again. A pyramid of up to 11 scales of the image is used, each 1.25 times smaller than the last.
The moving step of the 24x24 detector is one pixel, horizontally and vertically. For the resized image, the equivalent moving step is larger than 1. So the down-sampled image pyramid can adjust the equivalent detector size and the moving step size at the same time.

Normalization before Convolution
The selected 24x24 detectors are first normalized before the convolution with the features. If the variance of a detector is lower than a specific threshold, then it must be plain and has little information of interest which can be left out of consideration immediately, saving some time processing the backgrounds.

Performance Issue
In fact, for better detection performance we can discard 3 features and only keep features (c) and (b). Like in Figure 2. Given the computational efficiency, the detection process can be completed for an entire image at every scale at up to 15 fps. Of course, with advanced machines 20 years later it’s better to utilize all the 5 Harr-like features.

1.2 Fasten Convolution Process: Integral Image

Input: a N × M image I
Output: a N × M integral image II

Remember the +1/-1 rectangles in Haar-like features? It’ll be tedious if we simply multiply the feature masks with the pixels in the corresponding region. Integral image is a smart way to replace the convolution with only a few additions.

Figure 3

In an integral image, each pixel at (x,y)(x,y) is the sum of all the pixels on the top-left: II(x,y)=i=1xj=1yI(i,j)II(x,y) = \sum_{i=1}^{x} \sum_{j=1}^{y} I(i,j)

This simplified representation with sums of rectangles can speed up the feature extraction significantly. Only 1+2+1+2+4=101+2+1+2+4=10 operations are needed to compute the convolution of 5 masks on one position.

1.3 Feature Selection: AdaBoost

Input: the descriptor xx and label yy for each training sample (about 5k positive and 5k negative)
Output: a strong classifier h(x)h(x) consists of TT weak classifiers ht(x)h_t(x) from TT features

This step is to train a classification function using the feature set (a set of Haar-like feature masks with different sizes, patterns, and locations all over the images) and a training set of positive and negative images (no less than 5,000 24x24 faces and 5,000 24x24 non-faces). The illustration below can help build a good understanding quickly.

Full-width image

Figure 4 AdaBoost Algorithm

A weak classifier is chosen on each iteration. And the falsely classified samples are given higher weights in the next iteration. The next classifier will pay extra attention to the misclassified training samples. Combining the weak classifiers, in this case, the low-level boundaries, we have a high-level boundary, which is considered as a strong classifier.

People always describe AdaBoost as a Forest of Stumps. The analogy just tells the exact nature of AdaBoost. AdaBoost means Adaptive Boosting. There are many boosting algorithms. Remember the random forest binary classifier? Each of the trees is built with random samples and several features from the full dataset. Then the new sample is fed into each decision tree, being classified, and collect the labels to count the final result. During the training process, the order doesn’t matter at all, and all the trees are equal. In the forest of stumps, all the trees are 1-level, calling stumps (only root and 2 leaves in binary classification problems). Each stump is a weak classifier using only 1 feature from the extracted features. First, we find the best stump with the lowest error. It will suck. A weak classifier is only expected to function slightly better than randomly guessing. It may only classify the training data correctly 51% of the time. The next tree will then emphasize the misclassified samples and will be given a higher weight if the last tree doesn’t go well.

Figure 5

The algorithm from the paper can be summarized as above. It’s adequate for a basic implementation of the algorithm, like this project. Despite this, there are still some details that merit the discussion.

⚠️ Nitty-gritty alert!

Input: the ff-th feature of all nn samples, the weights of all samples Output: a weak classifier ht(x)h_t(x)

Now we know how to combine a series of weak classifiers to form a strong classifier. But the optimal decision stump search, that is, how to find the hj(x)h_j(x) in step 2 of the loop in AdaBoost algorithm shown in Figure 4, is always left out in some projects like this, and even in the paper itself.

Luckily, an explanation I found solved all my problems. The resources mentioned in this post are listed in the reference section.

A single decision stump is constructed with 4 parameters: threshold, toggle, error, and margin.

ParameterSymbolDefinition
Thresholdτ\tauThe boundary of decision.
ToggleT\TauDecide if it means +1 (positive) or -1 (negative) when >τ>\tau.
ErrorE\EpsilonThe classification error after threshold and toggle are set up.
MarginM\MuThe largest margin among the ascending feature values.

The exhaustive search algorithm of decision stumps takes the ff-th feature of all the nn training examples as input, and returns a decision stump’s 4-tuple {τ,T,E,M}\left\{ \tau, \Tau, \Epsilon, \Mu \right\}.

First, since we need to compute the margin M\Mu, the gap between each pair of adjacent feature values, the nn examples are rearranged in ascending order of feature ff. For the ff-th feature, we have xf1xf2xf3...xfnx_{f_1} \leq x_{f_2} \leq x_{f_3} \leq ... \leq x_{f_n}

xfnx_{f_n} is the largest value among the ff-th feature of all nn samples. It doesn’t mean the nn-th sample since they are sorted.

In the initialization step, τ\tau is set to be smaller than the smallest feature value xf1x_{f_1}. Margin M\Mu is set to 0 and error E\Epsilon is set to an arbitrary number larger than the upper bound of the empirical loss.

Inside the jj-th iteration, a new threshold τ\tau is computed, which is the mean of the adjacent values: τ^=xfj+xfj+12\hat{\tau} = \frac{x_{f_j} + x_{f_{j+1}}}{2}

And the margin is computed as: M^=xfj+1xfj\hat{\Mu} = x_{f_{j+1}} - x_{f_j}

We compute the error E^\hat{\Epsilon} by collecting the weight wfjw_{f_j} of the misclassified samples. Since the toggle is not decided yet, we have to calculate error+error_+ and errorerror_- for both situations. If error+<errorerror_+ < error_-, then we decide that toggle T^=+1\hat{\Tau} = +1, otherwise T^=1\hat{\Tau} = -1. Finally, the tuple of parameters with smallest error is kept as {τ,T,E,M}\left\{ \tau, \Tau, \Epsilon, \Mu \right\}.

Full-width image

Figure 6 Decision Stump by Exhaustive Search

1.3.2 Selecting the Best Stump

Input: dd stumps for dd given features
Output: the best decision stump

During each training round tt in AdaBoost, a weak classifier ht(x)h_t(x) is selected. For each feature, we utilize the method in 1.3.1 to decide on the best parameter set for the decision stump.

1tT1 \leq t \leq T, TT is the total number of weak classifiers, usually we let TT be the full length of the feature vector.

The best stump among all the stumps should have the lowest error E\Epsilon. We select the one with the largest margin M\Mu when the errors happen to be the same. It then becomes the tt-th weak classifier. The weights assigned to all nn samples will be adjusted, and thus the exhaustive search should be carried out again using the updated weights. The idea behind this part is quite straightforward.

Full-width image

Figure 7 Best Stump

Full-width image

Figure 8 AdaBoost Algorithm with Optimal ht(x)h_t(x)

1.4 Fasten Classification: Cascade Classifier

Input: samples and features
Output: a cascade classifier, each step consists of several weak classifiers

As described in the subtitle, the main goal of building the cascade is to accelerate the estimation when dealing with a new input image, say, 384x288 (preset in the paper). Most of the sub-windows used to detect potential faces will have negative output. We hope to drop out those with backgrounds and non-faces as soon as possible.

Some will call this an attentional cascade. We hope to pay more attention to those sub-windows that might have faces inside them.

Our goal can be summarized as:

  1. Weed out non-faces early on to reduce calculation.
  2. Pay more attention to those hard samples since negative samples are dropped in advance.

Figure 9

The decision cascade is formed using a series of stages, each consists of several weak classifiers. Recall the strong classifier we build in 1.3, h(x)h(x) is the sum of weighted votes of all the weak classifiers/stumps. This scheme is already sufficient to achieve face detection. Some projects just use this form of strong classifier instead of an attentional cascade. But the first goal reminds us that we shouldn’t use all TT weak classifiers to deal with every detector, which will be extremely time-consuming.

At each stage, or some may call it a layer or classifier, we can use only part of the weak classifiers, as long as a certain false positive rate and detection rate can be guaranteed. In the next stage, only the positive and false-positive samples (the hard samples) are used to train the classifier. All we need to do is set up the maximum acceptable false-positive rate and the minimum acceptable detection rate per layer, as well as the overall target false positive rate.

Figure 10

We iteratively add more features to train a layer so that we can achieve a higher detection rate and lower false-positive rate, till the requirements are satisfied. Then we move to the next layer. We obtain the final attentional cascade once the overall target false positive rate is low enough.

Hope my interpretation can help you understand the algorithm.

1.5 Additional Comments: AdaBoost in Paper Gestalt

Laziness is the mother of invention. Some people, mostly the CVPR reviewers, are trying to use AdaBoost to classify bad papers from good ones just by judging the layout and formatting of them. The algorithm from the paper Paper Gestalt can achieve a true negative rate of over 50% with a false negative rate of about 15%. The classification of a paper needs only 0.5s.

Full-width image

Figure 11 Training Procedure of Paper Gestalt

It’s very entertaining and also educational since it gives some comments on the characteristics of good and bad papers. It’s funny that in order to be classified as a qualified paper for CVPR, the authors intentionally added a set of Maxwell’s equations to it. Weird but effective!

Full-width image

Figure 12 A Good Paper

Full-width image

Figure 13 A Bad Paper

1.6 Reference

  1. Rapid Object Detection using a Boosted Cascade of Simple Features
  2. Haar-like feature - Wikipedia
  3. Viola–Jones object detection framework - Wikipedia
  4. Viola-Jones’ face detection claims 180k features - Stack Overflow
  5. An Analysis of the Viola-Jones Face Detection Algorithm
  6. AdaBoost, Clearly Explained - YouTube
  7. GitHub - Simon-Hohberg/Viola-Jones: Python implementation of the face detection algorithm by Paul Viola and Michael J. Jones
  8. Understanding and Implementing Viola-Jones (Part Two)
  9. Paper Gestalt

An Overview of EE5731 Visual Computing

An Overview of EE5731 Visual Computing

Visual Computing is a module I took this semester at NUS ECE, hosted by Assoc. Prof. Robby Tan. It covers some of the best-known classic CV algorithms.

These years, rapid progressions in Deep Learning have brought enormous improvements to vision-based applications. DL methods, mainly based on CNN frameworks, are constructed usually by training instead of the domain-specific designing and programming in the traditional CV algorithms. They are easy to build, often achieve better accuracy, and the trained models can be very lightweight. So why do we still need to learn about traditional CV methods?

Pagination