The Treasury

Global Navigation

Personal tools

Treasury
Publication

Survey Reweighting for Tax Microsimulation Modelling - WP 03/17

4  Alternative Distance Functions

The chi-squared distance function is convenient because it enables an explicit solution for the calibrated weights to be obtained, requiring only matrix inversion. However, a modified form of the same approach can be applied to a range of alternative distance functions, as shown in this section. These functions belong to a class of functions having two features: the first derivative with respect to can be expressed as a function of and its inverse can be obtained explicitly. An interactive solution procedure is required for the calculation of the Lagrange multipliers. The general case of this class is presented in subsection 1. An iterative approach based on Newton’s method is described in subsection 2. Several weighting functions are described in subsection 3 and illustrated in subsection 4.

4.1  The general case

The Lagrangean for the general case, stated in section 2, was written as:

(13)    

Suppose that has the property, shared with the chi-square distance function, that the differential with respect to can be expressed as a function of the ratio , so that:

(14)    

The first-order conditions for minimisation can therefore be written as:

(15)    

Write the inverse function of as so that if say, then In the case of the chi-square distance function used above, , and the inverse takes a simple linear form. In general, from (15) the values of are expressed as:

(16)    

If the inverse function, can be obtained explicitly, equation (16) can be used to compute the calibrated weights, given a solution for the vector, .

As before, the Lagrange multipliers can be obtained by post-multiplying (16) by the vector , summing over all and using the calibration equations, so that:

(17)    

Finally, subtracting from both sides of (17) gives:

(18)    

The term is of course a scalar, and the left hand side is a known vector. In general, (18) is nonlinear in the vector and so must be solved using an iterative procedure, as described in the following subsection.

4.2  An iterative procedure

Writing the equations in (18) can be written as:

(19)    

for . The roots can be obtained using Newton’s method, described in the Appendix. This involves the following iterative sequence, where denotes the value of in the th iteration:[7]

(20)    

The Hessian matrix and the vector on the right hand side of (20) are evaluated using .

The elements are given by:

(21)    

which can be written as:

(22)     

Starting from arbitrary initial values, the matrix equation in (20) is used repeatedly to adjust the values until convergence is reached, where possible.

As mentioned earlier, the application of the approach requires that it is limited to distance functions for which the form of the inverse function, can be obtained explicitly, given the specification for . Hence, the Hessian can easily be evaluated at each step using an explicit expression for . As these expressions avoid the need for the numerical evaluation of and for each individual at each step, the calculation of the new weights can be expected to be relatively quick, even for large samples.[8] However, it must be borne in mind that a solution does not necessarily exist, depending on the distance function used and the adjustment required to the vector .

Notes

  • [7]The approach described here differs somewhat from other routines described in the literature, for example in Singh and Mohl (1996) and Vanderhoeft (2001). However, it provides extremely rapid convergence.
  • [8]Using numerical methods to solve for each and for , for every individual in each iteration, would increase the computational burden substantially.
Page top