| HandoutNormalizing Variables
 This is a discussion of how to normalize (aka standardize) 
variables. The point of normalization is to make variables comparable to each 
other. The reason this is a problem is that measurements made using such scales 
of measurement as nominal, ordinal, interval and ratio are not unique. For 
example, you can measure temperature in both Fahrenheit and Centigrade. Both are 
valid, but they produce different numbers. If you want to know whether it is 
warmer in Seattle or Paris on a given day, and one is 68 degrees Fahrenheit and 
the other is 25 degrees Centigrade, you can't just say 68 is bigger than 25 so 
Seattle is warmer. Instead, you need to reduce the measurements to the same 
scale, and then compare. Normalization is the process of reducing measurements 
to a "neutral" or "standard" scale. Normalizing is done differently depending on the level of 
measurement of the variables, and is intimately related to the uniqueness 
properties of the measurement level. See the handout on
measurement theory for more information. Nominal Scale Variables 
        To compare two nominal variables that may be measured 
		using different scales, you would like to "normalize" the values so you 
		can see how well they correspond to each other. There is no simple normalization technique 
		to do this, but it can be done. 
      
        One approach: Construct a contingency table to cross-classify observations of 
        one variable against the other. Then, if all observations falling into a 
        given category or the row variable fall into just one category of the 
        column variable, you can establish a 1-1 mapping of one measurement to 
        the other. i.e., in variable A, a "2" corresponds to a "92" means in 
		variable BMore sophisticated approach: use a technique called 
		correspondence analysis (particularly a variant called optimal scaling) 
		to work out a set of scores that maximize correspondence Ordinal Scale 
        
        
          
            | Object | M1 | M2 | M3 |  
            | A | 22 | 0 | 99 |  
            | B | 22 | 0 | 99 |  
            | C | 22 | 0 | 99 |  
            | D | 23 | 1 | 150 |  
            | E | 24 | 67 | 152 |  
        To normalize an ordinal scale, you convert the values to rank order 
        values, for example, normalizing each of the scales above would yield: 
         
        
        
          
            | Object | M1* | M2* | M3* |  
            | A | 1 | 1 | 1 |  
            | B | 1 | 1 | 1 |  
            | C | 1 | 1 | 1 |  
            | D | 2 | 2 | 2 |  
            | E | 3 | 3 | 3 |  
        By normalizing variables, you can see whether a set of measured 
        variables are really measuring the same thing. i.e., you take away 
		numerical differences that are arbitrary (due to different measurement 
		properties) and leave only the differences that reflect differences in 
		the underlying property being measured.Note: we tend to use an asterisk after a variable 
		name to indicate the normalized version of the variable  Interval Scale 
        Uniqueness. Interval scales are unique up to a
        linear transformation (Y = mX+b). In 
        other words, if you measure a set of objects on an interval scale, and 
        then multiply and/or add a constant to each value, the resulting values 
        are equally as valid as the original values. This is because the ratios of 
        the intervals between the numbers are not affected by linear 
        transformations. The following measurements are equally valid: 
			
				
					| Object | M1 | M2 | M3 | M3 |  
					| A | 22 | 32 | 220 | 230 |  
					| B | 22 | 32 | 220 | 230 |  
					| C | 22 | 32 | 220 | 230 |  
					| D | 23 | 33 | 230 | 240 |  
					| E | 24 | 34 | 240 | 250 |  
        To normalize an interval scale, you perform a linear transformation 
        that creates a normalized version of the variable with the property that 
        the mean is zero and the standard deviation is one. This linear 
        transformation is called standardizing or reducing to z-scores. 
        Normalizing each of the variables above would yield: 
			
				
					| Object | M1 | M2 | M3 | M4 |  
					| A | -.75 | -.75 | -.75 | -.75 |  
					| B | -.75 | -.75 | -.75 | -.75 |  
					| C | -.75 | -.75 | -.75 | -.75 |  
					| D | 0.50 | 0.50 | 0.50 | 0.50 |  
					| E | 1.75 | 1.75 | 1.75 | 1.75 |  
        Note that all the values are the same -- this indicates that all 
        four columns are just linear transformations of each other and 
        therefore, from an interval scaling point of view, say exactly the same 
        thing.  Note all also that the standardized values can be interpreted as 
        (standard) deviations from the mean. D is just slightly above the mean of all 
        objects on this variable, while E is quite a bit higher than the mean. Ratio Scale 
        Uniqueness. Ratio scales are unique up to a congruence or 
        proportionality transformation (Y = mX). In other words, if you measure 
        a set of objects on a ratio scale, and then multiply each value by a 
        constant, the resulting values are equally as valid as the original 
        values. This is because the ratios of the intervals between the numbers 
        are not affected by congruence transformations. The measurements M1, M2 
        and M3 are equally valid measures of given object property, but M4 is 
        not measuring the same thing: 
			
				
					| Object | M1 | M2 | M3 | M4 |  
					| A | 22 | 220 | 11 | 12 |  
					| B | 22 | 220 | 11 | 12 |  
					| C | 22 | 220 | 11 | 12 |  
					| D | 23 | 230 | 11.5 | 13 |  
					| E | 24 | 240 | 12 | 14 |  
        To normalize a ratio scale, you perform a particular 
		"congruence" or 
		"similarity" 
        transformation that creates a normalized version of the variable with 
        the property that the length of the vector is 1 (i.e., the Euclidean or 
        L2 norm equals 1.0). In other words, to normalize a ratio-scaled 
		variable, we divide each value of the variable by the square root 
		of the sum of squares of all the original values. Normalizing each of the variables above would 
        yield: 
			
				
					
					
				
				
					| Object | M1 | M2 | M3 | M4 |  
					| A | 0.44 | 0.44 | 0.44 | 0.43 |  
					| B | 0.44 | 0.44 | 0.44 | 0.43 |  
					| C | 0.44 | 0.44 | 0.44 | 0.43 |  
					| D | 0.45 | 0.45 | 0.45 | 0.46 |  
					| E | 0.47 | 0.47 | 0.47 | 0.50 |  
        Note that all the values except the last column are the same -- this 
        indicates that the first three columns are just rescalings (in a ratio 
        sense) of each other and therefore, measure exactly the same thing. The 
        last column is different however, indicating that it measures something 
        else.  Note also that other ways of normalizing accomplish the same 
        goal of making different measurements comparable. So we could just 
        divide each column by the column sum, creating a new variable whose 
        values add to 1. This allows interpretation of the rescaled values as 
        proportions or shares of the whole. This is not the usual way but it 
		works fine. Difference Scale (aka Additive Scale) 
        Uniqueness. Additive scales are unique up to a "translation" 
        transformation (Y = X + b). In other words, if you measure a set 
        of objects on an additive scale, and then add a 
        constant to each value, the resulting values are equally as valid as the original 
        values. This is because the intervals between values are not affected by 
        translation transformations. The measurements M1 and M2 are equally valid measures of given object property, but M3 is 
        not measuring the same thing: 
			
				
					| Object | M1 | M2 | M3 |  
					| A | 22 | 12 | 11 |  
					| B | 22 | 12 | 11 |  
					| C | 22 | 12 | 11 |  
					| D | 23 | 13 | 11.5 |  
					| E | 25 | 15 | 12 |  
        To normalize an additive scale, you perform a particular translation transformation that creates a normalized version of the variable with 
        the property that the mean of the transformed vector is 0. To do this, 
		we just subtract the mean of the original values. Normalizing each of the variables above would 
        yield: 
			
				
					
					
				
				
					| Object | M1 | M2 | M3 |  
					| A | -0.8 | -0.8 | -0.3 |  
					| B | -0.8 | -0.8 | -0.3 |  
					| C | -0.8 | -0.8 | -0.3 |  
					| D | 0.2 | 0.2 | 0.2 |  
					| E | 2.2 | 2.2 | 0.7 |  
        Note that all the values except the last column are the same -- this 
		indicates that the first two columns are just rescalings of each other and therefore, say exactly the same thing. The 
        last column is different however, indicating that it measures something 
        else.  Note also that other ways of normalizing accomplish the same 
        goal of making different measurements comparable. So we could just 
        subtract the column sum from each value   Absolute Scale 
        Uniqueness. Absolute scales are unique up to an identity 
        transformation (Y = X). In other words, they are completely unique and 
        no (non-trivial) transformation of the numbers is permissible.  As a result of their uniqueness, no normalization of absolute-scaled 
		variables is needed (nor exists). |