Saturday 20 July 2013

A Day With New Concepts of Statistics - 19th July , 2013


                              A Day With New Concepts of Statistics - 19th July , 2013                                             

Today's class began with the introduction of PERMAP. 
                                                            
PERMAP is a free, Windows-based, real-time interactive program for making perceptual maps (also called product maps, strategic maps, sociograms, sociometric maps, psychometric maps, stimulus-response maps, relationship maps, concept maps, etc.). Its fundamental purpose is to uncover any "hidden structure" that might be residing in a complex data set.

PERMAP is a program that uses multidimensional scaling (MDS) to reduce multiple pair wise relationships to 2-D pictures, commonly called perceptual maps. The Churchill data are in the form of correlation coefficients that show the relationships between 10 factors that influence the image of a department store. These correlation coefficients were calculated from responses to semantic differential scale questions given to a random selection of shoppers.

Purpose of PERMAP. The use of MDS for the construction of perceptual maps is well developed and several computer programs are available. In fact, MDS was one of the earliest uses of high-speed computers in psychology and the social sciences. The purpose of PERMAP is to provide a particularly convenient method of producing perceptual maps and to do so in a way that helps the researcher avoid a number of common mistakes.

Usefulness of perceptual maps:  A major advantage of MDS and perceptual maps is that they deal with problems associated with substantiating and communicating results based on data involving more than two dimensions. They discussed the importance of graphical communications and the role of the eye in interpreting and distinguishing object (factor, stimulus, characteristic) grouping.
Although experts may be able to extract the subtle relationships represented in a matrix of numbers, this skill is not widespread. Another important aspect of perceptual maps is that they are forgiving of missing or imprecise data points. Whereas some analytical techniques cannot tolerate missing elements in the input matrix, MDS results are often unaffected. This is because it is not uncommon for there to be much redundancy in the information given by a complete matrix of dissimilarities.

Existing perceptual mapping difficulties: Although the theory behind making perceptual maps is well developed, its application has been controversial There are four major concerns that PERMAP can help alleviate. These include avoiding local minima (i.e., configurations that are optimal with respect to small changes in configuration but not optimal with respect to all possible changes), proving complete convergence, minimizing the influence of outliers, and combining multiple correlation matrices. With care, batch-operated programs can be used in such a manner that all of these difficulties are properly addressed, but moving to a visually interactive program renders these difficulties easier to deal with.
PERMAP provides an interactive, visual system for the construction of perceptual maps from multidimensional dissimilarity data. It can treat up to 30 objects and can aggregate an unlimited number of matrices (cases) describing the pair wise differences or similarities among the objects. Aggregation can be accomplished using any of three methods, and the use of weighting factors is available.
PERMAP was designed to be simple and easy to use by a novice and to offer enough advanced features that it would be of value to the expert. Its major improvement over existing perceptual mapping programs is that it was designed specifically to combat certain common errors associated with multidimensional scaling. For instance, it is particularly effective at showing incomplete convergence, trapping by a local minima, and outlier influence. It is also effective at revealing the importance, or lack of importance, of the choice of the distance metric used in the objective function. Overall, the program provides a means for the researcher to go beyond just finding a solution to developing a feel for the suitability, stability, and variability of the solution.

Then We learned What is Z-Score and how to calculate it.

A Z-Score is a statistical measurement of a score's relationship to the mean in a group of scores. A Z-score of 0 means the score is the same as the mean. A Z-score can also be positive or negative, indicating whether it is above or below the mean and by how many standard deviations.
In addition to showing a score's relationship to the mean, the Z-score shows statisticians whether a score is typical or atypical for a particular data set. Z-scores also allow analysts to convert scores from different data sets into scores that can be accurately compared to each other. One real-life application of z-scores occurs in usability testing.
Z-Score is used for normalization.

If a Z-Score….

ü      Has a value of 0, it is equal to the group mean.
ü      Is positive, it is above the group mean.
ü      Is negative, it is below the group mean.
ü      Is equal to +1, it is 1 Standard Deviation above the mean.
ü      Is equal to +2, it is 2 Standard Deviations above the mean.
ü      Is equal to -1, it is 1 Standard Deviation below the mean.
ü      Is equal to -2, it is 2 Standard Deviations below the mean.

If you've encountered the z-score in a statistics book you usually get some formula like:

The above formula is for obtaining a z-score for an entire population. Usability testing obviously samples a very small subset of the population and thus the following formula is used:

Where x-bar and s are used as estimators for the population's true mean and standard deviation. Both formulas essentially calculate the same thing:

                                                                                                                                                                    

Calculating a Z-Score Example

For example, lets say you took the GRE a few weeks ago and got scores of 630 Verbal and 700 Quantitative. How good are these scores? Which is better, the Verbal or Quantitative score? Using a z-score can tell you how far you are from the mean and thus how well you performed. If you know the mean and standard deviations for a set of GRE test takers you can compare your scores.
ETS publishes the means and standard deviations of a set of test takers on the GRE website. 


VerbalQuantitative
Mean469591
StDev119148

By plugging in your scores you get the following:
Verbal z = (630 - 469) ÷ 119 = 1.35σ
Quantitative z = (700 - 591) ÷ 148 = .736σ
To convert these sigma values into a percentage you can look them up in a standard z-table, use the Excel formula =NORMSDIST(1.35) or use the Z-Score to Percentile Calculator (choose 1-sided) and get the percentages : 91% Verbal and 77% Quantitative. You can see where your score falls within the sample of other test takers and also see that the verbal score was better than the quantitative score. Assuming the sample data was normally distributed, here's how the scores would look graphically:

Figure 1: Verbal Score


Figure 2: Quantitative Score


Z-Scores and Process Sigma

An interactive Graph of the Standard Normal Curve similar to Figures 1 & 2 is available for you to visualize how the z-scores and the area under the normal curve correspond. The graphs also allow you to see the difference between one and two-sided (also called two-tailed) areas. In Six Sigma the process sigma metric is derived using the same method as a z-score. However, in Six Sigma you are measuring the distance a sample mean is above a specification limit--there can be an upper and lower spec limit that a sample must fall between as well. As in the z-score, you still use the same normal-deviates from the z-table to approximate the area under the curve. The process sigma metric is essentially a Z equivalent.
When testing software with users, task times are usually a good metric that will reveal the individual differences in performance. For task times there typically is only an upper spec limit. That is, it usually doesn't matter how fast a user completes a task, but it does matter if a user takes too long. For example, say you and your product team determined that a task should be completed in 120 seconds. 120 seconds becomes your Upper Spec Limit (USL). You sampled 10 users and got these task times:
Sample
100
99
101
125
100
123
96
90
98
116
USL: 120
Mean: 104
StDev: 12
To calculate the process sigma you subtract the mean (104) of the sample from the target (120) and divide by the sample standard deviation (12). For Sample 1 the process sigma is -1.32σ. The visual representation of the data can be seen below:


In the case of task times, a negative process sigma is ideal--as you want more people completing the task below the task time, not above it. You can simply drop the negative when communicating the results in the event it causes confusion. If you were to make radical improvements to the UI and then sampled another set of ten users, here are more results:

Sample 2
USL: 120
Mean: 75.8
StDev: 12.14


In the redesign, the average of the new sample is well below the spec limit and the process sigma is now very high. The corresponding defect area is now only .01% and the quality area is 99.98% 


Of course having users perform that much below the spec limit is not very common due to the inherent variability in user performance.         

Next lecture began with world95 file 

This file contained data about countries, their population, birth rate, daeth rate etc.
First, we made the histogram for Population of countries.


Then we transformed the population to ln(population) to get normal graph. It is done so that larger values become small.


Last Thing we learned Was Bubble Graph

bubble chart is a type of chart that displays three dimensions of data. Each entity with its triplet (v1v2v3) of associated data is plotted as a disk that expresses two of the vi values through the disk's xy location and the third through its size. Bubble charts can facilitate the understanding of social, economical, medical, and other scientific relationships.
Bubble charts can be considered a variation of the scatter plot, in which the data points are replaced with bubbles. 

When to use a Bubble chart

Bubble charts are often used to present financial data. Use a Bubble chart when you want specific values to be more visually represented in your chart by different bubble sizes. Bubble charts are useful when your worksheet has any of the following types of data:
  • Three values per data point     Three values are required for each bubble. These values can be in rows or columns on the worksheet, but they must be in the following order: x value, y value, and then size value.
  • Negative values     Bubble sizes can represent negative values, although negative bubbles do not display in the chart by default. You can choose to display them by formatting that data series. When they are displayed, bubbles with negative values are colored white (which cannot be modified) and the size is based on their absolute value. Even though the size of negative bubbles is based on a positive value, their data labels will show the true negative value.
  • Multiple data series     Plotting multiple data series in a Bubble chart (multiple bubble series) is similar to plotting multiple data series in a Scatter chart (multiple scatter series). While Scatter charts use a single set of x values and multiple sets of y values, Bubble charts use a single set of x values and multiple sets of both y values and size values.

Then we made bubble chart for world95 file used earlier.

We used this table to make the bubble chart and we made the bubble chart in excel.

                                      Number of people /          Birth rate per           Death rate per 
                                                        sq. kilometer                1000 people              1000 people
                                                                   
                                              Mean                         Mean                  Mean
Region or      OECD                               108.0                          13.0                    10
economic    East Europe                        76.7                          13.4                    11
group    Pacific/Asia                       802.5                          26.3                      9
              Africa                                 62.2                          42.0                    15
          Middle East                       126.7                          32.7                     6
          Latn America                         88.3                          26.9                     7





With this we ended the today's lecture.

Blog Written by: 

Prerna Bansal (2013209)

Team Members:

Neeraj Garg 
Pallavi Gupta
Piyush
Prerna Bansal
Priya Jain

             
     

No comments:

Post a Comment