Integrasi Komputer    
   
Daftar Isi
(Sebelumnya) Hiraganahistory (Unix) (Berikutnya)

Histogram

Histogram
Histogram of arrivals per minute.svg
First described byKarl Pearson
PurposeTo roughly assess the probability distribution of a given variable by depicting the frequencies of observations occurring in certain ranges of values

In statistics, a histogram is a graphical representation showing a visual impression of the distribution of data. It is an estimate of the probability distribution of a continuous variable and was first introduced by Karl Pearson.[1] A histogram consists of tabular frequencies, shown as adjacent rectangles, erected over discrete intervals (bins), with an area equal to the frequency of the observations in the interval. The height of a rectangle is also equal to the frequency density of the interval, i.e., the frequency divided by the width of the interval. The total area of the histogram is equal to the number of data. A histogram may also be normalized displaying relative frequencies. It then shows the proportion of cases that fall into each of several categories, with the total area equaling 1. The categories are usually specified as consecutive, non-overlapping intervals of a variable. The categories (intervals) must be adjacent, and often are chosen to be of the same size.[2] The rectangles of a histogram are drawn so that they touch each other to indicate that the original variable is continuous.[3]

Histograms are used to plot density of data, and often for density estimation: estimating the probability density function of the underlying variable. The total area of a histogram used for probability density is always normalized to 1. If the length of the intervals on the x-axis are all 1, then a histogram is identical to a relative frequency plot.

An alternative to the histogram is kernel density estimation, which uses a kernel to smooth samples. This will construct a smooth probability density function, which will in general more accurately reflect the underlying variable.

The histogram is one of the seven basic tools of quality control.[4]

Contents

Etymology

An example histogram of the heights of 31 Black Cherry trees.

The etymology of the word histogram is uncertain. Sometimes it is said to be derived from the Greek histos 'anything set upright' (as the masts of a ship, the bar of a loom, or the vertical bars of a histogram); and gramma 'drawing, record, writing'. It is also said that Karl Pearson, who introduced the term in 1891, derived the name from "historical diagram".[5]

Examples

The U.S. Census Bureau found that there were 124 million people who work outside of their homes.[6] Using their data on the time occupied by travel to work, Table 2 below shows the absolute number of people who responded with travel times "at least 15 but less than 20 minutes" is higher than the numbers for the categories above and below it. This is likely due to people rounding their reported journey time.[citation needed] The problem of reporting values as somewhat arbitrarily rounded numbers is a common phenomenon when collecting data from people.[citation needed]

Histogram of travel time (to work), US 2000 census. Area under the curve equals the total number of cases. This diagram uses Q/width from the table.
Data by absolute numbers
IntervalWidthQuantityQuantity/width
054180836
55136872737
105186183723
155196343926
205179813596
25571901438
305163693273
3553212642
4054122824
45159200613
60306461215
9060343557

This histogram shows the number of cases per unit interval so that the height of each bar is equal to the proportion of total people in the survey who fall into that category. The area under the curve represents the total number of cases (124 million). This type of histogram shows absolute numbers, with Q in thousands.

Histogram of travel time (to work), US 2000 census. Area under the curve equals 1. This diagram uses Q/total/width from the table.
Data by proportion
IntervalWidthQuantity (Q)Q/total/width
0541800.0067
55136870.0221
105186180.0300
155196340.0316
205179810.0290
25571900.0116
305163690.0264
35532120.0052
40541220.0066
451592000.0049
603064610.0017
906034350.0005

This histogram differs from the first only in the vertical scale. The height of each bar is the decimal percentage of the total that each category represents, and the total area of all the bars is equal to 1, the decimal equivalent of 100%. The curve displayed is a simple density estimate. This version shows proportions, and is also known as a unit area histogram.

In other words, a histogram represents a frequency distribution by means of rectangles whose widths represent class intervals and whose areas are proportional to the corresponding frequencies. The intervals are placed together in order to show that the data represented by the histogram, while exclusive, is also continuous. (E.g., in a histogram it is possible to have two connecting intervals of 10.5–20.5 and 20.5–33.5, but not two connecting intervals of 10.5–20.5 and 22.5–32.5. Empty intervals are represented as empty and not skipped.)[7]

Mathematical definition

An ordinary and a cumulative histogram of the same data. The data shown is a random sample of 10,000 points from a normal distribution with a mean of 0 and a standard deviation of 1.

In a more general mathematical sense, a histogram is a function mi that counts the number of observations that fall into each of the disjoint categories (known as bins), whereas the graph of a histogram is merely one way to represent a histogram. Thus, if we let n be the total number of observations and k be the total number of bins, the histogram mi meets the following conditions:

n = \sum_{i=1}^k{m_i}.

Cumulative histogram

A cumulative histogram is a mapping that counts the cumulative number of observations in all of the bins up to the specified bin. That is, the cumulative histogram Mi of a histogram mj is defined as:

M_i = \sum_{j=1}^i{m_j}.

Number of bins and width

There is no "best" number of bins, and different bin sizes can reveal different features of the data. Some theoreticians have attempted to determine an optimal number of bins, but these methods generally make strong assumptions about the shape of the distribution. Depending on the actual data distribution and the goals of the analysis, different bin widths may be appropriate, so experimentation is usually needed to determine an appropriate width. There are, however, various useful guidelines and rules of thumb.[8]

The number of bins k can be assigned directly or can be calculated from a suggested bin width h as:[citation needed]

k = \left \lceil \frac{\max x - \min x}{h} \right \rceil.

The braces indicate the ceiling function.

Sturges' formula

Sturges' formula[9] is derived from a binomial distribution and implicitly assumes an approximately normal distribution.

k = \lceil \log_2 n + 1 \rceil, \,

It implicitly bases the bin sizes on the range of the data and can perform poorly if n < 30.[citation needed] It may also perform poorly if the data are not normally distributed.

Doane's formula

Doane's formula[10] is a modification of Sturges' formula which attempts to improve its performance with non-normal data.

 k = 1 + log_e( n ) + log_e ( 1 + \hat a ( \frac { n } { 6 } )^{ 1 / 2 } )

where a is the estimated kurtosis of the distribution.

Scott's normal reference rule
[11]
h = \frac{3.5 \hat \sigma}{n^{1/3}},

where \hat \sigma is the sample standard deviation. Scott's normal reference rule is optimal for random samples of normally distributed data, in the sense that it minimizes the integrated mean squared error of the density estimate.[12]

Square-root choice
k = \sqrt{n}, \,

which takes the square root of the number of data points in the sample (used by Excel histograms and many others).[citation needed]

Freedman–Diaconis' choice

The Freedman–Diaconis rule is:[13][12]

h = 2 \frac{\operatorname{IQR}(x)}{n^{1/3}},

which is based on the interquartile range, denoted by IQR. It replaces 3.5σ of Scott's rule with 2 IQR, which is less sensitive than the standard deviation to outliers in data.

Choice based on minimization of an estimated L2 risk function
[14]
 \underset{h}{\operatorname{arg\,min}} \frac{ 2 \bar{m} - v } {h^2}

where \textstyle \bar{m} and \textstyle v are mean and biased variance of a histogram with bin-width \textstyle h, \textstyle \bar{m}=\frac{1}{k} \sum_{i=1}^{k}  m_i and \textstyle v= \frac{1}{k} \sum_{i=1}^{k} (m_i - \bar{m})^2 .

See also

  • Data binning
  • Freedman–Diaconis rule
  • Image histogram
  • Density estimation
  • Kernel density estimation, a smoother but more complex method of density estimation
  • Pareto chart
  • Seven Basic Tools of Quality
  • V-optimal histograms

References

  1. ^ Pearson, K. (1895). "Contributions to the Mathematical Theory of Evolution. II. Skew Variation in Homogeneous Material". Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences 186: 343–414. Bibcode:1895RSPTA.186..343P. doi:10.1098/rsta.1895.0010.  edit
  2. ^ Howitt, D. and Cramer, D. (2008) Statistics in Psychology. Prentice Hall
  3. ^ Charles Stangor (2011) "Research Methods For The Behavioral Sciences". Wadsworth, Cengage Learning. ISBN 9780840031976.
  4. ^ Nancy R. Tague (2004). "Seven Basic Quality Tools". The Quality Toolbox. Milwaukee, Wisconsin: American Society for Quality. p. 15. Retrieved 2010-02-05. 
  5. ^ M. Eileen Magnello (December 2006). "Karl Pearson and the Origins of Modern Statistics: An Elastician becomes a Statistician". The New Zealand Journal for the History and Philosophy of Science and Technology. 1 volume. OCLC 682200824. 
  6. ^ US 2000 census.
  7. ^ Dean, S., & Illowsky, B. (2009, February 19). Descriptive Statistics: Histogram. Retrieved from the Connexions Web site: http://cnx.org/content/m16298/1.11/
  8. ^ e.g. § 5.6 "Density Estimation", W. N. Venables and B. D. Ripley, Modern Applied Statistics with S (2002), Springer, 4th edition. ISBN 0-387-95457-0.
  9. ^ Sturges, H. A. (1926). "The choice of a class interval". J American Statistical Association: 65–66. JSTOR 2965501. 
  10. ^ Doane DP (1976) Aesthetic frequency classification. American Statistician, 30: 181–183
  11. ^ Scott, David W. (1979). "On optimal and data-based histograms". Biometrika 66 (3): 605–610. doi:10.1093/biomet/66.3.605. 
  12. ^ a b Scott, D. (1992). Multivariate Density Estimation: Theory, Practice, and Visualization. New York: John Wiley. 
  13. ^ Freedman, David; Diaconis, P. (1981). "On the histogram as a density estimator: L2 theory". Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete 57 (4): 453–476. doi:10.1007/BF01025868. 
  14. ^ Shimazaki, H.; Shinomoto, S. (2007). "A method for selecting the bin size of a time histogram". Neural Computation 19 (6): 1503–1527. doi:10.1162/neco.2007.19.6.1503. PMID 17444758. 

Further reading

  • Lancaster, H.O. An Introduction to Medical Statistics. John Wiley and Sons. 1974. ISBN 0-471-51250-8

External links

(Sebelumnya) Hiraganahistory (Unix) (Berikutnya)