Id
stringlengths 1
6
| PostTypeId
stringclasses 7
values | AcceptedAnswerId
stringlengths 1
6
⌀ | ParentId
stringlengths 1
6
⌀ | Score
stringlengths 1
4
| ViewCount
stringlengths 1
7
⌀ | Body
stringlengths 0
38.7k
| Title
stringlengths 15
150
⌀ | ContentLicense
stringclasses 3
values | FavoriteCount
stringclasses 3
values | CreationDate
stringlengths 23
23
| LastActivityDate
stringlengths 23
23
| LastEditDate
stringlengths 23
23
⌀ | LastEditorUserId
stringlengths 1
6
⌀ | OwnerUserId
stringlengths 1
6
⌀ | Tags
sequence |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
1 | 1 | 15 | null | 49 | 5466 | How should I elicit prior distributions from experts when fitting a Bayesian model?
| Eliciting priors from experts | CC BY-SA 2.5 | null | 2010-07-19T19:12:12.510 | 2020-11-05T09:44:51.710 | null | null | 8 | [
"bayesian",
"prior",
"elicitation"
] |
2 | 1 | 59 | null | 34 | 33671 | In many different statistical methods there is an "assumption of normality". What is "normality" and how do I know if there is normality?
| What is normality? | CC BY-SA 2.5 | null | 2010-07-19T19:12:57.157 | 2022-11-23T13:03:42.033 | 2010-08-07T17:56:44.800 | null | 24 | [
"distributions",
"normality-assumption"
] |
3 | 1 | 5 | null | 71 | 6650 | What are some valuable Statistical Analysis open source projects available right now?
Edit: as pointed out by Sharpie, valuable could mean helping you get things done faster or more cheaply.
| What are some valuable Statistical Analysis open source projects? | CC BY-SA 2.5 | null | 2010-07-19T19:13:28.577 | 2022-11-27T23:33:13.540 | 2011-02-12T05:50:03.667 | 183 | 18 | [
"software",
"open-source"
] |
4 | 1 | 135 | null | 23 | 45871 | I have two groups of data. Each with a different distribution of multiple variables. I'm trying to determine if these two groups' distributions are different in a statistically significant way. I have the data in both raw form and binned up in easier to deal with discrete categories with frequency counts in each.
What tests/procedures/methods should I use to determine whether or not these two groups are significantly different and how do I do that in SAS or R (or Orange)?
| Assessing the significance of differences in distributions | CC BY-SA 2.5 | null | 2010-07-19T19:13:31.617 | 2010-09-08T03:00:19.690 | null | null | 23 | [
"distributions",
"statistical-significance"
] |
5 | 2 | null | 3 | 90 | null | The R-project
[http://www.r-project.org/](http://www.r-project.org/)
R is valuable and significant because it was the first widely-accepted Open-Source alternative to big-box packages. It's mature, well supported, and a standard within many scientific communities.
- Some reasons why it is useful and valuable
- There are some nice tutorials here.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:14:43.050 | 2010-07-19T19:21:15.063 | 2010-07-19T19:21:15.063 | 23 | 23 | null |
6 | 1 | null | null | 486 | 173164 | Last year, I read a blog post from [Brendan O'Connor](http://anyall.org/) entitled ["Statistics vs. Machine Learning, fight!"](http://anyall.org/blog/2008/12/statistics-vs-machine-learning-fight/) that discussed some of the differences between the two fields. [Andrew Gelman responded favorably to this](http://andrewgelman.com/2008/12/machine_learnin/):
Simon Blomberg:
>
From R's fortunes
package: To paraphrase provocatively,
'machine learning is statistics minus
any checking of models and
assumptions'.
-- Brian D. Ripley (about the difference between machine learning
and statistics) useR! 2004, Vienna
(May 2004) :-) Season's Greetings!
Andrew Gelman:
>
In that case, maybe we should get rid
of checking of models and assumptions
more often. Then maybe we'd be able to
solve some of the problems that the
machine learning people can solve but
we can't!
There was also the ["Statistical Modeling: The Two Cultures" paper](http://projecteuclid.org/euclid.ss/1009213726) by Leo Breiman in 2001 which argued that statisticians rely too heavily on data modeling, and that machine learning techniques are making progress by instead relying on the predictive accuracy of models.
Has the statistics field changed over the last decade in response to these critiques? Do the two cultures still exist or has statistics grown to embrace machine learning techniques such as neural networks and support vector machines?
| The Two Cultures: statistics vs. machine learning? | CC BY-SA 3.0 | null | 2010-07-19T19:14:44.080 | 2021-01-19T17:59:15.653 | 2017-04-08T17:58:18.247 | 11887 | 5 | [
"machine-learning",
"pac-learning"
] |
7 | 1 | 18 | null | 103 | 42538 | I've been working on a new method for analyzing and parsing datasets to identify and isolate subgroups of a population without foreknowledge of any subgroup's characteristics. While the method works well enough with artificial data samples (i.e. datasets created specifically for the purpose of identifying and segregating subsets of the population), I'd like to try testing it with live data.
What I'm looking for is a freely available (i.e. non-confidential, non-proprietary) data source. Preferably one containing bimodal or multimodal distributions or being obviously comprised of multiple subsets that cannot be easily pulled apart via traditional means. Where would I go to find such information?
| Locating freely available data samples | CC BY-SA 2.5 | null | 2010-07-19T19:15:59.303 | 2022-11-30T05:37:29.877 | 2013-09-26T21:50:36.963 | 253 | 38 | [
"dataset",
"sample",
"population",
"teaching"
] |
9 | 2 | null | 3 | 15 | null | [Incanter](http://incanter.org/) is a Clojure-based, R-like platform (environment + libraries) for statistical computing and graphics.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:16:27.553 | 2010-07-19T19:16:27.553 | null | null | 50 | null |
10 | 1 | 1887 | null | 63 | 119038 | Many studies in the social sciences use Likert scales. When is it appropriate to use Likert data as ordinal and when is it appropriate to use it as interval data?
| Under what conditions should Likert scales be used as ordinal or interval data? | CC BY-SA 2.5 | null | 2010-07-19T19:17:47.537 | 2021-02-04T18:28:52.953 | 2017-01-26T11:52:53.043 | 28666 | 24 | [
"ordinal-data",
"likert",
"scales",
"measurement"
] |
11 | 1 | 1201 | null | 5 | 355 | Is there a good, modern treatment covering the various methods of multivariate interpolation, including which methodologies are typically best for particular types of problems? I'm interested in a solid statistical treatment including error estimates under various model assumptions.
An example:
[Shepard's method](http://en.wikipedia.org/wiki/Inverse_distance_weighting)
Say we're sampling from a multivariate normal distribution with unknown parameters. What can we say about the standard error of the interpolated estimates?
I was hoping for a pointer to a general survey addressing similar questions for the various types of multivariate interpolations in common use.
| Multivariate Interpolation Approaches | CC BY-SA 2.5 | null | 2010-07-19T19:18:30.810 | 2010-08-03T21:50:09.007 | 2010-07-28T07:58:52.320 | 34 | 34 | [
"multivariable",
"interpolation"
] |
12 | 2 | null | 7 | 24 | null | See my response to ["Datasets for Running Statistical Analysis on"](https://stackoverflow.com/questions/2252144/datasets-for-running-statistical-analysis-on/2252450#2252450) in reference to datasets in R.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:18:41.370 | 2010-07-19T19:18:41.370 | 2017-05-23T12:39:26.143 | -1 | 5 | null |
13 | 2 | null | 6 | 27 | null | Machine Learning seems to have its basis in the pragmatic - a Practical observation or simulation of reality. Even within statistics, mindless "checking of models and assumptions" can lead to discarding methods that are useful.
For example, years ago, the very first commercially available (and working) Bankruptcy model implemented by the credit bureaus was created through a plain old linear regression model targeting a 0-1 outcome. Technically, that's a bad approach, but practically, it worked.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:18:56.800 | 2010-07-19T19:18:56.800 | null | null | 23 | null |
14 | 2 | null | 3 | 6 | null | I second that Jay. Why is R valuable? Here's a short list of reasons. [http://www.inside-r.org/why-use-r](http://www.inside-r.org/why-use-r). Also check out [ggplot2](http://had.co.nz/ggplot2/) - a very nice graphics package for R. Some nice tutorials [here](http://gettinggeneticsdone.blogspot.com/search/label/ggplot2).
| null | CC BY-SA 2.5 | null | 2010-07-19T19:19:03.990 | 2010-07-19T19:19:03.990 | null | null | 36 | null |
15 | 2 | null | 1 | 23 | null | John Cook gives some interesting recommendations. Basically, get percentiles/quantiles (not means or obscure scale parameters!) from the experts, and fit them with the appropriate distribution.
[http://www.johndcook.com/blog/2010/01/31/parameters-from-percentiles/](http://www.johndcook.com/blog/2010/01/31/parameters-from-percentiles/)
| null | CC BY-SA 2.5 | null | 2010-07-19T19:19:46.160 | 2010-07-19T19:19:46.160 | null | null | 6 | null |
16 | 2 | null | 3 | 17 | null | Two projects spring to mind:
- Bugs - taking (some of) the pain out of Bayesian statistics. It allows the user to focus more on the model and a bit less on MCMC.
- Bioconductor - perhaps the most popular statistical tool in Bioinformatics. I know it's a R repository, but there are a large number of people who want to learn R, just for Bioconductor. The number of packages available for cutting edge analysis, make it second to none.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:22:31.190 | 2010-07-19T20:43:02.683 | 2010-07-19T20:43:02.683 | 8 | 8 | null |
17 | 1 | 29 | null | 12 | 1903 | I have four competing models which I use to predict a binary outcome variable (say, employment status after graduating, 1 = employed, 0 = not-employed) for n subjects. A natural metric of model performance is hit rate which is the percentage of correct predictions for each one of the models.
It seems to me that I cannot use ANOVA in this setting as the data violates the assumptions underlying ANOVA. Is there an equivalent procedure I could use instead of ANOVA in the above setting to test for the hypothesis that all four models are equally effective?
| How can I adapt ANOVA for binary data? | CC BY-SA 2.5 | null | 2010-07-19T19:24:12.187 | 2012-01-22T23:34:51.837 | 2012-01-22T23:34:51.837 | 7972 | null | [
"anova",
"chi-squared-test",
"generalized-linear-model"
] |
18 | 2 | null | 7 | 43 | null | Also see the UCI machine learning Data Repository.
[http://archive.ics.uci.edu/ml/](http://archive.ics.uci.edu/ml/)
| null | CC BY-SA 2.5 | null | 2010-07-19T19:24:18.580 | 2010-07-19T19:24:18.580 | null | null | 36 | null |
19 | 2 | null | 7 | 16 | null | [Gapminder](http://www.gapminder.org/data/) has a number (430 at the last look) of datasets, which may or may not be of use to you.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:24:21.327 | 2010-07-19T19:24:21.327 | null | null | 55 | null |
20 | 2 | null | 2 | 3 | null | The assumption of normality assumes your data is normally distributed (the bell curve, or gaussian distribution). You can check this by plotting the data or checking the measures for kurtosis (how sharp the peak is) and skewdness (?) (if more than half the data is on one side of the peak).
| null | CC BY-SA 2.5 | null | 2010-07-19T19:24:35.803 | 2010-07-19T19:24:35.803 | null | null | 37 | null |
21 | 1 | null | null | 6 | 304 | What are some of the ways to forecast demographic census with some validation and calibration techniques?
Some of the concerns:
- Census blocks vary in sizes as rural
areas are a lot larger than condensed
urban areas. Is there a need to account for the area size difference?
- if let's say I have census data
dating back to 4 - 5 census periods,
how far can i forecast it into the
future?
- if some of the census zone change
lightly in boundaries, how can i
account for that change?
- What are the methods to validate
census forecasts? for example, if i
have data for existing 5 census
periods, should I model the first 3
and test it on the latter two? or is
there another way?
- what's the state of practice in
forecasting census data, and what are
some of the state of the art methods?
| Forecasting demographic census | CC BY-SA 2.5 | null | 2010-07-19T19:24:36.303 | 2019-08-04T09:23:05.420 | 2019-08-04T09:23:05.420 | 11887 | 59 | [
"forecasting",
"population",
"demography",
"census"
] |
22 | 1 | null | null | 431 | 276770 | How would you describe in plain English the characteristics that distinguish Bayesian from Frequentist reasoning?
| Bayesian and frequentist reasoning in plain English | CC BY-SA 3.0 | null | 2010-07-19T19:25:39.467 | 2022-10-28T14:15:46.433 | 2011-10-04T07:05:14.067 | 930 | 66 | [
"bayesian",
"frequentist"
] |
23 | 1 | 91 | null | 27 | 105467 | How can I find the PDF (probability density function) of a distribution given the CDF (cumulative distribution function)?
| Finding the PDF given the CDF | CC BY-SA 2.5 | null | 2010-07-19T19:26:04.363 | 2019-11-25T03:17:45.387 | null | null | 69 | [
"distributions",
"density-function",
"cumulative-distribution-function"
] |
24 | 2 | null | 3 | 22 | null | For doing a variety of MCMC tasks in Python, there's [PyMC](https://www.pymc.io/welcome.html), which I've gotten quite a bit of use out of. I haven't run across anything that I can do in BUGS that I can't do in PyMC, and the way you specify models and bring in data seems to be a lot more intuitive to me.
| null | CC BY-SA 4.0 | null | 2010-07-19T19:26:13.693 | 2022-11-27T23:15:56.147 | 2022-11-27T23:15:56.147 | 362671 | 61 | null |
25 | 1 | 32 | null | 10 | 4041 | What modern tools (Windows-based) do you suggest for modeling financial time series?
| Tools for modeling financial time series | CC BY-SA 2.5 | null | 2010-07-19T19:27:13.503 | 2022-12-03T14:50:44.067 | 2010-07-26T23:38:29.430 | 69 | 69 | [
"modeling",
"time-series",
"finance",
"software"
] |
26 | 1 | 61 | null | 38 | 25091 | What is a standard deviation, how is it calculated and what is its use in statistics?
| What is a standard deviation? | CC BY-SA 3.0 | null | 2010-07-19T19:27:43.860 | 2017-11-12T21:26:05.097 | 2017-11-12T21:26:05.097 | 11887 | 75 | [
"standard-deviation"
] |
28 | 2 | null | 3 | 7 | null | [GSL](http://www.gnu.org/software/gsl/) for those of you who wish to program in C / C++ is a valuable resource as it provides several routines for random generators, linear algebra etc. While GSL is primarily available for Linux there are also ports for Windows (See: [this](https://web.archive.org/web/20110817071546/http://gladman.plushost.co.uk/oldsite/computing/gnu_scientific_library.php) and [this](http://david.geldreich.free.fr/dev.html)).
| null | CC BY-SA 4.0 | null | 2010-07-19T19:28:12.830 | 2022-11-27T23:14:52.687 | 2022-11-27T23:14:52.687 | 362671 | null | null |
29 | 2 | null | 17 | 8 | null | Contingency table (chi-square). Also Logistic Regression is your friend - use dummy variables.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:28:15.640 | 2010-07-19T19:28:15.640 | null | null | 36 | null |
30 | 1 | 55 | null | 16 | 1992 | Which methods are used for testing random variate generation algorithms?
| Testing random variate generation algorithms | CC BY-SA 2.5 | null | 2010-07-19T19:28:34.220 | 2022-12-05T08:26:28.510 | 2010-08-25T14:12:54.547 | 919 | 69 | [
"algorithms",
"hypothesis-testing",
"random-variable",
"random-generation"
] |
31 | 1 | null | null | 289 | 545327 | After taking a statistics course and then trying to help fellow students, I noticed one subject that inspires much head-desk banging is interpreting the results of statistical hypothesis tests. It seems that students easily learn how to perform the calculations required by a given test but get hung up on interpreting the results. Many computerized tools report test results in terms of "p values" or "t values".
How would you explain the following points to college students taking their first course in statistics:
- What does a "p-value" mean in relation to the hypothesis being tested? Are there cases when one should be looking for a high p-value or a low p-value?
- What is the relationship between a p-value and a t-value?
| What is the meaning of p values and t values in statistical tests? | CC BY-SA 3.0 | null | 2010-07-19T19:28:44.903 | 2023-01-27T12:12:00.777 | 2019-09-27T13:22:45.057 | 919 | 13 | [
"hypothesis-testing",
"p-value",
"interpretation",
"intuition",
"faq"
] |
32 | 2 | null | 25 | 16 | null | I recommend R (see [the time series view on CRAN](http://cran.r-project.org/web/views/TimeSeries.html)).
Some useful references:
- Econometrics in R, by Grant Farnsworth
- Multivariate time series modelling in R
| null | CC BY-SA 2.5 | null | 2010-07-19T19:29:06.527 | 2010-07-19T19:29:06.527 | 2017-05-23T12:39:26.167 | -1 | 5 | null |
33 | 1 | 49 | null | 5 | 2578 | What R packages should I install for seasonality analysis?
| R packages for seasonality analysis | CC BY-SA 2.5 | null | 2010-07-19T19:30:03.717 | 2022-11-24T14:17:48.967 | 2010-09-16T06:56:44.040 | null | 69 | [
"r",
"seasonality"
] |
35 | 1 | 72 | null | 18 | 5678 | I have a data set that I'd expect to follow a Poisson distribution, but it is overdispersed by about 3-fold. At the present, I'm modelling this overdispersion using something like the following code in R.
```
## assuming a median value of 1500
med = 1500
rawdist = rpois(1000000,med)
oDdist = rawDist + ((rawDist-med)*3)
```
Visually, this seems to fit my empirical data very well. If I'm happy with the fit, is there any reason that I should be doing something more complex, like using a [negative binomial distribution, as described here](http://en.wikipedia.org/wiki/Overdispersion#Poisson)? (If so, any pointers or links on doing so would be much appreciated).
Oh, and I'm aware that this creates a slightly jagged distribution (due to the multiplication by three), but that shouldn't matter for my application.
---
Update: For the sake of anyone else who searches and finds this question, here's a simple R function to model an overdispersed poisson using a negative binomial distribution. Set d to the desired mean/variance ratio:
```
rpois.od<-function (n, lambda,d=1) {
if (d==1)
rpois(n, lambda)
else
rnbinom(n, size=(lambda/(d-1)), mu=lambda)
}
```
(via the R mailing list: [https://stat.ethz.ch/pipermail/r-help/2002-June/022425.html](https://stat.ethz.ch/pipermail/r-help/2002-June/022425.html))
| Modelling a Poisson distribution with overdispersion | CC BY-SA 2.5 | null | 2010-07-19T19:30:30.390 | 2010-07-25T01:44:14.513 | 2010-07-25T01:44:14.513 | 54 | 54 | [
"distributions",
"modeling",
"poisson-distribution",
"overdispersion"
] |
36 | 1 | null | null | 83 | 389933 | There is an old saying: "Correlation does not mean causation". When I teach, I tend to use the following standard examples to illustrate this point:
- number of storks and birth rate in Denmark;
- number of priests in America and alcoholism;
- in the start of the 20th century it was noted that there was a strong correlation between 'Number of radios' and 'Number of people in Insane Asylums'
- and my favorite: pirates cause global warming.
However, I do not have any references for these examples and whilst amusing, they are obviously false.
Does anyone have any other good examples?
| Examples for teaching: Correlation does not mean causation | CC BY-SA 3.0 | null | 2010-07-19T19:31:47.363 | 2019-08-23T10:09:53.320 | 2014-05-23T06:31:12.967 | 43889 | 8 | [
"correlation",
"teaching"
] |
38 | 2 | null | 35 | 4 | null | If your mean value for the Poisson is 1500, then you're very close to a normal distribution; you might try using that as an approximation and then modelling the mean and variance separately.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:32:28.063 | 2010-07-19T19:32:28.063 | null | null | 61 | null |
39 | 1 | 721 | null | 7 | 246 | I'm looking for worked out solutions using Bayesian and/or logit analysis similar to a workbook or an annal.
The worked out problems could be of any field; however, I'm interested in urban planning / transportation related fields.
| Sample problems on logit modeling and Bayesian methods | CC BY-SA 2.5 | null | 2010-07-19T19:32:29.513 | 2010-07-27T05:40:44.950 | 2010-07-27T04:57:47.780 | 190 | 59 | [
"modeling",
"bayesian",
"logit",
"transportation"
] |
40 | 1 | 111 | null | 14 | 1003 | What algorithms are used in modern and good-quality random number generators?
| Pseudo-random number generation algorithms | CC BY-SA 2.5 | null | 2010-07-19T19:32:47.750 | 2014-01-28T07:46:08.953 | 2010-08-25T14:13:48.740 | 919 | 69 | [
"algorithms",
"random-variable",
"random-generation"
] |
41 | 2 | null | 26 | 12 | null | A quote from [Wikipedia](http://en.wikipedia.org/wiki/Standard_deviation).
>
It shows how much variation there is from the "average" (mean, or expected/budgeted value). A low standard deviation indicates that the data points tend to be very close to the mean, whereas high standard deviation indicates that the data is spread out over a large range of values.
| null | CC BY-SA 3.0 | null | 2010-07-19T19:33:13.037 | 2015-06-17T08:59:36.287 | 2015-06-17T08:59:36.287 | -1 | 83 | null |
42 | 2 | null | 3 | 15 | null | [Weka](http://www.cs.waikato.ac.nz/ml/weka) for data mining - contains many classification and clustering algorithms in Java.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:33:19.073 | 2010-07-19T19:33:19.073 | null | null | 80 | null |
43 | 2 | null | 25 | 7 | null | R is great, but I wouldn't really call it "windows based" :) That's like saying the cmd prompt is windows based. I guess it is technically in a window...
RapidMiner is far easier to use [1]. It's a free, open-source, multi-platform, GUI. Here's a video on time series forecasting:
[Financial Time Series Modelling - Part 1](https://web.archive.org/web/20130522020008/http://rapidminerresources.com/index.php?page=financial-time-series-modelling---part-1)
Also, don't forget to read:
[Forecasting Methods and Principles](https://www.researchgate.net/project/Forecasting-Methods-and-Principles-forecastingprinciplescom-forprincom)
[1] No, I don't work for them.
| null | CC BY-SA 4.0 | null | 2010-07-19T19:33:37.497 | 2022-12-03T14:50:44.067 | 2022-12-03T14:50:44.067 | 362671 | 74 | null |
44 | 1 | 89 | null | 9 | 778 | How would you explain data visualization and why it is important to a layman?
| Explain data visualization | CC BY-SA 2.5 | null | 2010-07-19T19:34:42.643 | 2013-07-27T00:28:49.377 | 2013-07-27T00:28:49.377 | 12786 | 68 | [
"data-visualization",
"intuition"
] |
45 | 2 | null | 40 | 6 | null | The [Mersenne Twister](http://en.wikipedia.org/wiki/Mersenne_twister) is one I've come across and used before now.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:34:44.033 | 2010-07-19T19:34:44.033 | null | null | 55 | null |
46 | 2 | null | 26 | 1 | null | A standard deviation is the square root of the second central moment of a distribution. A central moment is the expected difference from the expected value of the distribution. A first central moment would usually be 0, so we define a second central moment as the expected value of the squared distance of a random variable from its expected value.
To put it on a scale that is more in line with the original observations, we take the square root of that second central moment and call it the standard deviation.
Standard deviation is a property of a population. It measures how much average "dispersion" there is to that population. Are all the obsrvations clustered around the mean, or are they widely spread out?
To estimate the standard deviation of a population, we often calculate the standard deviation of a "sample" from that population. To do this, you take observations from that population, calculate a mean of those observations, and then calculate the square root of the average squared deviation from that "sample mean".
To get an unbiased estimator of the variance, you don't actually calculate the average squared deviation from the sample mean, but instead, you divide by (N-1) where N is the number of observations in your sample. Note that this "sample standard deviation" is not an unbiased estimator of the standard deviation, but the square of the "sample standard deviation" is an unbiased estimator of the variance of the population.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:35:04.827 | 2010-07-20T02:13:12.733 | 2010-07-20T02:13:12.733 | 62 | 62 | null |
47 | 1 | 268 | null | 9 | 1102 | I have a dataset of 130k internet users characterized by 4 variables describing users' number of sessions, locations visited, avg data download and session time aggregated from four months of activity.
Dataset is very heavy-tailed. For example third of users logged only once during four months, whereas six users had more than 1000 sessions.
I wanted to come up with a simple classification of users, preferably with indication of the most appropriate number of clusters.
Is there anything you could recommend as a solution?
| Clustering of large, heavy-tailed dataset | CC BY-SA 4.0 | null | 2010-07-19T19:36:12.140 | 2020-05-13T03:35:50.177 | 2020-05-13T03:35:50.177 | 11887 | 22 | [
"clustering",
"large-data",
"kurtosis"
] |
49 | 2 | null | 33 | 7 | null | You don't need to install any packages because this is possible with base-R functions. Have a look at [the arima function](http://www.stat.ucl.ac.be/ISdidactique/Rhelp/library/ts/html/arima.html).
This is a basic function of [Box-Jenkins analysis](http://en.wikipedia.org/wiki/Box%E2%80%93Jenkins), so you should consider reading one of the R time series text-books for an overview; my favorite is Shumway and Stoffer. "[Time Series Analysis and Its Applications: With R Examples](http://rads.stackoverflow.com/amzn/click/0387293175)".
| null | CC BY-SA 2.5 | null | 2010-07-19T19:36:52.403 | 2010-07-19T19:36:52.403 | null | null | 5 | null |
50 | 1 | 85 | null | 103 | 24363 | What do they mean when they say "random variable"?
| What is meant by a "random variable"? | CC BY-SA 2.5 | null | 2010-07-19T19:37:31.873 | 2023-01-02T16:45:45.633 | 2017-04-15T16:49:34.860 | 11887 | 62 | [
"mathematical-statistics",
"random-variable",
"intuition",
"definition"
] |
51 | 1 | 64 | null | 5 | 232 | Are there any objective methods of assessment or standardized tests available to measure the effectiveness of a software that does pattern recognition?
| Measuring the effectiveness of a pattern recognition software | CC BY-SA 2.5 | null | 2010-07-19T19:37:55.630 | 2010-07-19T20:12:16.453 | null | null | 68 | [
"pattern-recognition"
] |
53 | 1 | 78 | null | 207 | 193152 | What are the main differences between performing principal component analysis (PCA) on the correlation matrix and on the covariance matrix? Do they give the same results?
| PCA on correlation or covariance? | CC BY-SA 3.0 | null | 2010-07-19T19:39:08.483 | 2022-04-09T04:48:09.847 | 2018-07-16T19:41:56.540 | 3277 | 17 | [
"correlation",
"pca",
"covariance",
"factor-analysis"
] |
54 | 1 | 65 | null | 18 | 1072 | As I understand UK Schools teach that the Standard Deviation is found using:
![alt text](https://upload.wikimedia.org/math/e/e/4/ee485814ab9e19908f2b39d5d70406d5.png)
whereas US Schools teach:
![alt text](https://upload.wikimedia.org/math/8/3/a/83a7338b851dcaafaf5b64353af56596.png)
(at a basic level anyway).
This has caused a number of my students problems in the past as they have searched on the Internet, but found the wrong explanation.
Why the difference?
With simple datasets say 10 values, what degree of error will there be if the wrong method is applied (eg in an exam)?
| Why do US and UK Schools Teach Different methods of Calculating the Standard Deviation? | CC BY-SA 2.5 | null | 2010-07-19T19:41:19.367 | 2020-01-13T00:55:13.780 | 2017-03-09T17:30:35.957 | -1 | 55 | [
"standard-deviation",
"error",
"teaching",
"unbiased-estimator"
] |
55 | 2 | null | 30 | 12 | null | The [Diehard Test Suite](http://en.wikipedia.org/wiki/Diehard_tests) is something close to a Golden Standard for testing random number generators. It includes a number of tests where a good random number generator should produce result distributed according to some know distribution against which the outcome using the tested generator can then be compared.
EDIT
I have to update this since I was not exactly right:
Diehard might still be used a lot, but it is no longer maintained and not state-of-the-art anymore. NIST has come up with [a set of improved tests](http://csrc.nist.gov/groups/ST/toolkit/rng/index.html) since.
| null | CC BY-SA 3.0 | null | 2010-07-19T19:41:39.880 | 2011-05-12T18:38:27.547 | 2011-05-12T18:38:27.547 | 56 | 56 | null |
56 | 2 | null | 22 | 261 | null | Here is how I would explain the basic difference to my grandma:
I have misplaced my phone somewhere in the home. I can use the phone locator on the base of the instrument to locate the phone and when I press the phone locator the phone starts beeping.
Problem: Which area of my home should I search?
## Frequentist Reasoning
I can hear the phone beeping. I also have a mental model which helps me identify the area from which the sound is coming. Therefore, upon hearing the beep, I infer the area of my home I must search to locate the phone.
## Bayesian Reasoning
I can hear the phone beeping. Now, apart from a mental model which helps me identify the area from which the sound is coming from, I also know the locations where I have misplaced the phone in the past. So, I combine my inferences using the beeps and my prior information about the locations I have misplaced the phone in the past to identify an area I must search to locate the phone.
| null | CC BY-SA 3.0 | null | 2010-07-19T19:42:28.040 | 2016-01-16T19:14:26.703 | 2016-01-16T19:14:26.703 | 100906 | null | null |
57 | 2 | null | 50 | 5 | null | From [Wikipedia](http://en.wikipedia.org/wiki/Random_variable):
>
In mathematics (especially probability
theory and statistics), a random
variable (or stochastic variable) is
(in general) a measurable function
that maps a probability space into a
measurable space. Random variables
mapping all possible outcomes of an
event into the real numbers are
frequently studied in elementary
statistics and used in the sciences to
make predictions based on data
obtained from scientific experiments.
In addition to scientific
applications, random variables were
developed for the analysis of games of
chance and stochastic events. The
utility of random variables comes from
their ability to capture only the
mathematical properties necessary to
answer probabilistic questions.
From [cnx.org](http://cnx.org/content/m13418/latest/):
>
A random variable is a function, which assigns unique numerical values to all possible
outcomes of a random experiment under fixed conditions. A random variable is not a
variable but rather a function that maps events to numbers.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:42:34.670 | 2010-07-19T19:47:54.623 | 2010-07-19T19:47:54.623 | 69 | 69 | null |
58 | 1 | 6988 | null | 14 | 2857 | What is the back-propagation algorithm and how does it work?
| Can someone please explain the back-propagation algorithm? | CC BY-SA 2.5 | null | 2010-07-19T19:42:57.990 | 2013-04-10T12:09:31.350 | 2011-04-29T00:38:41.203 | 3911 | 68 | [
"algorithms",
"optimization",
"neural-networks"
] |
59 | 2 | null | 2 | 29 | null | The assumption of normality is just the supposition that the underlying [random variable](http://en.wikipedia.org/wiki/Random_variable) of interest is distributed [normally](http://en.wikipedia.org/wiki/Normal_distribution), or approximately so. Intuitively, normality may be understood as the result of the sum of a large number of independent random events.
More specifically, normal distributions are defined by the following function:
$$ f(x) =\frac{1}{\sqrt{2\pi\sigma^2}}e^{ -\frac{(x-\mu)^2}{2\sigma^2} },$$
where $\mu$ and $\sigma^2$ are the mean and the variance, respectively, and which appears as follows:
![alt text](https://upload.wikimedia.org/wikipedia/commons/thumb/7/74/Normal_Distribution_PDF.svg/300px-Normal_Distribution_PDF.svg.png)
This can be checked in [multiple ways](http://en.wikipedia.org/wiki/Normality_test), that may be more or less suited to your problem by its features, such as the size of n. Basically, they all test for features expected if the distribution were normal (e.g. expected [quantile distribution](http://en.wikipedia.org/wiki/Q-Q_plot)).
| null | CC BY-SA 4.0 | null | 2010-07-19T19:43:20.423 | 2022-11-23T12:58:30.760 | 2022-11-23T12:58:30.760 | 362671 | 39 | null |
60 | 2 | null | 47 | 1 | null | [K-Means clustering](http://en.wikipedia.org/wiki/K-means_algorithm) should work well for this type of problem. However, it does require that you specify the number of clusters in advance.
Given the nature of this data, however, you may be able to work with a [hierarchical clustering algorithm](http://en.wikipedia.org/wiki/Cluster_analysis#Hierarchical_clustering) instead. Since all 4 variables are most likely fairly highly correlated, you can most likely break out clusters, and stop when you reach a small enough distance between clusters. This may be a much simpler approach in this specific case, and allows you to determine "how many clusters" by just stopping as soon as you've broken your set into fine enough clusters.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:43:20.720 | 2010-07-19T19:43:20.720 | null | null | 41 | null |
61 | 2 | null | 26 | 38 | null | Standard deviation is a number that represents the "spread" or "dispersion" of a set of data. There are other measures for spread, such as range and variance.
Here are some example sets of data, and their standard deviations:
```
[1,1,1] standard deviation = 0 (there's no spread)
[-1,1,3] standard deviation = 1.6 (some spread)
[-99,1,101] standard deviation = 82 (big spead)
```
The above data sets have the same mean.
Deviation means "distance from the mean".
"Standard" here means "standardized", meaning the standard deviation and mean are in the same units, unlike variance.
For example, if the mean height is 2 meters, the standard deviation might be 0.3 meters, whereas the variance would be 0.09 meters squared.
It is convenient to know that [at least 75%](http://en.wikipedia.org/wiki/Chebyshev%27s_inequality) of the data points always lie within 2 standard deviations of the mean (or [around 95%](http://en.wikipedia.org/wiki/68-95-99.7_rule) if the distribution is Normal).
For example, if the mean is 100, and the standard deviation is 15, then at least 75% of the values are between 70 and 130.
If the distribution happens to be Normal, then 95% of the values are between 70 and 130.
Generally speaking, IQ test scores are normally distributed and have an average of 100. Someone who is "very bright" is two standard deviations above the mean, meaning an IQ test score of 130.
| null | CC BY-SA 3.0 | null | 2010-07-19T19:44:35.037 | 2013-01-23T09:37:19.680 | 2013-01-23T09:37:19.680 | 74 | 74 | null |
62 | 1 | 80 | null | 8 | 267 | With the recent FIFA world cup, I decided to have some fun and determine which months produced world cup football players. Turned out, most footballers in the 2010 world cup were born in the first half of the year.
Someone pointed out, that children born in the first half of the year had a physical advantage over others and hence "survivorship bias" was involved in the equation. Is this an accurate observation? Can someone please explain why he says that?
Also, when trying to understand the concept, I found most examples revolved around the financial sector. Are they any other everyday life examples explaining it?
| A case of survivorship bias? | CC BY-SA 4.0 | null | 2010-07-19T19:44:58.033 | 2022-02-11T14:03:16.890 | 2022-02-11T14:03:16.890 | 11887 | 58 | [
"bias"
] |
63 | 2 | null | 36 | 43 | null | It might be useful to explain that "causes" is an asymmetric relation (X causes Y is different from Y causes X), whereas "is correlated with" is a symmetric relation.
For instance, homeless population and crime rate might be correlated, in that both tend to be high or low in the same locations. It is equally valid to say that homelesss population is correlated with crime rate, or crime rate is correlated with homeless population. To say that crime causes homelessness, or homeless populations cause crime are different statements. And correlation does not imply that either is true. For instance, the underlying cause could be a 3rd variable such as drug abuse, or unemployment.
The mathematics of statistics is not good at identifying underlying causes, which requires some other form of judgement.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:45:19.780 | 2010-07-19T19:45:19.780 | null | null | 87 | null |
64 | 2 | null | 51 | 8 | null | Yes, there are many methods. You would need to specify which model you're using, because it can vary.
For instance, Some models will be compared based on the [AIC](http://en.wikipedia.org/wiki/Akaike_information_criterion) or [BIC](http://en.wikipedia.org/wiki/Bayesian_information_criterion) criteria. In other cases, one would look at the [MSE from cross validation](http://en.wikipedia.org/wiki/Cross-validation_%28statistics%29) (as, for instance, with a support vector machine).
- I recommend reading Pattern Recognition and Machine Learning by Christopher Bishop.
- This is also discussed in Chapter 5 on Credibility, and particularly section 5.5 "Comparing data mining methods" of Data Mining: Practical Machine Learning Tools and Techniques by Witten and Frank (which discusses Weka in detail).
- Lastly, you should also have a look at The Elements of Statistical Learning by Hastie, Tibshirani and Friedman which is available for free online.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:46:08.217 | 2010-07-19T20:12:16.453 | 2010-07-19T20:12:16.453 | 5 | 5 | null |
65 | 2 | null | 54 | 20 | null | The first formula is the population standard deviation and the second formula is the the sample standard deviation. The second formula is also related to the unbiased estimator of the variance - see [wikipedia](http://en.wikipedia.org/wiki/Variance#Population_variance_and_sample_variance) for further details.
I suppose (here) in the UK they don't make the distinction between sample and population at high school. They certainly don't touch concepts such as biased estimators.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:46:11.383 | 2010-07-19T20:44:53.427 | 2010-07-19T20:44:53.427 | 8 | 8 | null |
66 | 2 | null | 54 | 6 | null | This is [Bessel's Correction](http://en.wikipedia.org/wiki/Bessel%27s_correction). The US version is showing the formula for the sample standard deviation, where the UK version above is the standard deviation of the sample.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:46:11.993 | 2010-07-19T19:46:11.993 | null | null | 41 | null |
67 | 2 | null | 44 | 2 | null | From [Wikipedia](http://en.wikipedia.org/wiki/Data_visualization): Data visualization is the study of the visual representation of data, meaning "information which has been abstracted in some schematic form, including attributes or variables for the units of information"
Data viz is important for visualizing trends in data, telling a story - See [Minard's map of Napoleon's march](http://www.edwardtufte.com/tufte/posters) - possibly one of the best data graphics ever printed.
Also see any of Edward Tufte's books - especially [Visual Display of Quantitative Information.](http://rads.stackoverflow.com/amzn/click/0961392142)
| null | CC BY-SA 2.5 | null | 2010-07-19T19:47:16.577 | 2010-07-19T19:47:16.577 | null | null | 36 | null |
69 | 2 | null | 54 | 3 | null | Since N is the number of points in the data set, one could argue that by calculating the mean one has reduced the degree of freedom in the data set by one (since one introduced a dependency into the data set), so one should use N-1 when estimating the standard deviation from a data set for which one had to estimate the mean before.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:47:49.587 | 2010-07-19T19:47:49.587 | null | null | 56 | null |
70 | 2 | null | 7 | 19 | null | [World Bank](http://data.worldbank.org/data-catalog) offers quite a lot of interesting data and has been recently very active in developing nice [API](http://data.worldbank.org/developers/api-overview) for it.
Also, [commugrate](http://www.cs.purdue.edu/commugrate/data_access/all_data_sets.php) project has an interesting list available.
For US health related data head for [Health Indicators Warehouse](http://www.healthindicators.gov/).
Daniel Lemire's blog [points](http://lemire.me/blog/archives/2012/03/27/publicly-available-large-data-sets-for-database-research/) to few interesting examples (mostly tailored towards DB research) including Canadian Census 1880 and synoptic cloud reports.
And as for today (03/04/2012) [US 1940 census records](http://1940census.archives.gov/) are also available to download.
| null | CC BY-SA 3.0 | null | 2010-07-19T19:48:45.853 | 2012-04-03T09:51:07.860 | 2012-04-03T09:51:07.860 | 22 | 22 | null |
71 | 2 | null | 58 | 3 | null | It's an algorithm for training feedforward multilayer neural networks (multilayer perceptrons). There are several nice java applets around the web that illustrate what's happening, like this one: [http://neuron.eng.wayne.edu/bpFunctionApprox/bpFunctionApprox.html](http://neuron.eng.wayne.edu/bpFunctionApprox/bpFunctionApprox.html). Also, [Bishop's book on NNs](http://rads.stackoverflow.com/amzn/click/0198538642) is the standard desk reference for anything to do with NNs.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:50:33.480 | 2010-07-19T19:50:33.480 | null | null | 36 | null |
72 | 2 | null | 35 | 11 | null | for overdispersed poisson, use the negative binomial, which allows you to parameterize the variance as a function of the mean precisely. rnbinom(), etc. in R.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:51:05.117 | 2010-07-19T19:51:05.117 | null | null | 96 | null |
73 | 1 | 104362 | null | 28 | 11307 | Duplicate thread: [I just installed the latest version of R. What packages should I obtain?](https://stats.stackexchange.com/questions/1676/i-just-installed-the-latest-version-of-r-what-packages-should-i-obtain)
What are the [R](http://www.r-project.org/) packages you couldn't imagine your daily work with data?
Please list both general and specific tools.
UPDATE:
As for 24.10.10 `ggplot2` seems to be the winer with 7 votes.
Other packages mentioned more than one are:
- plyr - 4
- RODBC, RMySQL - 4
- sqldf - 3
- lattice - 2
- zoo - 2
- Hmisc/rms - 2
- Rcurl - 2
- XML - 2
Thanks all for your answers!
| What R packages do you find most useful in your daily work? | CC BY-SA 3.0 | null | 2010-07-19T19:51:32.587 | 2015-08-20T09:00:50.943 | 2017-04-13T12:44:54.643 | -1 | 22 | [
"r"
] |
74 | 2 | null | 6 | 78 | null | In such a discussion, I always recall the famous Ken Thompson quote
>
When in doubt, use brute force.
In this case, machine learning is a salvation when the assumptions are hard to catch; or at least it is much better than guessing them wrong.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:51:34.287 | 2010-07-19T19:51:34.287 | null | null | null | null |
75 | 1 | 94 | null | 5 | 1803 | I'm using [R](http://www.r-project.org/) and the manuals on the R site are really informative. However, I'd like to see some more examples and implementations with R which can help me develop my knowledge faster. Any suggestions?
| Where can I find useful R tutorials with various implementations? | CC BY-SA 2.5 | null | 2010-07-19T19:52:31.180 | 2018-09-26T16:45:54.767 | 2018-09-26T16:45:54.767 | 7290 | 69 | [
"r",
"references"
] |
76 | 2 | null | 73 | 23 | null | I use [plyr](http://cran.r-project.org/web/packages/plyr/index.html) and [ggplot2](http://cran.r-project.org/web/packages/ggplot2/index.html) the most on a daily basis.
I also rely heavily on time series packages; most especially, the [zoo](http://cran.r-project.org/web/packages/zoo/index.html) package.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:52:49.387 | 2010-07-19T19:52:49.387 | null | null | 5 | null |
77 | 2 | null | 36 | 27 | null |
- Sometimes correlation is enough. For example, in car insurance, male drivers are correlated with more accidents, so insurance companies charge them more. There is no way you could actually test this for causation. You cannot change the genders of the drivers experimentally. Google has made hundreds of billions of dollars not caring about causation.
- To find causation, you generally need experimental data, not observational data. Though, in economics, they often use observed "shocks" to the system to test for causation, like if a CEO dies suddenly and the stock price goes up, you can assume causation.
- Correlation is a necessary but not sufficient condition for causation. To show causation requires a counter-factual.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:54:03.253 | 2010-09-09T18:16:59.853 | 2010-09-09T18:16:59.853 | 74 | 74 | null |
78 | 2 | null | 53 | 169 | null | You tend to use the covariance matrix when the variable scales are similar and the correlation matrix when variables are on different scales.
Using the correlation matrix is equivalent to standardizing each of the variables (to mean 0 and standard deviation 1). In general, PCA with and without standardizing will give different results. Especially when the scales are different.
As an example, take a look at this R `heptathlon` data set. Some of the variables have an average value of about 1.8 (the high jump), whereas other variables (run 800m) are around 120.
```
library(HSAUR)
heptathlon[,-8] # look at heptathlon data (excluding 'score' variable)
```
This outputs:
```
hurdles highjump shot run200m longjump javelin run800m
Joyner-Kersee (USA) 12.69 1.86 15.80 22.56 7.27 45.66 128.51
John (GDR) 12.85 1.80 16.23 23.65 6.71 42.56 126.12
Behmer (GDR) 13.20 1.83 14.20 23.10 6.68 44.54 124.20
Sablovskaite (URS) 13.61 1.80 15.23 23.92 6.25 42.78 132.24
Choubenkova (URS) 13.51 1.74 14.76 23.93 6.32 47.46 127.90
...
```
Now let's do PCA on covariance and on correlation:
```
# scale=T bases the PCA on the correlation matrix
hep.PC.cor = prcomp(heptathlon[,-8], scale=TRUE)
hep.PC.cov = prcomp(heptathlon[,-8], scale=FALSE)
biplot(hep.PC.cov)
biplot(hep.PC.cor)
```
[](https://i.stack.imgur.com/4IwjG.png)
Notice that PCA on covariance is dominated by `run800m` and `javelin`: PC1 is almost equal to `run800m` (and explains $82\%$ of the variance) and PC2 is almost equal to `javelin` (together they explain $97\%$). PCA on correlation is much more informative and reveals some structure in the data and relationships between variables (but note that the explained variances drop to $64\%$ and $71\%$).
Notice also that the outlying individuals (in this data set) are outliers regardless of whether the covariance or correlation matrix is used.
| null | CC BY-SA 4.0 | null | 2010-07-19T19:54:38.710 | 2018-10-10T11:27:18.523 | 2018-10-10T11:27:18.523 | 28666 | 8 | null |
79 | 2 | null | 54 | 5 | null | I am not sure this is purely a US vs. British issue. The rest of this page is excerpted from a faq I wrote.([http://www.graphpad.com/faq/viewfaq.cfm?faq=1383](http://www.graphpad.com/faq/viewfaq.cfm?faq=1383)).
How to compute the SD with n-1 in the denominator
- Compute the square of the difference between each value and the sample mean.
- Add those values up.
- Divide the sum by n-1. The result is called the variance.
- Take the square root to obtain the Standard Deviation.
Why n-1?
Why divide by n-1 rather than n when computing a standard deviation? In step 1, you compute the difference between each value and the mean of those values. You don't know the true mean of the population; all you know is the mean of your sample. Except for the rare cases where the sample mean happens to equal the population mean, the data will be closer to the sample mean than it will be to the true population mean. So the value you compute in step 2 will probably be a bit smaller (and can't be larger) than what it would be if you used the true population mean in step 1. To make up for this, divide by n-1 rather than n.v This is called Bessel's correction.
But why n-1? If you knew the sample mean, and all but one of the values, you could calculate what that last value must be. Statisticians say there are n-1 degrees of freedom.
When should the SD be computed with a denominator of n instead of n-1?
Statistics books often show two equations to compute the SD, one using n, and the other using n-1, in the denominator. Some calculators have two buttons.
The n-1 equation is used in the common situation where you are analyzing a sample of data and wish to make more general conclusions. The SD computed this way (with n-1 in the denominator) is your best guess for the value of the SD in the overall population.
If you simply want to quantify the variation in a particular set of data, and don't plan to extrapolate to make wider conclusions, then you can compute the SD using n in the denominator. The resulting SD is the SD of those particular values. It makes no sense to compute the SD this way if you want to estimate the SD of the population from which those points were drawn. It only makes sense to use n in the denominator when there is no sampling from a population, there is no desire to make general conclusions.
The goal of science is almost always to generalize, so the equation with n in the denominator should not be used. The only example I can think of where it might make sense is in quantifying the variation among exam scores. But much better would be to show a scatterplot of every score, or a frequency distribution histogram.
| null | CC BY-SA 4.0 | null | 2010-07-19T19:56:04.120 | 2020-01-13T00:55:13.780 | 2020-01-13T00:55:13.780 | 25 | 25 | null |
80 | 2 | null | 62 | 11 | null | The basic idea behind this is that football clubs have an age cut-off when determining teams. In the league my children participate in the age restrictions states that children born after July 31st are placed on the younger team. This means that two children that are effectively the same age can be playing with two different age groups. The child born July 31st will be playing on the older team and theoretically be the youngest and smallest on the team and in the league. The child born on August 1st will be the oldest and largest child in the league and will be able to benefit from that.
The survivorship bias comes because competitive leagues will select the best players for their teams. The best players in childhood are often the older players since they have additional time for their bodies to mature. This means that otherwise acceptable younger players are not selected simply because of their age. Since they are not given the same opportunities as the older kids, they don’t develop the same skills and eventually drop out of competitive soccer.
If the cut-off for competitive soccer in enough countries is January 1st, that would support the phenomena you see. A similar phenomena has been observed in several other sports including baseball and ice hockey.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:56:43.807 | 2010-07-19T19:56:43.807 | null | null | 93 | null |
81 | 2 | null | 73 | 8 | null | I use the [xtable](http://cran.r-project.org/web/packages/xtable/index.html) package. The xtable package turns tables produced by R (in particular, the tables displaying the anova results) into LaTeX tables, to be included in an article.
| null | CC BY-SA 2.5 | null | 2010-07-19T19:58:56.157 | 2010-07-19T19:58:56.157 | null | null | 69 | null |
83 | 2 | null | 75 | 2 | null | R is designed around ideas such as "reproducible research" and "trustworthy software", as John Chambers says [in his excellent book "Software for Data Analysis: Programming with R"](http://books.google.com/books?id=UXneuOIvhEAC&printsec=frontcover).
One of the best ways to learn R is to look at the wealth of source code that available on [CRAN](http://cran.r-project.org/) (with 2461 packages and counting). Simple `install.packages`, load a `library()`, and start browsing the code.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:02:51.297 | 2010-07-19T20:02:51.297 | null | null | 5 | null |
84 | 2 | null | 44 | 6 | null | I would explain it to a layman as:
>
Data visualization is taking data, and making a picture out of it. This allows you to easily see and understand relationships within the data much more easily than just looking at the numbers.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:03:34.307 | 2010-07-19T20:03:34.307 | null | null | 41 | null |
85 | 2 | null | 50 | 48 | null | A random variable is a variable whose value depends on unknown events. We can summarize the unknown events as "state", and then the random variable is a function of the state.
Example:
Suppose we have three dice rolls ($D_{1}$,$D_{2}$,$D_{3}$). Then the state $S=(D_{1},D_{2},D_{3})$.
- One random variable $X$ is the number of 5s. This is:
$$ X=(D_{1}=5?)+(D_{2}=5?)+(D_{3}=5?)$$
- Another random variable $Y$ is the sum of the dice rolls. This is:
$$ Y=D_{1}+D_{2}+D_{3} $$
| null | CC BY-SA 3.0 | null | 2010-07-19T20:08:00.927 | 2016-09-19T08:54:59.927 | 2016-09-19T08:54:59.927 | 35989 | 87 | null |
86 | 2 | null | 50 | 17 | null | Unlike a regular variable, a random variable may not be substituted for a single, unchanging value. Rather statistical properties such as the distribution of the random variable may be stated. The distribution is a function that provides the probability the variable will take on a given value, or fall within a range given certain parameters such as the mean or standard deviation.
Random variables may be classified as discrete if the distribution describes values from a countable set, such as the integers. The other classification for a random variable is continuous and is used if the distribution covers values from an uncountable set such as the real numbers.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:08:37.010 | 2010-07-20T17:46:55.840 | 2010-07-20T17:46:55.840 | 13 | 13 | null |
89 | 2 | null | 44 | 7 | null | When I teach very basic statistics to Secondary School Students I talk about evolution and how we have evolved to spot patterns in pictures rather than lists of numbers and that data visualisation is one of the techniques we use to take advantage of this fact.
Plus I try to talk about recent news stories where statistical insight contradicts what the press is implying, making use of sites like [Gapminder](http://www.gapminder.org/) to find the representation before choosing the story.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:11:47.797 | 2010-07-19T20:11:47.797 | null | null | 55 | null |
90 | 2 | null | 75 | 4 | null | [R bloggers](http://www.r-bloggers.com/) has been steadily supplying me with a lot of good pragmatic content.
From the author:
```
R-Bloggers.com is a central hub (e.g: A blog aggregator) of content
collected from bloggers who write about R (in English).
The site will help R bloggers and users to connect and follow
the “R blogosphere”.
```
| null | CC BY-SA 2.5 | null | 2010-07-19T20:12:24.130 | 2010-07-19T20:12:24.130 | null | null | 22 | null |
91 | 2 | null | 23 | 26 | null | As user28 said in comments above, the pdf is the first derivative of the cdf for a continuous random variable, and the difference for a discrete random variable.
In the continuous case, wherever the cdf has a discontinuity the pdf has an atom. Dirac delta "functions" can be used to represent these atoms.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:15:54.823 | 2010-07-20T08:52:27.083 | 2010-07-20T08:52:27.083 | 87 | 87 | null |
93 | 1 | null | null | 7 | 481 | We're trying to use a Gaussian process to model h(t) -- the hazard function -- for a very small initial population, and then fit that using the available data. While this gives us nice plots for credible sets for h(t) and so on, it unfortunately is also just pushing the inference problem from h(t) to the covariance function of our process. Perhaps predictably, we have several reasonable and equally defensible guesses for this that all produce different result.
Has anyone run across any good approaches for addressing such a problem? Gaussian-process related or otherwise?
| Robust nonparametric estimation of hazard/survival functions based on low count data | CC BY-SA 2.5 | null | 2010-07-19T20:17:07.643 | 2010-09-16T12:33:50.390 | 2010-09-16T12:33:50.390 | null | 61 | [
"nonparametric",
"survival",
"hazard"
] |
94 | 2 | null | 75 | 5 | null | Quick R site is basic, but quite nice for start [http://www.statmethods.net/index.html](http://www.statmethods.net/index.html) .
| null | CC BY-SA 2.5 | null | 2010-07-19T20:18:24.520 | 2010-07-19T20:18:24.520 | null | null | null | null |
95 | 1 | null | null | 7 | 745 | I have been using various GARCH-based models to forecast volatility for various North American equities using historical daily data as inputs.
Asymmetric GARCH models are often cited as a modification of the basic GARCH model to account for the 'leverage effect' i.e. volatility tends to increase more after a negative return than a similarly sized positive return.
What kind of a difference would you expect to see between a standard GARCH and an asymmetric GARCH forecast for a broad-based equity index like the S&P 500 or the NASDAQ-100?
There is nothing particularly special about these two indices, but I think it is helpful to give something concrete to focus the discussion, as I am sure the effect would be different depending on the equities used.
| How Large a Difference Can Be Expected Between Standard GARCH and Asymmetric GARCH Volatility Forecasts? | CC BY-SA 2.5 | null | 2010-07-19T20:21:35.683 | 2012-07-26T19:47:59.577 | 2010-07-20T06:14:40.363 | 154 | 57 | [
"time-series",
"garch",
"volatility-forecasting",
"finance"
] |
96 | 2 | null | 75 | 4 | null | Another great resource is [the LearnR blog](http://learnr.wordpress.com/), which went through an extensive study of visualizations with lattice and ggplot2.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:22:44.807 | 2010-07-19T20:22:44.807 | null | null | 5 | null |
97 | 1 | 104 | null | 69 | 68448 | I have some [ordinal data](http://en.wikipedia.org/wiki/Ordinal_scale#Ordinal_scale) gained from survey questions. In my case they are [Likert style](http://en.wikipedia.org/wiki/Likert_scale) responses (Strongly Disagree-Disagree-Neutral-Agree-Strongly Agree). In my data they are coded as 1-5.
I don't think means would mean much here, so what basic summary statistics are considered usefull?
| What are good basic statistics to use for ordinal data? | CC BY-SA 2.5 | null | 2010-07-19T20:23:22.603 | 2021-04-04T16:52:54.747 | 2021-04-04T16:52:54.747 | 11887 | 114 | [
"descriptive-statistics",
"ordinal-data",
"likert",
"faq"
] |
98 | 2 | null | 1 | 20 | null | Eliciting priors is a tricky business.
[Statistical Methods for Eliciting Probability Distributions](http://www.stat.cmu.edu/tr/tr808/tr808.pdf) and [Eliciting Probability Distributions](http://www.jeremy-oakley.staff.shef.ac.uk/Oakley_elicitation.pdf) are quite good practical guides for prior elicitation. The process in both papers is outlined as follows:
- background and preparation;
- identifying and recruiting the expert(s);
- motivation and training the expert(s);
- structuring and decomposition (typically deciding precisely what variables should
be elicited, and how to elicit joint distributions in the multivariate case);
- the elicitation itself.
Of course, they also review how the elicitation results in information that may be fit to or otherwise define distributions (for instance, in the Bayesian context, [Beta distributions](http://en.wikipedia.org/wiki/Beta_distribution)), but quite importantly, they also address common pitfalls in modeling expert knowledge (anchoring, narrow and small-tailed distributions, etc.).
| null | CC BY-SA 2.5 | null | 2010-07-19T20:23:57.330 | 2010-07-19T20:56:56.330 | 2010-07-19T20:56:56.330 | 39 | 39 | null |
99 | 2 | null | 73 | 8 | null | [multicore](http://cran.r-project.org/web/packages/multicore/index.html) is quite nice for tool for making faster scripts faster.
[cacheSweave](http://cran.r-project.org/web/packages/cacheSweave/index.html) saves a lot of time when using `Sweave`.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:25:08.793 | 2010-07-19T20:25:08.793 | null | null | null | null |
100 | 1 | 110 | null | 7 | 9712 | I'd like to see the answer with qualitative view on the problem, not just definition. Examples and analogous from other areas of applied math also would be good.
I understand, my question is silly, but I can't find good and intuitive introduction textbook on signal processing — if someone would suggest one, I will be happy.
| What's the purpose of window function in spectral analysis? | CC BY-SA 3.0 | null | 2010-07-19T20:30:23.320 | 2018-02-19T08:31:21.727 | 2018-02-19T08:31:21.727 | 128677 | 117 | [
"signal-processing",
"spectral-analysis"
] |
101 | 2 | null | 31 | 189 | null |
## Understanding $p$-value
Suppose, that you want to test the hypothesis that the average height of male students at your University is $5$ ft $7$ inches. You collect heights of $100$ students selected at random and compute the sample mean (say it turns out to be $5$ ft $9$ inches). Using an appropriate formula/statistical routine you compute the $p$-value for your hypothesis and say it turns out to be $0.06$.
In order to interpret $p=0.06$ appropriately, we should keep several things in mind:
- The first step under classical hypothesis testing is the assumption that the hypothesis under consideration is true. (In our context, we assume that the true average height is $5$ ft $7$ inches.)
- Imagine doing the following calculation: Compute the probability that the sample mean is greater than $5$ ft $9$ inches assuming that our hypothesis is in fact correct (see point 1).
In other words, we want to know $$\mathrm{P}(\mathrm{Sample\: mean} \ge 5 \:\mathrm{ft} \:9 \:\mathrm{inches} \:|\: \mathrm{True\: value} = 5 \:\mathrm{ft}\: 7\: \mathrm{inches}).$$
The calculation in step 2 is what is called the $p$-value. Therefore, a $p$-value of $0.06$ would mean that if we were to repeat our experiment many, many times (each time we select $100$ students at random and compute the sample mean) then $6$ times out of $100$ we can expect to see a sample mean greater than or equal to $5$ ft $9$ inches.
Given the above understanding, should we still retain our assumption that our hypothesis is true (see step 1)? Well, a $p=0.06$ indicates that one of two things have happened:
- (A) Either our hypothesis is correct and an extremely unlikely event has occurred (e.g., all $100$ students are student athletes)
or
- (B) Our assumption is incorrect and the sample we have obtained is not that unusual.
The traditional way to choose between (A) and (B) is to choose an arbitrary cut-off for $p$. We choose (A) if $p > 0.05$ and (B) if $p < 0.05$.
| null | CC BY-SA 3.0 | null | 2010-07-19T20:32:08.183 | 2015-09-14T16:14:44.860 | 2020-06-11T14:32:37.003 | -1 | null | null |
102 | 2 | null | 44 | 2 | null | For me [Illuminating the Path](http://nvac.pnl.gov/agenda.stm) report has been always good point of reference.
For more recent overview you can also have a look at good [article](http://queue.acm.org/detail.cfm?id=1805128) by Heer and colleagues.
But what would explain better than visualization itself?
![alt text](https://i.stack.imgur.com/uFvje.jpg)
([Source](http://blog.ffctn.com/what-is-data-visualization))
| null | CC BY-SA 2.5 | null | 2010-07-19T20:32:14.283 | 2010-11-23T20:42:01.350 | 2010-11-23T20:42:01.350 | 22 | 22 | null |
103 | 1 | null | null | 50 | 10817 | What is the best blog on data visualization?
I'm making this question a community wiki since it is highly subjective. Please limit each answer to one link.
---
Please note the following criteria for proposed answers:
>
[A]cceptable answers to questions like this ...need to supply adequate descriptions and reasoned justification. A mere hyperlink doesn't do it. ...[A]ny future replies [must] meet ...[these] standards; otherwise, they will be deleted without further comment.
| What is your favorite data visualization blog? | CC BY-SA 3.0 | null | 2010-07-19T20:33:26.713 | 2022-11-29T16:37:00.783 | 2016-10-13T15:18:16.230 | 28666 | 5 | [
"data-visualization",
"references"
] |
104 | 2 | null | 97 | 29 | null | A frequency table is a good place to start. You can do the count, and relative frequency for each level. Also, the total count, and number of missing values may be of use.
You can also use a contingency table to compare two variables at once. Can display using a mosaic plot too.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:34:46.820 | 2010-07-19T20:34:46.820 | null | null | 74 | null |
105 | 2 | null | 103 | 19 | null | [FlowingData](http://flowingdata.com/) | Data Visualization, Infographics, and Statistics
| null | CC BY-SA 3.0 | null | 2010-07-19T20:35:34.510 | 2012-10-24T14:50:25.800 | 2012-10-24T14:50:25.800 | 615 | 46 | null |
107 | 2 | null | 103 | 11 | null | [information aesthetics](https://web.archive.org/web/20120101181841/http://infosthetics.com/) - Data Visualization & Information Design
| null | CC BY-SA 4.0 | null | 2010-07-19T20:36:01.200 | 2022-11-29T16:27:06.847 | 2022-11-29T16:27:06.847 | 362671 | 46 | null |
108 | 2 | null | 103 | 17 | null | [Information Is Beautiful](https://informationisbeautiful.net/) | Ideas, issues, knowledge, data - visualized!
| null | CC BY-SA 4.0 | null | 2010-07-19T20:36:47.800 | 2022-11-29T16:27:52.770 | 2022-11-29T16:27:52.770 | 362671 | 5 | null |
109 | 1 | null | null | 5 | 11443 | Following one-way ANOVA, there are many possible follow-up multiple comparison tests. Holm's test (or better, the Holm-Sidak) test has lots of power, but because it works in a stepwise manner, it cannot compute confidence intervals. Its advantage over the tests than can compute confidence intervals (Tukey, Dunnett) is that is has more power. But is it fair to say that the Holm method always has more power than the methods of Tukey and Dunnet? Or does it depend...?
| Power of Holm's multiple comparison testing compared to others | CC BY-SA 2.5 | null | 2010-07-19T20:37:21.667 | 2010-08-04T14:38:22.450 | null | null | 25 | [
"multiple-comparisons",
"statistical-power"
] |
110 | 2 | null | 100 | 5 | null | It depends on where you apply the window function. If you do it in the time domain, it's because you only want to analyze the periodic behavior of the function in a short duration. You do this when you don't believe that your data is from a stationary process.
If you do it in the frequency domain, then you do it to isolate a specific set of frequencies for further analysis; you do this when you believe that (for instance) high-frequency components are spurious.
The first three chapters of "A Wavelet Tour of Signal Processing" by Stephane Mallat have an excellent introduction to signal processing in general, and chapter 4 goes into a very good discussion of windowing and time-frequency representations in both continuous and discrete time, along with a few worked-out examples.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:40:36.573 | 2010-07-19T20:40:36.573 | null | null | 61 | null |
111 | 2 | null | 40 | 11 | null | In R, the default setting for random number generation are:
- For U(0,1), use the Mersenne-Twister algorithm
- For Guassian numbers use the numerical inversion of the standard normal distribution function.
You can easily check this, viz.
```
> RNGkind()
[1] "Mersenne-Twister" "Inversion"
```
It is possible to change the default generator to other PRNGs, such as Super-Duper,Wichmann-Hill, Marsaglia-Multicarry or even a user-supplied PRNG. See the ?RNGkind for further details. I have never needed to change the default PRNG.
The [C GSL](http://www.gnu.org/software/gsl/manual/html_node/Random-number-environment-variables.html) library also uses the [Mersenne-Twister](http://www.gnu.org/software/gsl/manual/html_node/Random-number-generator-algorithms.html) by default.
| null | CC BY-SA 2.5 | null | 2010-07-19T20:41:24.917 | 2010-07-19T20:41:24.917 | null | null | 8 | null |
112 | 2 | null | 103 | 2 | null | I see all my favorite blogs have been listed. So I'll give you this one:
[I Love Charts](http://ilovecharts.tumblr.com/)
It's a bit light hearted.
| null | CC BY-SA 3.0 | null | 2010-07-19T20:52:27.263 | 2012-10-24T14:59:02.753 | 2012-10-24T14:59:02.753 | 615 | 12 | null |
Cross Validated / stats.stackexchange.com
Dataset Summary
This dataset contains all posts submitted to stats.stackexchange.com before the 30th of August 2023 formatted as Markdown text.
The data is sourced from Internet Archive StackExchange Data Dump and follows the format by mikex86/stackoverflow-posts
Dataset Structure
Each record corresponds to one post of a particular type.
Original ordering from the data dump is not exactly preserved due to parallelism in the script used to process the data dump.
The markdown content of each post is contained in the Body
field. The license for a particular post is contained in the ContentLicense
field.
Data Fields
{
Id: long,
PostTypeId: long, // 1=Question, 2=Answer, 3=Orphaned tag wiki, 4=Tag wiki excerpt, 5=Tag wiki, 6=Moderator nomination, 7=Wiki Placeholder, 8=Privilige Wiki
AcceptedAnswerId: long | null, // only present if PostTypeId=1
ParentId: long | null, // only present if PostTypeId=2
Score: long,
ViewCount: long | null,
Body: string | null,
Title: string | null,
ContentLicense: string | null,
FavoriteCount: long | null,
CreationDate: string | null,
LastActivityDate: string | null,
LastEditDate: string | null,
LastEditorUserId: long | null,
OwnerUserId: long | null,
Tags: array<string> | null
}
Also consider the StackExchange Datadump Schema Documentation, as all fields have analogs in the original dump format.
How to use?
from datasets import load_dataset
# predownload full dataset
ds = load_dataset('theblackcat102/crossvalidated-posts', split='train')
# dataset streaming (will only download the data as needed)
ds = load_dataset('theblackcat102/crossvalidated-posts', split='train', streaming=True)
for sample in iter(ds): print(sample["Body"])
How is the text stored?
The original Data Dump formats the "Body" field as HTML, using tags such as <code>
, <h1>
, <ul>
, etc.
This HTML format has been converted to Markdown following mikex86/stackoverflow-posts conversion rule.
Example:
After differencing I saw that my constant/intercept is not statistically significant. Does anybody know how to fit the same model without the const term?
im using statsmodels.tsa.arima.model
To give a relative example I have: ARIMA(data, order=(3,0,0))
an AR(3) model and say it that the second coefficient is insignificant. I can get rid of it by typing
ARMA(data,order=([1, 3], 0, 0)
but how can I get rid of coefficient??
- Downloads last month
- 33