Este trabalho do Ollin Boer Bohan é simplesmente fenomenal. E além de tudo tem o repositório no GitHub.
Abstract: Softmax GAN is a novel variant of Generative Adversarial Network (GAN). The key idea of Softmax GAN is to replace the classification loss in the original GAN with a softmax cross-entropy loss in the sample space of one single batch. In the adversarial learning of N real training samples and M generated samples, the target of discriminator training is to distribute all the probability mass to the real samples, each with probability 1M, and distribute zero probability to generated data. In the generator training phase, the target is to assign equal probability to all data points in the batch, each with probability 1M+N. While the original GAN is closely related to Noise Contrastive Estimation (NCE), we show that Softmax GAN is the Importance Sampling version of GAN. We futher demonstrate with experiments that this simple change stabilizes GAN training.
Esse post da Data Robot é um daqueles tipos de post que mostra muito como a evolução das plataformas de Big Data, aliado com um maior arsenal computacional e preditivo estão varrendo para baixo do tapete qualquer bullshit disfarçado com tecnicalidades em relação à Data Science.
Vou reproduzir na íntegra, pois vale a pena usar esse post quando você tiver que justificar a qualquer burocrata de números (não vou dar nome aos bois dado o butthurt que isso poderia causar) porque ninguém mais dá a mínima para P-Valor, testes de hipóteses, etc na era em que temos uma abundância de dados; e principalmente está havendo a morte da significância estatística.
“Underpinning many published scientific conclusions is the concept of ‘statistical significance,’ typically assessed with an index called the p-value. While the p-value can be a useful statistical measure, it is commonly misused and misinterpreted.” ASA Statement on Statistical Significance and p-Values
If you’ve ever heard the words “statistically significant” or “fail to reject,” then you are among the countless thousands who have been traumatized by an academic approach building predictive models. Unfortunately, I can’t claim innocence in this matter. I taught statistics when I was in grad school, and I do have a Ph.D. in applied statistics. I was born into the world that uses formal hypothesis testing to justify every decision made in the model building process:
Should I include this variable in my model? How about an F-test?
Do my two samples have different means? Student’s t-test!
Are my variables correlated? How about a test using a Pearson Correlation Coefficient?
And on, and on, and on, and on…
These tests are all based on various theoretical assumptions. If the assumptions are valid, then they allegedly tell you whether or not your results are “statistically significant.”
Over the last century, as businesses and governments have begun to incorporate data science into their business processes, these “statistical tests” have also leaked into commercial and regulatory practices.
For instance, federal regulators in the banking industry issued this tortured guidance in 2011:
“… statistical tests depend on specific distributional assumptions and the purpose of the model… Any single test is rarely sufficient, so banks should apply a variety of tests to develop a sound model.”
In other words, statistical tests have lots of assumptions that are often (always) untrue, so use lots of them. (?!)
Here’s why statistical significance is a waste of time
If assumptions are invalid, the tests are invalid — even if your model is good
I developed a statistical test of my very own for my dissertation. The procedure for doing this is pretty simple. First, you make some assumptions about independence and data distributions, and variance, and so on. Then, you do some math that relies (heavily) on these assumptions in order to come up with a p-value. The p-value tells you what decision to make.
As an example, let’s take linear regression. Every business stats student memorizes the three assumptions associated with the p-values in this approach: independence (for which no real test exists), constant variance, and normality. If all these assumptions aren’t met, then none of the statistical tests that you might do are valid; yet regulators, professors, scientists, and statisticians all expect you to rely (heavily) on these tests.
What’s are you to do if your assumptions are invalid? In practice, the general practice is to wave your hands about “robustness” or some such thing and then continue along the same path.
If your data is big enough, EVERYTHING is significant
“The primary product of a research inquiry is one or more measures of effect size, not P values.” Jacob Cohen
As your data gets bigger and bigger (as data tends to do these days), everything becomes statistically significant. On one hand, this makes intuitive sense. For example, the larger a dataset is, the most likely an F-test is to tell you that your GLM coefficients are nonzero; i.e., larger datasets can support more complex models, as expected. On the other hand, for many assumption validity tests — e.g., tests for constant variance — statistical significance indicates invalid assumptions. So, for big datasets, you end up with tests telling you every feature is significant, but assumption tests telling you to throw out all of your results.
Validating assumptions is expensive and doesn’t add value
Nobody ever generated a single dollar of revenue by validating model assumptions (except of course the big consulting firms that are doing the work). No prospect was converted; no fraud was detected; no marketing message was honed by the drudgery of validating model assumptions. To make matters worse, it’s a never ending task. Every time a model is backtested, refreshed, or evaluated, the same assumption-validation-song-and-dance has to happen again. And that’s assuming that the dozens of validity tests don’t give you inconsistent results. It’s a gigantic waste of resources because there is a better way.
You can cheat, and nobody will ever know
Known as data dredging, data snooping, or p-hacking, it is very easy and relatively undetectable to manufacture statistically significant results. Andrew Gelman observed that most modelers have a (perverse) incentive to produce statistically significantresults — even at the expense of reality. It’s hardly surprising that these techniques exist, given the pressure to produce valuable data driven solutions. This risk, on its own, should be sufficient reason to abandon p-values entirely in some settings, like financial services, where cheating could result in serious consequences for the economy.
If the model is misspecified, then your p-values are likely to be misleading
Suppose you’re investigating whether or not a gender gap exists in America. Lots of things are correlated with gender; e.g., career choice, hours worked per week, percentage of vacation taken, participation in a STEM career, and so on. To the extent that any of these variables are excluded from your investigation — whether you know about them or not — the significance of gender will be overstated. In other words, statistical significance will give the impression that a gender gap exists, when it may not — simply due to model misspecification.
Only out-of-sample accuracy matters
Whether or not results are statistically significant is the wrong question. The only metric that actually matters when building models is whether or not your models can make accurate predictions on new data. Not only is this metric difficult to fake, but it also perfectly aligns with the business motivation for building the model in the first place. Fraud models that do a good job predicting fraud actually prevent losses. Underwriting models that accurately segment credit risk really do increase profits. Optimizing model accuracy instead of identifying statistical significance makes good business sense.
Over the course of the last few decades lots and lots of tools have been developed outside of the hypothesis testing framework. Cross-validation, partial dependence, feature importance, and boosting/bagging methods are just some of the tools in the machine learning toolbox. They provide a means not only for ensuring out-of-sample accuracy, but also understanding which features are important and how complex models work.
A survey of these methods is out of scope, but let me close with a final point. Unlike traditional statistical methods, tasks like cross-validation, model tuning, feature selection, and model selection are highly automatable. Custom coded solutions of any kind are inherently error prone, even for the most experienced data scientist
Many of the world’s biggest companies are recognizing that bespoke models, hand-built by Ph.D.’s are too slow and expensive to develop and maintain. Solutions like DataRobot provide a way for business experts to build predictive models in a safe, repeatable, systematic way that yields business value much more quickly and much cheaper than other approaches.
By Greg Michaelson, Director – DataRobot Labs
Abstract:In the contemporary information society, constructing an effective sales prediction model is challenging due to the sizeable amount of purchasing information obtained from diverse consumer preferences. Many empirical cases shown in the existing literature argue that the traditional forecasting methods, such as the index of smoothness, moving average, and time series, have lost their dominance of prediction accuracy when they are compared with modern forecasting approaches such as neural network (NN) and support vector machine (SVM) models. To verify these findings, this paper utilizes the Taiwanese cosmetic sales data to examine three forecasting models: i) the back propagation neural network (BPNN), ii) least-square support vector machine (LSSVM), and iii) auto regressive model (AR). The result concludes that the LS-SVM has the smallest mean absolute percent error (MAPE) and largest Pearson correlation coefficient ( R2 ) between model and predicted values.
Analytical challenges in multivariate data analysis and predictive modeling include identifying redundant and irrelevant variables. A recommended analytics approach is to first address the redundancy; which can be achieved by identifying groups of variables that are as correlated as possible among themselves and as uncorrelated as possible with other variable groups in the same data set. On the other hand, relevancy is about potential predictor variables and involves understanding the relationship between the target variable and input variables.
Multiple correspondence analysis (MCA) is a multivariate data analysis and data mining tool for finding and constructing a low-dimensional visual representation of variable associations among groups of categorical variables. Variable clustering as a tool for identifying redundancy is often applied to get a first impression of variable associations and multivariate data structure.
The motivations of this post are to illustrate the applications of: 1) preparing input variables for analysis and predictive modeling, 2) MCA as a multivariate exploratory data analysis and categorical data mining tool for business insights of customer churn data, and 3) variable clustering of categorical variables for the identification of redundant variables.
Agora o Matt Bogard do Econometric Sense dá a dica de como interpretar esse número:
From the basic probabilities above, we know that the probability of event Y is greater for males than females. The odds of event Y are also greater for males than females. These relationships are also reflected in the odds ratios. The odds of event Y for males is 3 times the odds of females. The odds of event Y for females are only .33 times the odds of males. In other words, the odds of event Y for males are greater and the odds of event Y for females is less.
This can also be seen from the formula for odds ratios. If the OR M vs F = odds(M)/odds(F), we can see that if the odds (M) > odds(F), the odds ratio will be greater than 1. Alternatively, for OR F vs M = odds(F)/odds(M), we can see that if the odds(F) < odds(M) then the ratio will be less than 1. If the odds for both groups are equal, the odds ratio will be 1 exactly.
RELATION TO LOGISTIC REGRESSION
Odds ratios can be obtained from logistic regression by exponentiating the coefficient or beta for a given explanatory variable. For categorical variables, the odds ratios are interpreted as above. For continuous variables, odds ratios are in terms of changes in odds as a result of a one-unit change in the variable.