Thursday December 2 at the University of Amsterdam
On Thursday December 2, the Social Science Section of the Dutch Statistical Society will be hosting our 2021 Fall meeting (‘najaarbijeenkomst’) at Utrecht of Amsterdam, which takes the form of an afternoon symposium (starting at 13:15). The topic of this symposium will be
Machine Learning Day
How machine learning can help the social sciences
in which recent methodological developments relating to machine learning will be discussed, as well as their relevance for and application in the Social Sciences. During the symposium international experts on the topic – including prof. dr. Felix Wichmann from University of Tübingen – will share their thoughts and contributions on this topic with you, and there will be room for discussion. With machine learning becoming of ever more importance, we believe that the symposium will be of interest to researchers throughout the Social Sciences and hope to welcome you there. More details about the program of the symposium (inclduing abstracts) can be found below. The Symposium will be held in room Doelenzaal (Singel 425, Amsterdam).
Attendance is free, but please register with M.L.Geelhoed@uu.nl before the 25th of November with subject ‘VvS-OR Machine Learning day’.
If you value our Fall meetings or other VvS-OR meetings, you can become a member of VvS-OR, see vvsor.nl/join.
All the best,
the board of the Social Science Section of the Dutch Statistical Society:
Jesper Tijmstra (Tilburg University)
Don van Ravenzwaaij (Groningen University)
Joost van Ginkel (Leiden University)
Lourens Waldorp (University of Amsterdam)
Rebecca Kuiper (Utrecht University)
Seeing the forest for the trees
Machine learning (ML) is by now a familiar buzzword, both within and outside the field of statistics. Opinions vary on what exactly is machine learning, and how it distinguishes itself from statistics. The author would like to avoid making the distinction altogether, and to start from the more familiar distinction between parametric and non-parametric methods. After a short introduction to ML, I will focus on one class of non-parametric methods for prediction: decision trees. Single decision trees are well known for their ease of interpretation. Techniques such as bagging, boosting and random forests, in which the predictions of a large number of trees are averaged, are well known for providing state-of-the-art predictive accuracy in many applications. These techniques will be used to discuss trade-offs between interpretability and predictive accuracy, and to illustrate the potential benefits of flexible non-parametric methods for applications in social sciences.
dr. Caspar J. van Lissa
Assistant professor of developmental data science,
FSW ambassador to the Open Science Community Utrecht
Utrecht University, dept. Methodology & Statistics
Closing the empirical cycle: Using machine learning for rigorous scientific exploration
Should applied scientists care about machine learning? What can data scientists learn from applied science? These are questions that Caspar van Lissa seeks to address. His Veni-funded research “Will the kids be alright?” applies modern machine learning methods to a classical pedagogical problem: Emotional problems in adolescence. Confirmatory, theory-testing research is now the dominant approach in the social sciences. This approach typically tests whether a specific theoretically relevant predictor, like attachment, has a significant effect. Machine learning, by contrast, offers an unprecedentedly effective way to conduct exploratory research. It can encompass all possible predictors, and rank their importance in predicting the outcome (e.g., emotional problems). The results have relevance for theory formation, as they may emphasize which theoretical factors are empirically most important, and can reveal blind spots of under-theorized factors. These improved theories may then be tested using confirmatory research. By combining both approaches, we close the “empirical cycle”.
Prof. Felix A. Wichmann,
University of Tübingen,
Faculty of Science
Neural Information Processing Group
Object recognition in man and machine
Convolutional neural networks (CNNs) have been proposed as computational
models for (rapid) human object recognition and the
(feedforward-component) of the primate ventral stream. The usefulness of
CNNs as such models obviously depends on the degree of similarity they
share with human visual processing and human visual behaviour. In my
talk I will discuss both similarities and differences between human
vision and currently used standard CNNs in terms of their robustness to
image distortions (out-of-distribution generalisation) and texture bias.
In addition, I will argue for the importance of comparing decision
makers on a trial-by-trial basis rather than simply on aggregated
performance measures if we want to understand to what degree different
decision makers may use similar strategies or internal representations.
Safe statistics: anytime-valid hypothesis tests and confidence sequences
Safe statistics is a new framework for collecting evidence for hypotheses, particularly suitable for online and sequential learning. Most currently available methods for hypothesis testing and parameter estimation should not be used in the online or sequential setting, as bounds on error probabilities are then not guaranteed. Safe statistics are feasible, easily implementable equivalents of classical methods that do offer these guarantees: Type-I error bounds and/or confidence intervals can be given at any point in time, under arbitrary stopping rules.
Within the framework, nonnegative random variables called E-variables represent evidence for a hypothesis in the data. To optimize the amount of evidence collected, the information-theoretic GRO-criterion is used for E-variable design. GRO stands for Growth Rate Optimal. These GRO E-variables are special Bayes factors, but with corresponding prior distributions that are sometimes quite different from what Bayesian machine learners or statisticians would normally use. These concepts can also be extended and inverted safe tests can be used to construct anytime-valid confidence sequences.
In this talk, the theoretical background of safe statistics and its foundations in machine learning will be discussed, and the principles will be illustrated with specific implementations for t-tests and a two-sample data stream setting.