Senior Research Fellow
Department of Statistical Science
University College London
Noteworthy News ‖ CV ‖ Email: try or ‖ Math Genealogy ‖ Follow:
|Algorithmic Learning Theory.   Machine Learning.   Mathematics.   Probability and Statistics.
|I work at the Department of Statistical Science at UCL, where I lead the DELTA group.
Mine is one of two departments of the upcoming Institute for Mathematical and Statistical Sciences, the other one being the Department of Mathematics.
My outstanding colleagues include no less than the likes of Ricardo Silva, Ioanna Manolopoulou, Codina Cotar, Sam Livingstone, Brieuc Lehmann, F-X Briol, Serge Guillas, Gianluca Baio, and many others from my home department (Statistical Science); and Andrea Macrina, Timo Betcke, Helen Wilson, and many others from our sister department (Mathematics). I have close connections and collaboration with outstanding colleagues across UCL, inlcuding the Centre for Artificial Intelligence, the Advanced Research Computing centre, and the Information, Inference and Machine Learning group. I have a network of connections at other UK institutions and abroad. I am a member of the European Laboratory for Learning and Intelligent Systems and the London Mathematical Society, and a fellow of the Royal Statistical Society and the Institute of Mathematics and its Applications.
My work is on machine learning research. This field is fascinating! One of the things I enjoy most about it being the confluence of maths and stats, and computer science experiments, to answer research questions. Besides statistical learning I am interested also in other learning frameworks such as online learning and reinforcement learning, and of course deep learning, which is quite popular these days. It looks that optimization is one pervasive theme across machine learning theory and practice, though it comes up in such a variety of flavours and colours that it isn't boring. It reminds of the least action principle of Maupertuis, saying that "everything happens as if some quantity was to be made as small as possible." (This principle has lead the optimists to believe that we live in the best possible world.) But just optimization doesn't quite do it for machine learning... to really be talking about learning one has to pay attention to generalization!
|I did research studies in statistical machine learning at the Department of Computer Science, University College London, sponsored by DeepMind. In parallel with these studies, I was affiliated with DeepMind as a research scientist intern, for three years.
Outstanding ML/AI people with whom I have worked at UCL Computer Science include David Barber, Marc Deisenroth, Mark Herbster, María Pérez-Ortiz, and John Shawe-Taylor.
Outstanding ML/AI people with whom I have worked at DeepMind include Marcus Hutter, Laurent Orseau, Ilja Kuzborskij, Csaba Szepesvári, András György from my own team; and many others from friend teams including Amal Rannen-Triki, Razvan Pascanu, Agnieszka Grabska-Barwińska, Thore Graepel, Arnaud Doucet, Benjamin Van Roy, and Geoffrey Irving.
|I spent a year at the Department of Computing Science, University of Alberta. During this time I started building my mental model of the machine learning field and fine-tuning my hyperparameters. My host was Rich Sutton in theory, but in practice I was developing my research plans together with Csaba Szepesvári.
|I spent some time with Mauricio Sacchi's group looking at problems related to seismic signal analysis. Before that I worked with Sasha Litvak and Nicole Tomczak-Jaegermann on the smallest singular value of a sparse random matrix, using methods from geometric functional analysis and probability. Even before that I worked with Byron Schmuland on reversibility of a Brownian motion with drift. As an undergrad, with Loretta Gasco I worked on a fun project about repeated two-player games with incomplete information on one side.
- Tighter risk certificates for (probabilistic) neural networks.
UCL Centre for AI.
- Statistical Learning Theory: A Hitchhiker's Guide.
NeurIPS 2018 Tutorial. (with J. Shawe-Taylor)
|Conference & Journal Papers
- I. Kuzborskij, Cs. Szepesvári, O. Rivasplata, A. Rannen-Triki, R. Pascanu,
On the Role of Optimization in Double Descent: A Least Squares Study.
- M. Pérez-Ortiz, O. Rivasplata, J. Shawe-Taylor, Cs. Szepesvári,
Tighter risk certificates for neural networks.
JMLR, 22, 227 (2021), 1-40.
revised PDF /
- M. Haddouche, B. Guedj, O. Rivasplata, J. Shawe-Taylor,
PAC-Bayes unleashed: generalisation bounds with unbounded losses.
Entropy, 23, 10 (2021).
- L. Orseau, M. Hutter, O. Rivasplata,
Logarithmic pruning is all you need.
NeurIPS 2020 .
- O. Rivasplata, I. Kuzborskij, Cs. Szepesvári, J. Shawe-Taylor,
PAC-Bayes analysis beyond the usual bounds.
- O. Rivasplata, E. Parrado-Hernández, J. Shawe-Taylor, S. Sun, Cs. Szepesvári,
PAC-Bayes bounds for stable algorithms with instance-dependent priors.
- A.E. Litvak, O. Rivasplata,
Smallest singular value of sparse random matrices.
Studia Math., 212, 3 (2012), 195-218.
- O. Rivasplata, J. Rychtar, B. Schmuland,
Reversibility for diffusions via quasi-invariance.
Acta Univ. Carolin. Math. Phys., 48, 1 (2007), 3-10.
- O. Rivasplata, J. Rychtar, C. Sykes,
Evolutionary games in finite populations.
Pro Mathematica, 20, 39/40 (2006), 147-164.
- O. Rivasplata, B. Schmuland,
Invariant and reversible measures for random walks on Z.
Pro Mathematica, 19, 37/38 (2005), 117-124.
- O. Rivasplata,
A note on a confidence bound of Kuzborskij and Szepesvári.
- O. Rivasplata,
Subgaussian random variables: An expository note.
- O. Rivasplata, V. Tankasali, Cs. Szepesvári,
PAC-Bayes with Backprop.
|Probability Links (probably accessible)
|Stats Links (most significant)
|My birth town is Trujillo, the marinera dance town.
|Sometimes people ask me about Machu Picchu, it's a great place to see.
|They ask me less about Arequipa though it is also a great place to visit.
|Last link, in case you care to know, is about Pisco.