Header logo is


2019


no image
Convolutional neural networks: A magic bullet for gravitational-wave detection?

Gebhard, T., Kilbertus, N., Harry, I., Schölkopf, B.

Physical Review D, 100(6):063015, American Physical Society, September 2019 (article)

ei

link (url) DOI [BibTex]

2019


link (url) DOI [BibTex]


no image
Data scarcity, robustness and extreme multi-label classification

Babbar, R., Schölkopf, B.

Machine Learning, 108(8):1329-1351, September 2019, Special Issue of the ECML PKDD 2019 Journal Track (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
SPINDLE: End-to-end learning from EEG/EMG to extrapolate animal sleep scoring across experimental settings, labs and species

Miladinovic, D., Muheim, C., Bauer, S., Spinnler, A., Noain, D., Bandarabadi, M., Gallusser, B., Krummenacher, G., Baumann, C., Adamantidis, A., Brown, S. A., Buhmann, J. M.

PLOS Computational Biology, 15(4):1-30, Public Library of Science, April 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
A 32-channel multi-coil setup optimized for human brain shimming at 9.4T

Aghaeifar, A., Zhou, J., Heule, R., Tabibian, B., Schölkopf, B., Jia, F., Zaitsev, M., Scheffler, K.

Magnetic Resonance in Medicine, 2019, (Early View) (article)

ei

DOI [BibTex]

DOI [BibTex]


Multidimensional Contrast Limited Adaptive Histogram Equalization
Multidimensional Contrast Limited Adaptive Histogram Equalization

Stimper, V., Bauer, S., Ernstorfer, R., Schölkopf, B., Xian, R. P.

IEEE Access, 7, pages: 165437-165447, 2019 (article)

ei

arXiv link (url) DOI [BibTex]

arXiv link (url) DOI [BibTex]


no image
TD-regularized actor-critic methods

Parisi, S., Tangkaratt, V., Peters, J., Khan, M. E.

Machine Learning, 108(8):1467-1501, (Editors: Karsten Borgwardt, Po-Ling Loh, Evimaria Terzi, and Antti Ukkonen), 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Probabilistic solutions to ordinary differential equations as nonlinear Bayesian filtering: a new perspective

Tronarp, F., Kersting, H., Särkkä, S. H. P.

Statistics and Computing, 29(6):1297-1315, 2019 (article)

ei pn

DOI [BibTex]

DOI [BibTex]


Learning to Control Highly Accelerated Ballistic Movements on Muscular Robots
Learning to Control Highly Accelerated Ballistic Movements on Muscular Robots

Büchler, D., Calandra, R., Peters, J.

2019 (article) Submitted

Abstract
High-speed and high-acceleration movements are inherently hard to control. Applying learning to the control of such motions on anthropomorphic robot arms can improve the accuracy of the control but might damage the system. The inherent exploration of learning approaches can lead to instabilities and the robot reaching joint limits at high speeds. Having hardware that enables safe exploration of high-speed and high-acceleration movements is therefore desirable. To address this issue, we propose to use robots actuated by Pneumatic Artificial Muscles (PAMs). In this paper, we present a four degrees of freedom (DoFs) robot arm that reaches high joint angle accelerations of up to 28000 °/s^2 while avoiding dangerous joint limits thanks to the antagonistic actuation and limits on the air pressure ranges. With this robot arm, we are able to tune control parameters using Bayesian optimization directly on the hardware without additional safety considerations. The achieved tracking performance on a fast trajectory exceeds previous results on comparable PAM-driven robots. We also show that our system can be controlled well on slow trajectories with PID controllers due to careful construction considerations such as minimal bending of cables, lightweight kinematics and minimal contact between PAMs and PAMs with the links. Finally, we propose a novel technique to control the the co-contraction of antagonistic muscle pairs. Experimental results illustrate that choosing the optimal co-contraction level is vital to reach better tracking performance. Through the use of PAM-driven robots and learning, we do a small step towards the future development of robots capable of more human-like motions.

ei

Arxiv Video [BibTex]


no image
Robustifying Independent Component Analysis by Adjusting for Group-Wise Stationary Noise

Pfister*, N., Weichwald*, S., Bühlmann, P., Schölkopf, B.

Journal of Machine Learning Research, 20(147):1-50, 2019, *equal contribution (article)

ei

ArXiv Code Project page PDF link (url) Project Page Project Page [BibTex]

ArXiv Code Project page PDF link (url) Project Page Project Page [BibTex]


no image
Enhancing Human Learning via Spaced Repetition Optimization

Tabibian, B., Upadhyay, U., De, A., Zarezade, A., Schölkopf, B., Gomez Rodriguez, M.

Proceedings of the National Academy of Sciences, 116(10):3988-3993, National Academy of Sciences, 2019 (article)

ei

link (url) DOI Project Page Project Page [BibTex]

link (url) DOI Project Page Project Page [BibTex]


no image
Entropic Regularization of Markov Decision Processes

Belousov, B., Peters, J.

Entropy, 21(7):674, 2019 (article)

ei

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Searchers adjust their eye-movement dynamics to target characteristics in natural scenes

Rothkegel, L., Schütt, H., Trukenbrod, H., Wichmann, F. A., Engbert, R.

Scientific Reports, 9(1635), 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Spatial statistics for gaze patterns in scene viewing: Effects of repeated viewing

Trukenbrod, H. A., Barthelmé, S., Wichmann, F. A., Engbert, R.

Journal of Vision, 19(6):19, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Quantum mean embedding of probability distributions

Kübler, J. M., Muandet, K., Schölkopf, B.

Physical Review Research, 1(3):033159, American Physical Society, 2019 (article)

ei

link (url) DOI [BibTex]

link (url) DOI [BibTex]


no image
Co-Contraction facilitates Body Stiffness Modulation during Swimming with Sensory Feedback in a Soft Biorobotic Physical Model

Jusufi, A., Vogt, D., Wood, R. J.

Integrative and Comparative Biology, 59(Supplement 1):E116-E116, Society of Integrative and Comparative Biology, McLean, VA, 2019 (article)

bio

DOI [BibTex]

DOI [BibTex]


no image
Inferring causation from time series with perspectives in Earth system sciences

Runge, J., Bathiany, S., Bollt, E., Camps-Valls, G., Coumou, D., Deyle, E., Glymour, C., Kretschmer, M., Mahecha, M., Munoz-Mari, J., van Nes, E., Peters, J., Quax, R., Reichstein, M., Scheffer, M., Schölkopf, B., Spirtes, P., Sugihara, G., Sun, J., Zhang, K., Zscheischler, J.

Nature Communications, 10(2553), 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Analysis of cause-effect inference by comparing regression errors

Blöbaum, P., Janzing, D., Washio, T., Shimizu, S., Schölkopf, B.

PeerJ Computer Science, 5, pages: e169, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Learning Intention Aware Online Adaptation of Movement Primitives

Koert, D., Pajarinen, J., Schotschneider, A., Trick, S., Rothkopf, C., Peters, J.

IEEE Robotics and Automation Letters, 4(4):3719-3726, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Spread-spectrum magnetic resonance imaging

Scheffler, K., Loktyushin, A., Bause, J., Aghaeifar, A., Steffen, T., Schölkopf, B.

Magnetic Resonance in Medicine, 82(3):877-885, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
How Cognitive Models of Human Body Experience Might Push Robotics

Schürmann, T., Mohler, B. J., Peters, J., Beckerle, P.

Frontiers in Neurorobotics, 13(14), 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Dense connectomic reconstruction in layer 4 of the somatosensory cortex

Motta, A., Berning, M., Boergens, K. M., Staffler, B., Beining, M., Loomba, S., Hennig, P., Wissler, H., Helmstaedter, M.

Science, 366(6469):eaay3134, American Association for the Advancement of Science, 2019 (article)

ei pn

DOI [BibTex]

DOI [BibTex]


no image
Learning Trajectory Distributions for Assisted Teleoperation and Path Planning

Ewerton, M., Arenz, O., Maeda, G., Koert, D., Kolev, Z., Takahashi, M., Peters, J.

Frontiers in Robotics and AI, 6, pages: 89, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Brainglance: Visualizing Group Level MRI Data at One Glance

Stelzer, J., Lacosse, E., Bause, J., Scheffler, K., Lohmann, G.

Frontiers in Neuroscience, 13(972), 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Eigendecompositions of Transfer Operators in Reproducing Kernel Hilbert Spaces

Klus, S., Schuster, I., Muandet, K.

Journal of Nonlinear Science, 2019, First Online: 21 August 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Workshops of the seventh international brain-computer interface meeting: not getting lost in translation

Huggins, J. E., Guger, C., Aarnoutse, E., Allison, B., Anderson, C. W., Bedrick, S., Besio, W., Chavarriaga, R., Collinger, J. L., Do, A. H., Herff, C., Hohmann, M., Kinsella, M., Lee, K., Lotte, F., Müller-Putz, G., Nijholt, A., Pels, E., Peters, B., Putze, F., Rupp, R. S. G., Scott, S., Tangermann, M., Tubig, P., Zander, T.

Brain-Computer Interfaces, 6(3):71-101, Taylor & Francis, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Compatible natural gradient policy search

Pajarinen, J., Thai, H. L., Akrour, R., Peters, J., Neumann, G.

Machine Learning, 108(8):1443-1466, (Editors: Karsten Borgwardt, Po-Ling Loh, Evimaria Terzi, and Antti Ukkonen), 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Learning stable and predictive structures in kinetic systems

Pfister, N., Bauer, S., Peters, J.

Proceedings of the National Academy of Sciences (PNAS), 116(51):25405-25411, 2019 (article)

ei

DOI [BibTex]

DOI [BibTex]


no image
Fairness Constraints: A Flexible Approach for Fair Classification

Zafar, M. B., Valera, I., Gomez-Rodriguez, M., Krishna, P.

Journal of Machine Learning Research, 20(75):1-42, 2019 (article)

ei

link (url) [BibTex]

link (url) [BibTex]

2003


no image
Concentration Inequalities for Sub-Additive Functions Using the Entropy Method

Bousquet, O.

Stochastic Inequalities and Applications, 56, pages: 213-247, Progress in Probability, (Editors: Giné, E., C. Houdré and D. Nualart), November 2003 (article)

Abstract
We obtain exponential concentration inequalities for sub-additive functions of independent random variables under weak conditions on the increments of those functions, like the existence of exponential moments for these increments. As a consequence of these general inequalities, we obtain refinements of Talagrand's inequality for empirical processes and new bounds for randomized empirical processes. These results are obtained by further developing the entropy method introduced by Ledoux.

ei

PostScript [BibTex]

2003


PostScript [BibTex]


no image
Statistical Learning Theory, Capacity and Complexity

Schölkopf, B.

Complexity, 8(4):87-94, July 2003 (article)

Abstract
We give an exposition of the ideas of statistical learning theory, followed by a discussion of how a reinterpretation of the insights of learning theory could potentially also benefit our understanding of a certain notion of complexity.

ei

Web DOI [BibTex]


no image
Dealing with large Diagonals in Kernel Matrices

Weston, J., Schölkopf, B., Eskin, E., Leslie, C., Noble, W.

Annals of the Institute of Statistical Mathematics, 55(2):391-408, June 2003 (article)

Abstract
In kernel methods, all the information about the training data is contained in the Gram matrix. If this matrix has large diagonal values, which arises for many types of kernels, then kernel methods do not perform well: We propose and test several methods for dealing with this problem by reducing the dynamic range of the matrix while preserving the positive definiteness of the Hessian of the quadratic programming problem that one has to solve when training a Support Vector Machine, which is a common kernel approach for pattern recognition.

ei

PDF DOI [BibTex]

PDF DOI [BibTex]


no image
The em Algorithm for Kernel Matrix Completion with Auxiliary Data

Tsuda, K., Akaho, S., Asai, K.

Journal of Machine Learning Research, 4, pages: 67-81, May 2003 (article)

ei

PDF [BibTex]

PDF [BibTex]


no image
Constructing Descriptive and Discriminative Non-linear Features: Rayleigh Coefficients in Kernel Feature Spaces

Mika, S., Rätsch, G., Weston, J., Schölkopf, B., Smola, A., Müller, K.

IEEE Transactions on Pattern Analysis and Machine Intelligence, 25(5):623-628, May 2003 (article)

Abstract
We incorporate prior knowledge to construct nonlinear algorithms for invariant feature extraction and discrimination. Employing a unified framework in terms of a nonlinearized variant of the Rayleigh coefficient, we propose nonlinear generalizations of Fisher‘s discriminant and oriented PCA using support vector kernel functions. Extensive simulations show the utility of our approach.

ei

DOI [BibTex]

DOI [BibTex]


no image
Tractable Inference for Probabilistic Data Models

Csato, L., Opper, M., Winther, O.

Complexity, 8(4):64-68, April 2003 (article)

Abstract
We present an approximation technique for probabilistic data models with a large number of hidden variables, based on ideas from statistical physics. We give examples for two nontrivial applications. © 2003 Wiley Periodicals, Inc.

ei

PDF GZIP Web [BibTex]

PDF GZIP Web [BibTex]


no image
Feature selection and transduction for prediction of molecular bioactivity for drug design

Weston, J., Perez-Cruz, F., Bousquet, O., Chapelle, O., Elisseeff, A., Schölkopf, B.

Bioinformatics, 19(6):764-771, April 2003 (article)

Abstract
Motivation: In drug discovery a key task is to identify characteristics that separate active (binding) compounds from inactive (non-binding) ones. An automated prediction system can help reduce resources necessary to carry out this task. Results: Two methods for prediction of molecular bioactivity for drug design are introduced and shown to perform well in a data set previously studied as part of the KDD (Knowledge Discovery and Data Mining) Cup 2001. The data is characterized by very few positive examples, a very large number of features (describing three-dimensional properties of the molecules) and rather different distributions between training and test data. Two techniques are introduced specifically to tackle these problems: a feature selection method for unbalanced data and a classifier which adapts to the distribution of the the unlabeled test data (a so-called transductive method). We show both techniques improve identification performance and in conjunction provide an improvement over using only one of the techniques. Our results suggest the importance of taking into account the characteristics in this data which may also be relevant in other problems of a similar type.

ei

Web [BibTex]


no image
Use of the Zero-Norm with Linear Models and Kernel Methods

Weston, J., Elisseeff, A., Schölkopf, B., Tipping, M.

Journal of Machine Learning Research, 3, pages: 1439-1461, March 2003 (article)

Abstract
We explore the use of the so-called zero-norm of the parameters of linear models in learning. Minimization of such a quantity has many uses in a machine learning context: for variable or feature selection, minimizing training error and ensuring sparsity in solutions. We derive a simple but practical method for achieving these goals and discuss its relationship to existing techniques of minimizing the zero-norm. The method boils down to implementing a simple modification of vanilla SVM, namely via an iterative multiplicative rescaling of the training data. Applications we investigate which aid our discussion include variable and feature selection on biological microarray data, and multicategory classification.

ei

PDF PostScript PDF [BibTex]

PDF PostScript PDF [BibTex]


no image
An Introduction to Variable and Feature Selection.

Guyon, I., Elisseeff, A.

Journal of Machine Learning, 3, pages: 1157-1182, 2003 (article)

ei

[BibTex]

[BibTex]


no image
New Approaches to Statistical Learning Theory

Bousquet, O.

Annals of the Institute of Statistical Mathematics, 55(2):371-389, 2003 (article)

Abstract
We present new tools from probability theory that can be applied to the analysis of learning algorithms. These tools allow to derive new bounds on the generalization performance of learning algorithms and to propose alternative measures of the complexity of the learning task, which in turn can be used to derive new learning algorithms.

ei

PostScript [BibTex]

PostScript [BibTex]

2001


no image
Regularized principal manifolds

Smola, A., Mika, S., Schölkopf, B., Williamson, R.

Journal of Machine Learning Research, 1, pages: 179-209, June 2001 (article)

Abstract
Many settings of unsupervised learning can be viewed as quantization problems - the minimization of the expected quantization error subject to some restrictions. This allows the use of tools such as regularization from the theory of (supervised) risk minimization for unsupervised learning. This setting turns out to be closely related to principal curves, the generative topographic map, and robust coding. We explore this connection in two ways: (1) we propose an algorithm for finding principal manifolds that can be regularized in a variety of ways; and (2) we derive uniform convergence bounds and hence bounds on the learning rates of the algorithm. In particular, we give bounds on the covering numbers which allows us to obtain nearly optimal learning rates for certain types of regularization operators. Experimental results demonstrate the feasibility of the approach.

ei

PDF [BibTex]

2001


PDF [BibTex]


no image
The psychometric function: II. Bootstrap-based confidence intervals and sampling

Wichmann, F., Hill, N.

Perception and Psychophysics, 63 (8), pages: 1314-1329, 2001 (article)

ei

PDF [BibTex]

PDF [BibTex]


no image
The psychometric function: I. Fitting, sampling and goodness-of-fit

Wichmann, F., Hill, N.

Perception and Psychophysics, 63 (8), pages: 1293-1313, 2001 (article)

Abstract
The psychometric function relates an observer'sperformance to an independent variable, usually some physical quantity of a stimulus in a psychophysical task. This paper, together with its companion paper (Wichmann & Hill, 2001), describes an integrated approach to (1) fitting psychometric functions, (2) assessing the goodness of fit, and (3) providing confidence intervals for the function'sparameters and other estimates derived from them, for the purposes of hypothesis testing. The present paper deals with the first two topics, describing a constrained maximum-likelihood method of parameter estimation and developing several goodness-of-fit tests. Using Monte Carlo simulations, we deal with two specific difficulties that arise when fitting functions to psychophysical data. First, we note that human observers are prone to stimulus-independent errors (or lapses ). We show that failure to account for this can lead to serious biases in estimates of the psychometric function'sparameters and illustrate how the problem may be overcome. Second, we note that psychophysical data sets are usually rather small by the standards required by most of the commonly applied statistical tests. We demonstrate the potential errors of applying traditional X^2 methods to psychophysical data and advocate use of Monte Carlo resampling techniques that do not rely on asymptotic theory. We have made available the software to implement our methods

ei

PDF [BibTex]

PDF [BibTex]


no image
Markovian domain fingerprinting: statistical segmentation of protein sequences

Bejerano, G., Seldin, Y., Margalit, H., Tishby, N.

Bioinformatics, 17(10):927-934, 2001 (article)

ei

PDF Web [BibTex]

PDF Web [BibTex]