liu.seSearch for publications in DiVA
Change search
Link to record
Permanent link

Direct link
Publications (10 of 80) Show all publications
Ducrocq, G., runewald, L., Westenhoff, S. & Lindsten, F. (2025). cryoSPHERE: Single-Particle HEterogeneous REconstruction from cryo EM. In: : . Paper presented at The Thirteenth International Conference on Learning Representations, ICLR 2025, Singapore, Apr 24 – 28, 2025.
Open this publication in new window or tab >>cryoSPHERE: Single-Particle HEterogeneous REconstruction from cryo EM
2025 (English)Conference paper, Published paper (Other academic)
Abstract [en]

The three-dimensional structure of proteins plays a crucial role in determining their function. Protein structure prediction methods, like AlphaFold, offer rapid access to a protein’s structure. However, large protein complexes cannot be reliably predicted, and proteins are dynamic, making it important to resolve their full conformational distribution. Single-particle cryo-electron microscopy (cryo-EM) is a powerful tool for determining the structures of large protein complexes. Importantly, the numerous images of a given protein contain underutilized information about conformational heterogeneity. These images are very noisy projections of the protein, and traditional methods for cryo-EM reconstruction are limited to recovering only one or a few consensus conformations.In this paper, we introduce cryoSPHERE, which is a deep learning method that uses a nominal protein structure (e.g., from AlphaFold) as input, learns how to divide it into segments, and moves these segments as approximately rigid bodies to fit the different conformations present in the cryo-EM dataset. This approach provides enough constraints to enable meaningful reconstructions of single protein structural ensembles. We demonstrate this with two synthetic datasets featuring varying levels of noise, as well as two real dataset. We show that cryoSPHERE is very resilient to the high levels of noise typically encountered in experiments, where we see consistent improvements over the current state-of-the-art for heterogeneous reconstruction.

Identifiers
urn:nbn:se:liu:diva-212790 (URN)
Conference
The Thirteenth International Conference on Learning Representations, ICLR 2025, Singapore, Apr 24 – 28, 2025
Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-04
Ekström Kelvinius, F. & Lindsten, F. (2024). Discriminator Guidance for Autoregressive Diffusion Models. In: Proceedings of The 27th International Conference on Artificial Intelligence and Statistics: . Paper presented at International Conference on Artificial Intelligence and Statistics, 2-4 May 2024, Palau de Congressos, Valencia, Spain (pp. 3403-3411). PMLR, 238
Open this publication in new window or tab >>Discriminator Guidance for Autoregressive Diffusion Models
2024 (English)In: Proceedings of The 27th International Conference on Artificial Intelligence and Statistics, PMLR , 2024, Vol. 238, p. 3403-3411Conference paper, Published paper (Refereed)
Abstract [en]

We introduce discriminator guidance in the setting of Autoregressive Diffusion Models. The use of a discriminator to guide a diffusion process has previously been used for continuous diffusion models, and in this work we derive ways of using a discriminator together with a pretrained generative model in the discrete case. First, we show that using an optimal discriminator will correct the pretrained model and enable exact sampling from the underlying data distribution. Second, to account for the realistic scenario of using a sub-optimal discriminator, we derive a sequential Monte Carlo algorithm which iteratively takes the predictions from the discriminator into account during the generation process. We test these approaches on the task of generating molecular graphs and show how the discriminator improves the generative performance over using only the pretrained model.

Place, publisher, year, edition, pages
PMLR, 2024
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:liu:diva-203716 (URN)001286500302034 ()
Conference
International Conference on Artificial Intelligence and Statistics, 2-4 May 2024, Palau de Congressos, Valencia, Spain
Note

Funding Agencies|Swedish Research Council via the project Handling Uncertainty in Machine Learning Systems [2020-04122]; Wallenberg AI, Autonomous Systems and Software Program (WASP) - Knut and Alice Wallenberg Foundation; Excellence Center at Linkoping-Lund in Information Technology (ELLIIT)

Available from: 2024-05-27 Created: 2024-05-27 Last updated: 2024-11-12Bibliographically approved
Konold, P. E., Monrroy, L., Bellisario, A., Filipe, D., Adams, P., Alvarez, R., . . . Westenhoff, S. (2024). Microsecond time-resolved X-ray scattering by utilizing MHz repetition rate at second-generation XFELs. Nature Methods, 21(9), 1608-1611
Open this publication in new window or tab >>Microsecond time-resolved X-ray scattering by utilizing MHz repetition rate at second-generation XFELs
Show others...
2024 (English)In: Nature Methods, ISSN 1548-7091, E-ISSN 1548-7105, Vol. 21, no 9, p. 1608-1611Article in journal (Refereed) Published
Abstract [en]

Detecting microsecond structural perturbations in biomolecules has wide relevance in biology, chemistry and medicine. Here we show how MHz repetition rates at X-ray free-electron lasers can be used to produce microsecond time-series of protein scattering with exceptionally low noise levels of 0.001%. We demonstrate the approach by examining J alpha helix unfolding of a light-oxygen-voltage photosensory domain. This time-resolved acquisition strategy is easy to implement and widely applicable for direct observation of structural dynamics of many biochemical processes. The MHz repetition rates available at second-generation X-ray free-electron lasers enable the collection of microsecond time-resolved X-ray scattering data with exceptionally low noise, providing insights into protein structural dynamics.

Place, publisher, year, edition, pages
NATURE PORTFOLIO, 2024
National Category
Atom and Molecular Physics and Optics
Identifiers
urn:nbn:se:liu:diva-206401 (URN)10.1038/s41592-024-02344-0 (DOI)001262907600003 ()38969722 (PubMedID)
Note

Funding Agencies|Swedish Research Council [2019-06092, 2018-00234, 2022-06725]; BMBF; Universities Australia; German Academic Exchange Service; European Union [101004728]; National Science Foundation; BioXFEL Science and Technology Center [1231306]; Directorate for Biological Sciences [1943448]

Available from: 2024-08-19 Created: 2024-08-19 Last updated: 2025-04-15Bibliographically approved
Govindarajan, H., Sidén, P., Roll, J. & Lindsten, F. (2024). On Partial Prototype Collapse in the DINO Family of Self-Supervised Methods. In: 35th British Machine Vision Conference 2024, Glasgow, UK, November 25-28, 2024: . Paper presented at British Machine Vision Conference (BMVC) 2024.
Open this publication in new window or tab >>On Partial Prototype Collapse in the DINO Family of Self-Supervised Methods
2024 (English)In: 35th British Machine Vision Conference 2024, Glasgow, UK, November 25-28, 2024, 2024Conference paper, Oral presentation only (Other academic)
Abstract [en]

A prominent self-supervised learning paradigm is to model the representations as clusters, or more generally as a mixture model. Learning to map the data samples to compact representations and fitting the mixture model simultaneously leads to the representation collapse problem. Regularizing the distribution of data points over the clusters is the prevalent strategy to avoid this issue. While this is sufficient to prevent full representation collapse, we show that a partial prototype collapse problem still exists in the DINO family of methods, that leads to significant redundancies in the prototypes. Such prototype redundancies serve as shortcuts for the method to achieve a marginal latent class distribution that matches the prescribed prior. We show that by encouraging the model to use diverse prototypes, the partial prototype collapse can be mitigated. We study the downstream impact of effective utilization of the prototypes during pre-training. We show that it enables the methods to learn more fine-grained clusters, encouraging more informative representations. We demonstrate that this is especially beneficial when pre-training on a long-tailed fine-grained dataset.

Keywords
self-supervised learning, vision transformers, long-tailed classification, dino, collapse, few-shot learning, clustering based methods, representation learning
National Category
Computer graphics and computer vision
Identifiers
urn:nbn:se:liu:diva-211237 (URN)
Conference
British Machine Vision Conference (BMVC) 2024
Funder
Swedish Research Council, 2020-04122Wallenberg AI, Autonomous Systems and Software Program (WASP)ELLIIT - The Linköping‐Lund Initiative on IT and Mobile Communications
Available from: 2025-01-28 Created: 2025-01-28 Last updated: 2025-02-13
Olmin, A., Lindqvist, J., Svensson, L. & Lindsten, F. (2024). On the connection between Noise-Contrastive Estimation and Contrastive Divergence. In: INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238: . Paper presented at International Conference on Artificial Intelligence and Statistics, 2-4 May 2024, Palau de Congressos, Valencia, Spain (pp. 3016-3024). JMLR-JOURNAL MACHINE LEARNING RESEARCH, 238
Open this publication in new window or tab >>On the connection between Noise-Contrastive Estimation and Contrastive Divergence
2024 (English)In: INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, JMLR-JOURNAL MACHINE LEARNING RESEARCH , 2024, Vol. 238, p. 3016-3024Conference paper, Published paper (Refereed)
Abstract [en]

Noise-contrastive estimation (NCE) is a popular method for estimating unnormalised probabilistic models, such as energy-based models, which are effective for modelling complex data distributions. Unlike classical maximum likelihood (ML) estimation that relies on importance sampling (resulting in ML-IS) or MCMC (resulting in contrastive divergence, CD), NCE uses a proxy criterion to avoid the need for evaluating an often intractable normalisation constant. Despite apparent conceptual differences, we show that two NCE criteria, ranking NCE (RNCE) and conditional NCE (CNCE), can be viewed as ML estimation methods. Specifically, RNCE is equivalent to ML estimation combined with conditional importance sampling, and both RNCE and CNCE are special cases of CD. These findings bridge the gap between the two method classes and allow us to apply techniques from the ML-IS and CD literature to NCE, offering several advantageous extensions.

Place, publisher, year, edition, pages
JMLR-JOURNAL MACHINE LEARNING RESEARCH, 2024
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
Keywords
Unnormalised models, noise-contrastive estimation, contrastive divergence
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:liu:diva-204020 (URN)001286500301029 ()
Conference
International Conference on Artificial Intelligence and Statistics, 2-4 May 2024, Palau de Congressos, Valencia, Spain
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Note

Funding Agencies|Swedish Research Council via the project Handling Uncertainty in Machine Learning Systems [2020-04122]; Wallenberg AI, Autonomous Systems and Software Program (WASP) - Knut and Alice Wallenberg Foundation; Excellence Center at Linkoping-Lund in Information Technology (EL-LIIT)

Available from: 2024-05-31 Created: 2024-05-31 Last updated: 2024-11-12Bibliographically approved
Oskarsson, J., Landelius, T., Deisenroth, M. P. & Lindsten, F. (2024). Probabilistic Weather Forecasting with Hierarchical Graph Neural Networks. In: A. Globerson and L. Mackey and D. Belgrave and A. Fan and U. Paquet and J. Tomczak and C. Zhang (Ed.), Advances in Neural Information Processing Systems: 38th Conference on Neural Information Processing Systems (NeurIPS 2024). Paper presented at 38th Conference on Neural Information Processing Systems (NeurIPS 2024), 10-15 December 2024, Vancouver, Canada. (pp. 41577-41648). Neural Information Processing Systems, 37
Open this publication in new window or tab >>Probabilistic Weather Forecasting with Hierarchical Graph Neural Networks
2024 (English)In: Advances in Neural Information Processing Systems: 38th Conference on Neural Information Processing Systems (NeurIPS 2024) / [ed] A. Globerson and L. Mackey and D. Belgrave and A. Fan and U. Paquet and J. Tomczak and C. Zhang, Neural Information Processing Systems, 2024, Vol. 37, p. 41577-41648Conference paper, Published paper (Refereed)
Abstract [en]

In recent years, machine learning has established itself as a powerful tool forhigh-resolution weather forecasting. While most current machine learning modelsfocus on deterministic forecasts, accurately capturing the uncertainty in thechaotic weather system calls for probabilistic modeling. We propose a probabilisticweather forecasting model called Graph-EFM, combining a flexible latent-variableformulation with the successful graph-based forecasting framework. The use of ahierarchical graph construction allows for efficient sampling of spatially coherentforecasts. Requiring only a single forward pass per time step, Graph-EFM allowsfor fast generation of arbitrarily large ensembles. We experiment with the modelon both global and limited area forecasting. Ensemble forecasts from Graph-EFMachieve equivalent or lower errors than comparable deterministic models, with theadded benefit of accurately capturing forecast uncertainty.

Place, publisher, year, edition, pages
Neural Information Processing Systems, 2024
Keywords
weather forecasting, graph neural network, probabilistic, ensemble forecasting, latent variable model, earth system modeling
National Category
Computer Sciences
Identifiers
urn:nbn:se:liu:diva-212805 (URN)9798331314385 (ISBN)
Conference
38th Conference on Neural Information Processing Systems (NeurIPS 2024), 10-15 December 2024, Vancouver, Canada.
Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-04Bibliographically approved
Varga, J., Karlsson, E., Raidl, G. R., Rönnberg, E., Lindsten, F. & Rodemann, T. (2024). Speeding Up Logic-Based Benders Decomposition by Strengthening Cuts with Graph Neural Networks. In: Giuseppe Nicosia, Varun Ojha, Emanuele La Malfa, Gabriele La Malfa, Panos M. Pardalos, Renato Umeton (Ed.), Machine Learning, Optimization, and Data Science: . Paper presented at 9th International Conference, LOD 2023, Grasmere, UK, September 22–26, 2023 (pp. 24-38). Cham: Springer
Open this publication in new window or tab >>Speeding Up Logic-Based Benders Decomposition by Strengthening Cuts with Graph Neural Networks
Show others...
2024 (English)In: Machine Learning, Optimization, and Data Science / [ed] Giuseppe Nicosia, Varun Ojha, Emanuele La Malfa, Gabriele La Malfa, Panos M. Pardalos, Renato Umeton, Cham: Springer, 2024, p. 24-38Conference paper, Published paper (Refereed)
Abstract [en]

Logic-based Benders decomposition is a technique to solve optimization problems to optimality. It works by splitting the problem into a master problem, which neglects some aspects of the problem, and a subproblem, which is used to iteratively produce cuts for the master problem to account for those aspects. It is critical for the computational performance that these cuts are strengthened, but the strengthening of cuts comes at the cost of solving additional subproblems. In this work we apply a graph neural network in an autoregressive fashion to approximate the compilation of an irreducible cut, which then only requires few postprocessing steps to ensure its validity. We test the approach on a job scheduling problem with a single machine and multiple time windows per job and compare to approaches from the literature. Results show that our approach is capable of considerably reducing the number of subproblems that need to be solved and hence the total computational effort.

Place, publisher, year, edition, pages
Cham: Springer, 2024
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349 ; 14505
Keywords
Logic-based Benders Decomposition; Cut Strengthening; Graph Neural Networks; Job Scheduling
National Category
Computational Mathematics
Identifiers
urn:nbn:se:liu:diva-200891 (URN)10.1007/978-3-031-53969-5_3 (DOI)001217088300003 ()9783031539688 (ISBN)9783031539695 (ISBN)
Conference
9th International Conference, LOD 2023, Grasmere, UK, September 22–26, 2023
Note

Funding Agencies|Honda Research Institute Europe

Available from: 2024-02-15 Created: 2024-02-15 Last updated: 2025-02-16Bibliographically approved
Ahmadian, A., Ding, Y., Eilertsen, G. & Lindsten, F. (2024). Unsupervised Novelty Detection in Pretrained Representation Space with Locally Adapted Likelihood Ratio. In: International Conference on Artificial Intelligence and Statistics 2024, Proceedings of Machine Learning Research: . Paper presented at 27th International Conference on Artificial Intelligence and Statistics (AISTATS), Valencia, SPAIN, MAY 02-04, 2024. , 238
Open this publication in new window or tab >>Unsupervised Novelty Detection in Pretrained Representation Space with Locally Adapted Likelihood Ratio
2024 (English)In: International Conference on Artificial Intelligence and Statistics 2024, Proceedings of Machine Learning Research, 2024, Vol. 238Conference paper, Published paper (Refereed)
Abstract [en]

Detecting novelties given unlabeled examples of normal data is a challenging task in machine learning, particularly when the novel and normal categories are semantically close. Large deep models pretrained on massive datasets can provide a rich representation space in which the simple k-nearest neighbor distance works as a novelty measure. However, as we show in this paper, the basic k-NN method might be insufficient in this context due to ignoring the 'local geometry' of the distribution over representations as well as the impact of irrelevant 'background features'. To address this, we propose a fully unsupervised novelty detection approach that integrates the flexibility of k-NN with a locally adapted scaling of dimensions based on the 'neighbors of nearest neighbor' and computing a 'likelihood ratio' in pretrained (self-supervised) representation spaces. Our experiments with image data show the advantage of this method when off-the-shelf vision transformers (e.g., pretrained by DINO) are used as the feature extractor without any fine-tuning.

Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Computer Sciences Computer graphics and computer vision Signal Processing
Identifiers
urn:nbn:se:liu:diva-203391 (URN)001221034002024 ()
Conference
27th International Conference on Artificial Intelligence and Statistics (AISTATS), Valencia, SPAIN, MAY 02-04, 2024
Available from: 2024-05-08 Created: 2024-05-08 Last updated: 2025-04-14
Zimmermann, H., Lindsten, F., Meent, J.-W. v. & Naesseth, C. A. (2023). A Variational Perspective on Generative Flow Networks. Transactions on Machine Learning Research
Open this publication in new window or tab >>A Variational Perspective on Generative Flow Networks
2023 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856Article in journal (Refereed) Epub ahead of print
Abstract [en]

Generative flow networks (GFNs) are a class of probabilistic models for sequential samplingof composite objects, proportional to a target distribution that is defined in terms of anenergy function or a reward. GFNs are typically trained using a flow matching or trajectorybalance objective, which matches forward and backward transition models over trajectories.In this work we introduce a variational objective for training GFNs, which is a convexcombination of the reverse- and forward KL divergences, and compare it to the trajectorybalance objective when sampling from the forward- and backward model, respectively. Weshow that, in certain settings, variational inference for GFNs is equivalent to minimizing thetrajectory balance objective, in the sense that both methods compute the same score-functiongradient. This insight suggests that in these settings, control variates, which are commonlyused to reduce the variance of score-function gradient estimates, can also be used with thetrajectory balance objective. We evaluate our findings and the performance of the proposedvariational objective numerically by comparing it to the trajectory balance objective on twosynthetic tasks.

National Category
Probability Theory and Statistics Computer Sciences
Identifiers
urn:nbn:se:liu:diva-204028 (URN)
Funder
ELLIIT - The Linköping‐Lund Initiative on IT and Mobile CommunicationsWallenberg AI, Autonomous Systems and Software Program (WASP)Swedish Research Council, 2020-04122
Available from: 2024-06-01 Created: 2024-06-01 Last updated: 2024-10-30Bibliographically approved
Olmin, A., Lindqvist, J., Svensson, L. & Lindsten, F. (2023). Active Learning with Weak Supervision for Gaussian Processes. In: M. Tanveer et al. (Ed.), Neural Information Processing 29th International Conference, ICONIP 2022, Virtual Event, November 22–26, 2022, Proceedings, Part V: . Paper presented at 29th International Conference on Neural Information Processing, ICONIP 2022, Virtual Event, November 22–26, 2022 (pp. 195-204). Singapore: Springer Nature
Open this publication in new window or tab >>Active Learning with Weak Supervision for Gaussian Processes
2023 (English)In: Neural Information Processing 29th International Conference, ICONIP 2022, Virtual Event, November 22–26, 2022, Proceedings, Part V / [ed] M. Tanveer et al., Singapore: Springer Nature, 2023, p. 195-204Conference paper, Published paper (Refereed)
Abstract [en]

Annotating data for supervised learning can be costly. When the annotation budget is limited, active learning can be used to select and annotate those observations that are likely to give the most gain in model performance. We propose an active learning algorithm that, in addition to selecting which observation to annotate, selects the precision of the annotation that is acquired. Assuming that annotations with low precision are cheaper to obtain, this allows the model to explore a larger part of the input space, with the same annotation budget. We build our acquisition function on the previously proposed BALD objective for Gaussian Processes, and empirically demonstrate the gains of being able to adjust the annotation precision in the active learning loop.

Place, publisher, year, edition, pages
Singapore: Springer Nature, 2023
Series
Communications in Computer and Information Science, ISSN 1865-0929, E-ISSN 1865-0937 ; 1792
Keywords
Machine learning, Active learning, Weak supervision
National Category
Computer Sciences
Identifiers
urn:nbn:se:liu:diva-195039 (URN)10.1007/978-981-99-1642-9_17 (DOI)978-981-99-1641-2 (ISBN)978-981-99-1642-9 (ISBN)
Conference
29th International Conference on Neural Information Processing, ICONIP 2022, Virtual Event, November 22–26, 2022
Available from: 2023-06-14 Created: 2023-06-14 Last updated: 2024-09-19
Projects
Sequential Monte Carlo Workshop [2017-00515_VR]; Uppsala UniversityFuture Leaders Program of STS Forum [2018-03088_Vinnova]; Uppsala University
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-3749-5820

Search in DiVA

Show all publications