liu.seSearch for publications in DiVA
Change search
Link to record
Permanent link

Direct link
Publications (10 of 87) Show all publications
Andrae, M., Landelius, T., Oskarsson, J. & Lindsten, F. (2025). Continuous Ensemble Weather Forecasting with Diffusion models. In: The Thirteenth International Conference on Learning Representations (ICLR 2025): . Paper presented at ICLR 2025, Singapore, Apr 24 – 28, 2025.
Open this publication in new window or tab >>Continuous Ensemble Weather Forecasting with Diffusion models
2025 (English)In: The Thirteenth International Conference on Learning Representations (ICLR 2025), 2025Conference paper, Published paper (Refereed)
Abstract [en]

Weather forecasting has seen a shift in methods from numerical simulations to data-driven systems. While initial research in the area focused on deterministic forecasting, recent works have used diffusion models to produce skillful ensemble forecasts. These models are trained on a single forecasting step and rolled out autoregressively. However, they are computationally expensive and accumulate errors for high temporal resolution due to the many rollout steps. We address these limitations with Continuous Ensemble Forecasting, a novel and flexible method for sampling ensemble forecasts in diffusion models. The method can generate temporally consistent ensemble trajectories completely in parallel, with no autoregressive steps. Continuous Ensemble Forecasting can also be combined with autoregressive rollouts to yield forecasts at an arbitrary fine temporal resolution without sacrificing accuracy. We demonstrate that the method achieves competitive results for global weather forecasting with good probabilistic properties.

National Category
Artificial Intelligence
Identifiers
urn:nbn:se:liu:diva-219565 (URN)
Conference
ICLR 2025, Singapore, Apr 24 – 28, 2025
Available from: 2025-11-18 Created: 2025-11-18 Last updated: 2025-11-28
Ducrocq, G., runewald, L., Westenhoff, S. & Lindsten, F. (2025). cryoSPHERE: Single-Particle HEterogeneous REconstruction from cryo EM. In: : . Paper presented at The Thirteenth International Conference on Learning Representations, ICLR 2025, Singapore, Apr 24 – 28, 2025.
Open this publication in new window or tab >>cryoSPHERE: Single-Particle HEterogeneous REconstruction from cryo EM
2025 (English)Conference paper, Published paper (Other academic)
Abstract [en]

The three-dimensional structure of proteins plays a crucial role in determining their function. Protein structure prediction methods, like AlphaFold, offer rapid access to a protein’s structure. However, large protein complexes cannot be reliably predicted, and proteins are dynamic, making it important to resolve their full conformational distribution. Single-particle cryo-electron microscopy (cryo-EM) is a powerful tool for determining the structures of large protein complexes. Importantly, the numerous images of a given protein contain underutilized information about conformational heterogeneity. These images are very noisy projections of the protein, and traditional methods for cryo-EM reconstruction are limited to recovering only one or a few consensus conformations.In this paper, we introduce cryoSPHERE, which is a deep learning method that uses a nominal protein structure (e.g., from AlphaFold) as input, learns how to divide it into segments, and moves these segments as approximately rigid bodies to fit the different conformations present in the cryo-EM dataset. This approach provides enough constraints to enable meaningful reconstructions of single protein structural ensembles. We demonstrate this with two synthetic datasets featuring varying levels of noise, as well as two real dataset. We show that cryoSPHERE is very resilient to the high levels of noise typically encountered in experiments, where we see consistent improvements over the current state-of-the-art for heterogeneous reconstruction.

Identifiers
urn:nbn:se:liu:diva-212790 (URN)
Conference
The Thirteenth International Conference on Learning Representations, ICLR 2025, Singapore, Apr 24 – 28, 2025
Available from: 2025-04-03 Created: 2025-04-03 Last updated: 2025-04-04
Ohl, L. & Lindsten, F. (2025). Discriminative ordering through ensemble consensus. In: Chiappa S., Magliacane S. (Ed.), Proceedings of the 41st Conference on Uncertainty in Artificial Intelligence: . Paper presented at UAI '25: Conference on Uncertainty in Artificial Intelligence Rio de Janeiro Brazil July 21 - 25, 2025 (pp. 3252-3271). ML Research Press, 286
Open this publication in new window or tab >>Discriminative ordering through ensemble consensus
2025 (English)In: Proceedings of the 41st Conference on Uncertainty in Artificial Intelligence / [ed] Chiappa S., Magliacane S., ML Research Press , 2025, Vol. 286, p. 3252-3271Conference paper, Published paper (Refereed)
Abstract [en]

Evaluating the performance of clustering models isa challenging task where the outcome depends onthe definition of what constitutes a cluster. Due tothis design, current existing metrics rarely handlemultiple clustering models with diverse cluster def-initions, nor do they comply with the integration ofconstraints when available. In this work, we takeinspiration from consensus clustering and assumethat a set of clustering models is able to uncoverhidden structures in the data. We propose to con-struct a discriminative ordering through ensembleconsensus based on the distance between the con-nectivity of a clustering model and the consensusmatrix. We first validate the proposed method withsynthetic scenarios, highlighting that the proposedscore ranks the models that best match the consen-sus first. We then show that this simple rankingscore significantly outperforms other scoring meth-ods when comparing sets of different clusteringalgorithms that are not restricted to a fixed num-ber of clusters and is compatible with clusteringconstraints.

Place, publisher, year, edition, pages
ML Research Press, 2025
Series
Proceedings of Machine Learning Research, ISSN 2640-3498 ; 286
National Category
Artificial Intelligence
Identifiers
urn:nbn:se:liu:diva-217109 (URN)001592914500140 ()2-s2.0-105014724063 (Scopus ID)
Conference
UAI '25: Conference on Uncertainty in Artificial Intelligence Rio de Janeiro Brazil July 21 - 25, 2025
Funder
Knut and Alice Wallenberg Foundation, 2020.0033Wallenberg AI, Autonomous Systems and Software Program (WASP)ELLIIT - The Linköping‐Lund Initiative on IT and Mobile CommunicationsSwedish Research Council, 2020-04122Swedish Research Council, 2024-05011
Note

Funding Agencies|Swedish Research Council [2020-04122, 2024-05011]; Knut and Alice Wallenberg Foundation [KAW 2020.0033]; Wallenberg AI, Autonomous Systems and Software Program (WASP); Excellence Center at Linkoping-Lund in Information Technology (ELLIIT)

Available from: 2025-09-01 Created: 2025-09-01 Last updated: 2025-12-18
Ahmadian, A. & Lindsten, F. (2025). Improved Contrastive Predictive Coding for Time Series Out-Of-Distribution Detection Applied to Human Activity Data. Pattern Recognition Letters, 197, 132-138
Open this publication in new window or tab >>Improved Contrastive Predictive Coding for Time Series Out-Of-Distribution Detection Applied to Human Activity Data
2025 (English)In: Pattern Recognition Letters, ISSN 0167-8655, E-ISSN 1872-7344, Vol. 197, p. 132-138Article in journal (Refereed) Published
Abstract [en]

Contrastive Predictive Coding (CPC) is a well-established self-supervised learning method that naturally fits time series data. This method has been recently leveraged to detect anomalous inputs, viewed as the task of classifying positive pairs of context-feature representations versus negative ones in order to employ classifier uncertainty measures. In this paper, by taking a different perspective, we propose a CPC-based Out-Of-Distribution (OOD) detection method for time series data that does not require any negative samples at test time and is theoretically related to a probabilistic type of uncertainty estimation in the latent representation space. Our method extends the standard CPC by using a radial (distance-based) score function both in the training loss and as the OOD measure, in addition to quantizing the context (replacing it by cluster prototypes) during inference. The proposed method is applied to detecting OOD human activities with smartphone sensors data and shows promising performance on two primary datasets without using activity labels in training.

Place, publisher, year, edition, pages
ELSEVIER, 2025
Keywords
Out-Of-Distribution/anomaly/novelty; detection; Self-supervised machine learning; Contrastive learning; Human activity recognition; Time series analysis; Deep learning
National Category
Computer Sciences
Identifiers
urn:nbn:se:liu:diva-217248 (URN)10.1016/j.patrec.2025.07.011 (DOI)001545804100001 ()2-s2.0-105012301703 (Scopus ID)
Note

Funding Agencies|Swedish Research Council [2024-05011]; Wallenberg AI, Autonomous Systems and Software Program (WASP) - Knut and Alice Wallenberg Foundation; Excellence Center at Linkoeping-Lund in Information Technology (ELLIIT)

Available from: 2025-09-03 Created: 2025-09-03 Last updated: 2025-12-18
Athanasiadis, I., Lindsten, F. & Felsberg, M. (2025). Prior Learning in Introspective VAEs. Transactions on Machine Learning Research (06), 1-41
Open this publication in new window or tab >>Prior Learning in Introspective VAEs
2025 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856, no 06, p. 1-41Article in journal (Refereed) Published
Abstract [en]

Variational Autoencoders (VAEs) are a popular framework for unsupervised learning anddata generation. A plethora of methods have been proposed focusing on improving VAEs,with the incorporation of adversarial objectives and the integration of prior learning mechanismsbeing prominent directions. When it comes to the former, an indicative instance is therecently introduced family of Introspective VAEs aiming at ensuring that a low likelihood isassigned to unrealistic samples. In this study, we focus on the Soft-IntroVAE (S-IntroVAE),one of only two members of the Introspective VAE family, the other being the originalIntroVAE. We select S-IntroVAE for its state-of-the-art status and its training stability.In particular, we investigate the implication of incorporating a multimodal and trainableprior into this S-IntroVAE. Namely, we formulate the prior as a third player and show thatwhen trained in cooperation with the decoder constitutes an effective way for prior learning,which shares the Nash Equilibrium with the vanilla S-IntroVAE. Furthermore, basedon a modified formulation of the optimal ELBO in S-IntroVAE, we develop theoreticallymotivated regularizations, namely (i) adaptive variance clipping to stabilize training whenlearning the prior and (ii) responsibility regularization to discourage the formation of inactiveprior modes. Finally, we perform a series of targeted experiments on a 2D densityestimation benchmark and in an image generation setting comprised of the (F)-MNIST andCIFAR-10 datasets demonstrating the effect of prior learning in S-IntroVAE in generationand representation learning.

Keywords
probaiblity theory and statistics, computer sciences
National Category
Computer Sciences
Identifiers
urn:nbn:se:liu:diva-214678 (URN)
Funder
Wallenberg AI, Autonomous Systems and Software Program (WASP)
Available from: 2025-06-12 Created: 2025-06-12 Last updated: 2025-06-12Bibliographically approved
Ekström Kelvinius, F., Zhao, Z. & Lindsten, F. (2025). Solving Linear-Gaussian Bayesian Inverse Problems with Decoupled Diffusion Sequential Monte Carlo. In: Proceedings of the 42nd International Conference on Machine Learning: . Paper presented at ICML 2025,Forty-Second International Conference on Machine Learning, Vancouver Convention Center Sun. July 13th through Sat. July 19th (pp. 15148-15181). PMLR, 267
Open this publication in new window or tab >>Solving Linear-Gaussian Bayesian Inverse Problems with Decoupled Diffusion Sequential Monte Carlo
2025 (English)In: Proceedings of the 42nd International Conference on Machine Learning, PMLR , 2025, Vol. 267, p. 15148-15181Conference paper, Published paper (Refereed)
Abstract [en]

A recent line of research has exploited pre-trained generative diffusion models as priors for solving Bayesian inverse problems. We contribute to this research direction by designing a sequential Monte Carlo method for linear-Gaussian inverse problems which builds on “decoupled diffusion", where the generative process is designed such that larger updates to the sample are possible. The method is asymptotically exact and we demonstrate the effectiveness of our Decoupled Diffusion Sequential Monte Carlo (DDSMC) algorithm on both synthetic as well as protein and image data. Further, we demonstrate how the approach can be extended to discrete data.

Place, publisher, year, edition, pages
PMLR, 2025
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:liu:diva-218522 (URN)
Conference
ICML 2025,Forty-Second International Conference on Machine Learning, Vancouver Convention Center Sun. July 13th through Sat. July 19th
Available from: 2025-10-07 Created: 2025-10-07 Last updated: 2025-10-13
Ekström Kelvinius, F., Andersson, O., Parackal, A. S., Qian, D., Armiento, R. & Lindsten, F. (2025). WyckoffDiff– A Generative Diffusion Model for Crystal Symmetry. In: Proceedings of the 42nd International Conference on Machine Learning: . Paper presented at ICML 2025, Forty-Second International Conference on Machine Learning, Vancouver Convention Center, Sun. July 13th through Sat. July 19th (pp. 15130-15147). PMLR, 267
Open this publication in new window or tab >>WyckoffDiff– A Generative Diffusion Model for Crystal Symmetry
Show others...
2025 (English)In: Proceedings of the 42nd International Conference on Machine Learning, PMLR , 2025, Vol. 267, p. 15130-15147Conference paper, Published paper (Refereed)
Abstract [en]

Crystalline materials often exhibit a high level of symmetry. However, most generative models do not account for symmetry, but rather model each atom without any constraints on its position or element. We propose a generative model, Wyckoff Diffusion (WyckoffDiff), which generates symmetry-based descriptions of crystals. This is enabled by considering a crystal structure representation that encodes all symmetry, and we design a novel neural network architecture which enables using this representation inside a discrete generative model framework. In addition to respecting symmetry by construction, the discrete nature of our model enables fast generation. We additionally present a new metric, Fréchet Wrenformer Distance, which captures the symmetry aspects of the materials generated, and we benchmark WyckoffDiff against recently proposed generative models for crystal generation. As a proof-of-concept study, we use WyckoffDiff to find new materials below the convex hull of thermodynamical stability.

Place, publisher, year, edition, pages
PMLR, 2025
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Condensed Matter Physics Artificial Intelligence
Identifiers
urn:nbn:se:liu:diva-218524 (URN)
Conference
ICML 2025, Forty-Second International Conference on Machine Learning, Vancouver Convention Center, Sun. July 13th through Sat. July 19th
Available from: 2025-10-07 Created: 2025-10-07 Last updated: 2025-10-13
Zimmermann, H., Lindsten, F., Meent, J.-W. v. & Naesseth, C. A. (2024). A Variational Perspective on Generative Flow Networks. Transactions on Machine Learning Research
Open this publication in new window or tab >>A Variational Perspective on Generative Flow Networks
2024 (English)In: Transactions on Machine Learning Research, E-ISSN 2835-8856Article in journal (Refereed) Published
Abstract [en]

Generative flow networks (GFNs) are a class of probabilistic models for sequential samplingof composite objects, proportional to a target distribution that is defined in terms of anenergy function or a reward. GFNs are typically trained using a flow matching or trajectorybalance objective, which matches forward and backward transition models over trajectories.In this work we introduce a variational objective for training GFNs, which is a convexcombination of the reverse- and forward KL divergences, and compare it to the trajectorybalance objective when sampling from the forward- and backward model, respectively. Weshow that, in certain settings, variational inference for GFNs is equivalent to minimizing thetrajectory balance objective, in the sense that both methods compute the same score-functiongradient. This insight suggests that in these settings, control variates, which are commonlyused to reduce the variance of score-function gradient estimates, can also be used with thetrajectory balance objective. We evaluate our findings and the performance of the proposedvariational objective numerically by comparing it to the trajectory balance objective on twosynthetic tasks.

National Category
Probability Theory and Statistics Computer Sciences
Identifiers
urn:nbn:se:liu:diva-204028 (URN)
Funder
ELLIIT - The Linköping‐Lund Initiative on IT and Mobile CommunicationsWallenberg AI, Autonomous Systems and Software Program (WASP)Swedish Research Council, 2020-04122
Available from: 2024-06-01 Created: 2024-06-01 Last updated: 2025-09-24Bibliographically approved
Ekström Kelvinius, F. & Lindsten, F. (2024). Discriminator Guidance for Autoregressive Diffusion Models. In: Proceedings of The 27th International Conference on Artificial Intelligence and Statistics: . Paper presented at International Conference on Artificial Intelligence and Statistics, 2-4 May 2024, Palau de Congressos, Valencia, Spain (pp. 3403-3411). PMLR, 238
Open this publication in new window or tab >>Discriminator Guidance for Autoregressive Diffusion Models
2024 (English)In: Proceedings of The 27th International Conference on Artificial Intelligence and Statistics, PMLR , 2024, Vol. 238, p. 3403-3411Conference paper, Published paper (Refereed)
Abstract [en]

We introduce discriminator guidance in the setting of Autoregressive Diffusion Models. The use of a discriminator to guide a diffusion process has previously been used for continuous diffusion models, and in this work we derive ways of using a discriminator together with a pretrained generative model in the discrete case. First, we show that using an optimal discriminator will correct the pretrained model and enable exact sampling from the underlying data distribution. Second, to account for the realistic scenario of using a sub-optimal discriminator, we derive a sequential Monte Carlo algorithm which iteratively takes the predictions from the discriminator into account during the generation process. We test these approaches on the task of generating molecular graphs and show how the discriminator improves the generative performance over using only the pretrained model.

Place, publisher, year, edition, pages
PMLR, 2024
Series
Proceedings of Machine Learning Research, ISSN 2640-3498
National Category
Probability Theory and Statistics
Identifiers
urn:nbn:se:liu:diva-203716 (URN)001286500302034 ()
Conference
International Conference on Artificial Intelligence and Statistics, 2-4 May 2024, Palau de Congressos, Valencia, Spain
Note

Funding Agencies|Swedish Research Council via the project Handling Uncertainty in Machine Learning Systems [2020-04122]; Wallenberg AI, Autonomous Systems and Software Program (WASP) - Knut and Alice Wallenberg Foundation; Excellence Center at Linkoping-Lund in Information Technology (ELLIIT)

Available from: 2024-05-27 Created: 2024-05-27 Last updated: 2025-10-06Bibliographically approved
Konold, P. E., Monrroy, L., Bellisario, A., Filipe, D., Adams, P., Alvarez, R., . . . Westenhoff, S. (2024). Microsecond time-resolved X-ray scattering by utilizing MHz repetition rate at second-generation XFELs. Nature Methods, 21(9), 1608-1611
Open this publication in new window or tab >>Microsecond time-resolved X-ray scattering by utilizing MHz repetition rate at second-generation XFELs
Show others...
2024 (English)In: Nature Methods, ISSN 1548-7091, E-ISSN 1548-7105, Vol. 21, no 9, p. 1608-1611Article in journal (Refereed) Published
Abstract [en]

Detecting microsecond structural perturbations in biomolecules has wide relevance in biology, chemistry and medicine. Here we show how MHz repetition rates at X-ray free-electron lasers can be used to produce microsecond time-series of protein scattering with exceptionally low noise levels of 0.001%. We demonstrate the approach by examining J alpha helix unfolding of a light-oxygen-voltage photosensory domain. This time-resolved acquisition strategy is easy to implement and widely applicable for direct observation of structural dynamics of many biochemical processes. The MHz repetition rates available at second-generation X-ray free-electron lasers enable the collection of microsecond time-resolved X-ray scattering data with exceptionally low noise, providing insights into protein structural dynamics.

Place, publisher, year, edition, pages
NATURE PORTFOLIO, 2024
National Category
Atom and Molecular Physics and Optics
Identifiers
urn:nbn:se:liu:diva-206401 (URN)10.1038/s41592-024-02344-0 (DOI)001262907600003 ()38969722 (PubMedID)
Note

Funding Agencies|Swedish Research Council [2019-06092, 2018-00234, 2022-06725]; BMBF; Universities Australia; German Academic Exchange Service; European Union [101004728]; National Science Foundation; BioXFEL Science and Technology Center [1231306]; Directorate for Biological Sciences [1943448]

Available from: 2024-08-19 Created: 2024-08-19 Last updated: 2025-04-15Bibliographically approved
Projects
Sequential Monte Carlo Workshop [2017-00515_VR]; Uppsala UniversityFuture Leaders Program of STS Forum [2018-03088_Vinnova]; Uppsala University
Organisations
Identifiers
ORCID iD: ORCID iD iconorcid.org/0000-0003-3749-5820

Search in DiVA

Show all publications