{"id":2434,"date":"2025-06-13T09:16:01","date_gmt":"2025-06-13T09:16:01","guid":{"rendered":"https:\/\/ivi.fnwi.uva.nl\/ellis\/?post_type=news-&#038;p=2434"},"modified":"2025-06-13T12:43:08","modified_gmt":"2025-06-13T12:43:08","slug":"promoting-local-research-excellence-to-the-global-stage-ellis-unit-amsterdam-travel-grants","status":"publish","type":"news-","link":"https:\/\/ivi.fnwi.uva.nl\/ellis\/news-\/promoting-local-research-excellence-to-the-global-stage-ellis-unit-amsterdam-travel-grants\/","title":{"rendered":"Promoting Local Research Excellence to the Global Stage: ELLIS unit Amsterdam Travel Grants"},"content":{"rendered":"<p>The Amsterdam unit of the European Laboratory for Learning and Intelligent Systems&#8211; ELLIS Society network is proud to support the next generation of Machine Learning and Artificial Intelligence researchers through its <strong data-start=\"393\" data-end=\"435\">Travel Grants for BSc and MSc students<\/strong>. The grants are designed to help students take their cutting-edge research beyond the lab and onto the global stage\u2014presenting their works at the world\u2019s leading AI and ML conferences.<\/p>\n<p data-start=\"623\" data-end=\"1022\">Between <strong>2024 and 2025<\/strong>, the ELLIS Amsterdam Travel Grants have enabled students to present at prestigious venues including the eighteenth European Conference on Computer Vision (<strong>ECCV 2024<\/strong>), the thirty-eighth Conference on Neural Information Processing Systems (<strong>NeurIPS 2024<\/strong>), the twelfth and thirteenth International Conference on Learning Representations (<strong>ICLR 2024 and 2025<\/strong>), the forty-first and forty-second International Conference on Machine Learning (<strong>ICML 2024 and 2025<\/strong>), and the forty-first Conference on Uncertainty in Artificial Intelligence (<strong>UAI 2025<\/strong>).<\/p>\n<p>In 2023, four travel grants were awarded to MSc students who published at the seventeeth Conference of the European Chapter of the Association for Computational Linguistics (EACL 2023), the sixty-first Annual Meeting of the Association for Computational Linguistics (ACL 2023), the Fortieth International Conference on Machine Learning (ICML 2023), the twenty-sixth\u00a0International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2023). These conferences are key venues in AI\/ML research, the forefront of innovation, and participation at the top-tier conferences not only amplifies the impact of the students&#8217; research but also connects them with the wider research community.<\/p>\n<p>By enabling young researchers to showcase their work at international conferences, the unit <span data-preserver-spaces=\"true\">encourages the development of high-quality research locally and provides research exposure and recognition, as Amsterdam\u2019s growing contribution to the European and global AI landscapes. <\/span><\/p>\n<p>The unit also notes that several of the research projects being presented are the\u00a0<span style=\"box-sizing: border-box; margin: 0px; padding: 0px;\">result of\u00a0<strong>cross-border collaboration<\/strong> within the wider ELLIS Society network, made possible by, among others, the <a href=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/funding\/ellis-msc-honours-programme\/\">ELLIS unit Amsterdam\u2019s MSc Honours Programme<\/a>, which features co-supervision by<\/span>\u00a0two ELLIS Members in different countries. This underscores the collaborative spirit that ELLIS champions: pushing the boundaries of AI research through shared knowledge and international teamwork.<\/p>\n<p data-start=\"1685\" data-end=\"2004\">The grants are currently awarded to students whose papers are accepted to the <strong data-start=\"1763\" data-end=\"1797\">main tracks of top conferences<\/strong>, ensuring the highest standards of research are maintained. The travel grant initiative serves a dual purpose: celebrating local academic excellence and showcasing Amsterdam\u2019s thriving AI ecosystem to the world.<\/p>\n<p data-start=\"2006\" data-end=\"2134\">The UvA BSc and MSc Students who are interested in applying can find detailed information about the eligibility criteria and application process of the <a href=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/funding\/travel-grants\/\">ELLIS unit Amsterdam&#8217;s Travel Grants<\/a>, which is fully funded by the University of Amsterdam.<\/p>\n<p data-start=\"2006\" data-end=\"2134\">We highlight ten main track papers funded by the ELLIS unit Amsterdam travel grants in 2024-2025, as follows (in chronological order):<\/p>\n<p><em><strong>1. GeneralAD: Anomaly Detection Across Domains by Attending to Distorted Features<\/strong><\/em><\/p>\n<p><span style=\"font-size: 16px;\">Authors- <strong>Luc P.J. Str\u00e4ter<\/strong>, Mohammadreza Salehi, Efstratios Gavves, Cees G.M. Snoek, Yuki M. Asano\u202f<\/span><\/p>\n<p>Conference- <strong>The Eighteenth European Conference on Computer Vision (ECCV 20224)<\/strong><\/p>\n<p>In the domain of anomaly detection, methods often excel in either high-level semantic or low-level industrial benchmarks, rarely achieving cross-domain proficiency. Semantic anomalies are novelties that differ in meaning from the training set, like unseen objects in self-driving cars. In contrast, industrial anomalies are subtle defects that preserve semantic meaning, such as cracks in airplane components. In this paper, we present GeneralAD, an anomaly detection framework designed to operate in semantic, near-distribution, and industrial settings with minimal per-task adjustments. In our approach, we capitalize on the inherent design of Vision Transformers, which are trained on image patches, thereby ensuring that the last hidden states retain a patch-based structure. We propose a novel self-supervised anomaly generation module that employs straightforward operations like noise addition and shuffling to patch features to construct pseudo-abnormal samples. These features are fed to an attention-based discriminator, which is trained to score every patch in the image. With this, our method can both accurately identify anomalies at the image level and also generate interpretable anomaly maps. We extensively evaluated our approach on ten datasets, achieving state-of-the-art results in six and on-par performance in the remaining for both localization and detection tasks.<\/p>\n<p><img fetchpriority=\"high\" decoding=\"async\" class=\" wp-image-2435 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/1.jpg\" alt=\"\u2018GeneralAD: Anomaly Detection Across Domains by Attending to Distorted Features\u2019\" width=\"741\" height=\"318\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/1.jpg 774w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/1-300x129.jpg 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/1-768x329.jpg 768w\" sizes=\"(max-width: 741px) 100vw, 741px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>2. No Train, all Gain: Self-Supervised Gradients Improve Deep Frozen Representations<\/strong><\/em><\/p>\n<p>Authors &#8211; <strong>Walter Simoncini<\/strong>, Andrei Bursuc, Spyros Gidaris, Yuki M. Asano<\/p>\n<p>Conference &#8211; <strong>The<\/strong> <strong>Thirty-Eighth Conference on Neural Information Processing Systems (NeurIPS 2024)<\/strong><\/p>\n<p>Summary &#8211; this paper introduces FUNGI, Features from UNsupervised GradIents, a method to enhance the features of transformer encoders by leveraging self-supervised gradients. Our method is simple: given any pretrained model, we first compute gradients from various self-supervised objectives for each input. These gradients are projected to a lower dimension and then concatenated with the model&#8217;s output embedding. The resulting features are evaluated on k-nearest neighbor classification over 11 datasets from vision, 5 from natural language processing, and 2 from audio. Across backbones spanning various sizes and pretraining strategies, FUNGI features provide consistent performance improvements over the embeddings. We also show that using FUNGI features can benefit linear classification, clustering and image retrieval, and that they significantly improve the retrieval-based in-context scene understanding abilities of pretrained models, for example improving upon DINO by +17% for semantic segmentation &#8211; without any training.<span data-ccp-props=\"{&quot;134233117&quot;:true,&quot;134233118&quot;:true}\">\u00a0<\/span>We propose using self-supervised gradients to enhance pretrained embedding features and achieve significant improvements in k-nearest neighbor classification, in-context learning, linear probing and clustering, across images, text and audio.<br \/>\n<span data-contrast=\"none\">(i) Given a transformer backbone, <\/span>e.g. a vision transformer (ViT), we forward an image or its views\/patches (1), compute a self-supervised loss, e.g. SimCLR, and backpropagate (2). We then extract gradients with respect to the weights of a linear layer and project them (3, 4).<\/p>\n<p><img decoding=\"async\" class=\"wp-image-2436 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2a.png\" alt=\"\" width=\"801\" height=\"191\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2a.png 1384w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2a-300x72.png 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2a-1024x244.png 1024w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2a-768x183.png 768w\" sizes=\"(max-width: 801px) 100vw, 801px\" \/><\/p>\n<p><span data-contrast=\"auto\">(ii) We call these FUNGI (Features from UNsupervised GradIents). They can be used as-is or combined with FUNGIs from other SSL losses or the model embeddings.<\/span><\/p>\n<p><img decoding=\"async\" class=\" wp-image-2437 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2b.png\" alt=\"\" width=\"807\" height=\"219\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2b.png 1384w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2b-300x81.png 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2b-1024x277.png 1024w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2b-768x208.png 768w\" sizes=\"(max-width: 807px) 100vw, 807px\" \/><br \/>\n<span data-contrast=\"auto\">\u00a0(iii) <\/span><span data-contrast=\"auto\">These features can be used for multiple tasks (retrieval, linear classification, clustering). E.g. in retrieval, our features improve k-nn classification for several backbones (on average across 11 vision datasets), especially in few-shot setups.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-2438 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2c.png\" alt=\"\" width=\"827\" height=\"226\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2c.png 1384w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2c-300x82.png 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2c-1024x280.png 1024w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2c-768x210.png 768w\" sizes=\"(max-width: 827px) 100vw, 827px\" \/><br \/>\n<span data-contrast=\"auto\">\u00a0(iv) <\/span><span data-contrast=\"auto\">They can also improve in-context semantic segmentation, producing sharper and denser masks, even for complex objects such as horses and trains.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-2439 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2d.png\" alt=\"\" width=\"604\" height=\"220\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2d.png 604w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2d-300x109.png 300w\" sizes=\"(max-width: 604px) 100vw, 604px\" \/><br \/>\n<span data-contrast=\"auto\">\u00a0(v) <\/span><span data-contrast=\"auto\">Where GT stands for Ground Truth, and DINO is the baseline. While the experiments illustrated above are for vision, the findings also extend to language and audio.<\/span><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"size-full wp-image-2440 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2e.png\" alt=\"\" width=\"604\" height=\"220\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2e.png 604w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/2e-300x109.png 300w\" sizes=\"(max-width: 604px) 100vw, 604px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>3. VeRA: Vector-based Random Matrix Adaptation<\/strong><\/em><\/p>\n<p>Authors- <strong>Dawid J. Kopiczko<\/strong>, Tijmen Blankevoort, Yuki M. Asano<\/p>\n<p>Conference- <strong>The Twelfth International Conference on Learning Representations (ICLR\u00a02024)<\/strong><\/p>\n<p>Summary- Low-rank adapation (LoRA) is a popular method that reduces the number of trainable parameters when finetuning large language models, but still faces acute storage challenges when scaling to even larger models or deploying numerous per-user or per-task adapted models. In this work, we present Vector-based Random Matrix Adaptation (VeRA), which significantly reduces the number of trainable parameters compared to LoRA, yet maintains the same performance. It achieves this by using a single pair of low-rank matrices shared across all layers and learning small scaling vectors instead. We demonstrate its effectiveness on the GLUE and E2E benchmarks, image classification tasks, and show its application in instruction-tuning of 7B and 13B language models.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-2450 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/methods-1.jpg\" alt=\"\" width=\"823\" height=\"319\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/methods-1.jpg 916w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/methods-1-300x116.jpg 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/methods-1-768x298.jpg 768w\" sizes=\"(max-width: 823px) 100vw, 823px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>4. A Probabilistic Approach To Learning the Degree of Equivariance in Steerable CNNs<\/strong><\/em><\/p>\n<p>Authors- <strong>Lars\u00a0Veefkind<\/strong> and Gabriele Cesa<\/p>\n<p>Conference- <strong>The Forty-First International Conference on Machine Learning\u00a0(ICML 2024)<\/strong><\/p>\n<p>Summary- Steerable convolutional neural networks (SCNNs) enhance task performance by modelling geometric symmetries through equivariance constraints on weights. Yet, unknown or varying symmetries can lead to overconstrained weights and decreased performance. To address this, this paper introduces a probabilistic method to learn the degree of equivariance in SCNNs. We parameterise the degree of equivariance as a likelihood distribution over the transformation group using Fourier coefficients, offering the option to model layer-wise and shared equivariance. These likelihood distributions are regularised to ensure an interpretable degree of equivariance across the network. Advantages include the applicability to many types of equivariant networks through the flexible framework of SCNNs and the ability to learn equivariance with respect to any subgroup of any compact group without requiring additional layers. Our experiments reveal competitive performance on datasets with mixed symmetries, with learnt likelihood distributions that are representative of the underlying degree of equivariance.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-2442 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/4.png\" alt=\"\" width=\"546\" height=\"493\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/4.png 782w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/4-300x271.png 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/4-768x693.png 768w\" sizes=\"(max-width: 546px) 100vw, 546px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>5. Utility-Directed Conformal Prediction: A Decision-Aware Framework for Actionable Uncertainty Quantification<\/strong><\/em><\/p>\n<p>Authors &#8211; Santiago Cortes-Gomez, <strong>Carlos Pati\u00f1o<\/strong>, Yewon Byun, Steven Wu, Eric Horvitz, Bryan Wilder<\/p>\n<p>Conference &#8211; <strong>The Thirteenth International Conference on Learning Representations (ICLR 2025)<\/strong><\/p>\n<p>Summary &#8211; This paper introduces a novel approach to conformal prediction that integrates downstream utility functions, improving decision-making in high-stakes domains. Traditional conformal predictors provide statistical coverage guarantees but lack alignment with specific decision objectives. This work bridges this gap by proposing methods to generate prediction sets that minimize user-defined cost functions while maintaining coverage guarantees. The main result is on a dataset for detecting skin diseases, where the proposed methods are more useful than standard conformal prediction because their predictions do not mix benign\u00a0and malignant diagnoses.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-2451 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/6-1.png\" alt=\"\" width=\"788\" height=\"328\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/6-1.png 904w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/6-1-300x125.png 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/6-1-768x319.png 768w\" sizes=\"(max-width: 788px) 100vw, 788px\" \/><\/p>\n<p><em><strong>6. Compositional Entailment Learning for Hyperbolic Vision-Language Models<\/strong><\/em><\/p>\n<p>Authors &#8211; <strong>Avik Pal<\/strong>, Max van Spengler, Guido Maria D&#8217;Amely di Melendugno, Alessandro Flaborea, Fabio Galasso, Pascal Mettes<\/p>\n<p>Conference &#8211;\u00a0<strong>The Thirteenth International Conference on Learning Representations (ICLR 2025)<\/strong><\/p>\n<p>Summary &#8211;\u00a0Images and texts are naturally hierarchical. This work proposes Compositional Entailment Learning for hyperbolic vision-language models. The idea is that an image is not only described by a sentence but is a composition of multiple object boxes, each with its textual description. Such information can be obtained freely by extracting nouns from sentences and using openly available localized grounding models. This method hierarchically organizes images, image boxes, and textual descriptions through contrastive and entailment-based objectives. Empirical evaluation on a hyperbolic vision-language model trained with millions of image-text pairs shows that the proposed compositional learning approach outperforms conventional Euclidean CLIP learning, as well as recent hyperbolic alternatives, with better zero-shot and retrieval generalization and stronger hierarchical performance.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\" wp-image-2445 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/7.jpg\" alt=\"\" width=\"810\" height=\"303\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/7.jpg 904w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/7-300x112.jpg 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/7-768x287.jpg 768w\" sizes=\"(max-width: 810px) 100vw, 810px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>7. BrainACTIV: Identifying visuo-semantic properties driving cortical selectivity using diffusion-based image manipulation<\/strong><\/em><\/p>\n<p>Authors- <strong>Diego Garc\u00eda Cerdas<\/strong>, Christina Sartzetaki, Magnus Petersen, Gemma Roig, Pascal Mettes, Iris Groen<\/p>\n<p>Conference- <strong>The Thirteenth International Conference on Learning Representations (ICLR 2025)<\/strong><\/p>\n<p>Summary- The human brain efficiently represents visual inputs through specialized neural populations that selectively respond to specific categories. Brain Activation Control Through Image Variation (BrainACTIV) is a method for manipulating a reference image to enhance or decrease activity in a target cortical region using pretrained diffusion models. Starting from a reference image allows for fine-grained and reliable offline identification of optimal visuo-semantic properties, as well as producing controlled stimuli for novel neuroimaging studies.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-2447\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/8b.png\" alt=\"\" width=\"492\" height=\"442\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/8b.png 416w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/8b-300x270.png 300w\" sizes=\"(max-width: 492px) 100vw, 492px\" \/><\/p>\n<h4><\/h4>\n<p><strong><i>8. Exponential Family Variational Flow Matching for Tabular Data Generation<\/i><\/strong><\/p>\n<p>Authors- <i><strong><span class=\"outlook-search-highlight\" data-markjs=\"true\"><span class=\"outlook-search-highlight\" data-markjs=\"true\">Andr\u00e9s<\/span><\/span> Guzm\u00e1n-Cordero<\/strong>, Floor Eijkelboom and Jan-Willem van de Meent<\/i><\/p>\n<p>Conference &#8211; <strong>The Forty-Second International Conference on Machine Learning (ICML 2025)<\/strong><\/p>\n<p>Summary- <i>We<span class=\"x_note-content-value\">\u00a0extend variational flow matching by leveraging exponential family distributions to handle heterogeneous data types efficiently, and further show a prototype of this framework, TabbyFlow, to showcase its use in tabular data generation.<\/span><\/i><\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter wp-image-2459\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/EF-VFM-Visual-Abstract.jpg\" alt=\"\" width=\"882\" height=\"303\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/EF-VFM-Visual-Abstract.jpg 1152w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/EF-VFM-Visual-Abstract-300x103.jpg 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/EF-VFM-Visual-Abstract-1024x352.jpg 1024w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/EF-VFM-Visual-Abstract-768x264.jpg 768w\" sizes=\"(max-width: 882px) 100vw, 882px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>9. Gridded Transformer Neural Processes for Spatio-Temporal Data<\/strong><\/em><\/p>\n<p>Authors &#8211; Matthew Ashman*, Cristiana Diaconu*, <strong>Eric Langezaal*<\/strong>, Adrian Weller, Richard Turner<\/p>\n<p>Conference &#8211; <strong>The Forty-Second International Conference on Machine Learning (ICML 2025)<\/strong><\/p>\n<p>Summary &#8211;\u00a0Effective modelling of large-scale spatio-temporal\u00a0datasets is essential for many domains, yet existing approaches often impose rigid constraints on\u00a0the input data, such as requiring them to lie on\u00a0fixed-resolution grids. With the rise of foundation\u00a0models, the ability to process diverse, heterogenous\u00a0data structures is becoming increasingly\u00a0important. Neural processes (NPs), particularly transformer neural processes (TNPs), offer a\u00a0promising framework for such tasks, but struggle\u00a0to scale to large spatio-temporal datasets due to\u00a0the lack of an efficient attention mechanism. To\u00a0address this, we introduce gridded pseudo-token\u00a0TNPs which employ specialised encoders and decoders to handle unstructured data and utilise a\u00a0processor comprising gridded pseudo-tokens with\u00a0efficient attention mechanisms. Furthermore, we\u00a0develop equivariant gridded TNPs for applications\u00a0where exact or approximate translation equivariance\u00a0is a useful inductive bias, improving accuracy and training efficiency. Our method consistently outperforms a range of strong baselines in\u00a0various synthetic and real-world regression tasks\u00a0involving large-scale data, while maintaining competitive computational efficiency. Experiments\u00a0with weather data highlight the potential of gridded TNPs and serve as just one example of a\u00a0domain where they can have a significant impact.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-2443 aligncenter\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5.png\" alt=\"\" width=\"844\" height=\"340\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5.png 2394w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5-300x121.png 300w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5-1024x413.png 1024w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5-768x310.png 768w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5-1536x620.png 1536w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/5-2048x826.png 2048w\" sizes=\"(max-width: 844px) 100vw, 844px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p><em><strong>10. $\\sigma$-Maximal Ancestral Graphs<\/strong><\/em><\/p>\n<p>Authors-<strong>Binghua Yao<\/strong>, Joris M. Mooij<\/p>\n<p>Conference- <strong>The Forty-first Conference on Uncertainty in Artificial Intelligence (UAI 2025)<\/strong><\/p>\n<p>Summary- Graphs\u2014networks of nodes connected by edges\u2014are a powerful tool for understanding relationships between variables. In many fields such as biology, economics, and computer science, we use causal graphs to represent how one variable can influence another. A commonly used type of graph for this is called a DAG (Directed Acyclic Graph), where arrows represent cause-and-effect, and cycles (like feedback loops) are not allowed.<\/p>\n<p>However, real-world systems often include hidden (latent) variables and selection bias, and may even involve feedback\u2014like when a system\u2019s output affects its input. To represent such systems abstractly, researchers use more general graphs called MAGs (Maximal Ancestral Graphs). These are great for modeling hidden variables, but they cannot handle cycles, limiting their use in more complex situations.<\/p>\n<p>In this thesis, we fix that limitation. We introduce a new kind of graph, called a $\\sigma$-MAG ($\\sigma$-Maximal Ancestral Graph), which can represent cyclic causal systems while still accounting for hidden variables. These new graphs generalize MAGs and make it possible to model more realistic systems where feedback and information flow matter.<\/p>\n<p>We also study the mathematical properties of these new graphs\u2014such as when two different graphs actually represent the same system (this is called Markov equivalence)\u2014and lay the groundwork for using them in future algorithms for discovering causal relationships from data.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"wp-image-2455 alignleft\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/illustration1.png\" alt=\"\" width=\"350\" height=\"518\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/illustration1.png 427w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/illustration1-203x300.png 203w\" sizes=\"(max-width: 350px) 100vw, 350px\" \/><\/p>\n<p style=\"text-align: center;\"><img loading=\"lazy\" decoding=\"async\" class=\"alignleft wp-image-2454\" src=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/illustration2.png\" alt=\"\" width=\"291\" height=\"497\" srcset=\"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/illustration2.png 465w, https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-content\/uploads\/2025\/06\/illustration2-176x300.png 176w\" sizes=\"(max-width: 291px) 100vw, 291px\" \/><\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n<p>&nbsp;<\/p>\n","protected":false},"featured_media":0,"template":"","news-or-blog":[16],"class_list":["post-2434","news-","type-news-","status-publish","hentry","news-or-blog-blogs"],"_links":{"self":[{"href":"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-json\/wp\/v2\/news-\/2434","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-json\/wp\/v2\/news-"}],"about":[{"href":"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-json\/wp\/v2\/types\/news-"}],"wp:attachment":[{"href":"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-json\/wp\/v2\/media?parent=2434"}],"wp:term":[{"taxonomy":"news-or-blog","embeddable":true,"href":"https:\/\/ivi.fnwi.uva.nl\/ellis\/wp-json\/wp\/v2\/news-or-blog?post=2434"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}