Scientific Publication | Page 3
Revisiting Supervision for Continual Representation Learning
Abstract: In the field of continual learning, models are designed to learn tasks one after the other. While most research has centered on supervised continual learning, there is a growing interest in unsupervised continual learning, which makes use of the vast amounts of unlabeled data. Recent studies have highlighted the strengths of unsupervised methods, particularly self-supervised learning, in providing robust representations.
The improved transferability of those representations built with selfsupervised methods is often associated with the role played by the multilayer perceptron projector. In this work, we depart from this observation
and reexamine the role of supervision in continual representation learning. We reckon that additional information, such as human annotations, should not deteriorate the quality of representations. Our findings show that supervised models when enhanced with a multi-layer perceptron head, can outperform self-supervised models in continual representation learning. This highlights the importance of the multi-layer perceptron projector in shaping feature transferability across a sequence of tasks in continual learning.
Type of Publication: conference paper
Type of Conference: The 18th European Conference on Computer Vision (ECCV), Milano, 2024
Authors: Marczak, Daniel; Cygert, Sebastian; Trzciński, Tomasz; Twardowski, Bartłomiej
Category Adaptation Meets Projected Distillation in Generalized Continual Category Discovery
Abstract: Generalized Continual Category Discovery (GCCD) tackles learning from sequentially arriving, partially labeled datasets while uncovering new categories. Traditional methods depend on feature distillation to prevent forgetting the old knowledge. However, this strategy restricts the model’s ability to adapt and effectively distinguish new categories. To address this, we introduce a novel technique integrating a learnable
projector with feature distillation, thus enhancing model adaptability without sacrificing past knowledge. The resulting distribution shift of the previously learned categories is mitigated with the auxiliary category
adaptation network. We demonstrate that while each component offers modest benefits individually, their combination – dubbed CAMP (Category Adaptation Meets Projected distillation) – significantly improves the
balance between learning new information and retaining old. CAMP exhibits superior performance across several GCCD and Class Incremental Learning scenarios. The code is available on Github.
Type of Publication: conference paper
Type of Conference: The 18th European Conference on Computer Vision (ECCV) , Milano, 2024
Authors: Rypeść, Grzegorz; Marczak, Daniel; Cygert, Sebastian; Trzciński, Tomasz; Twardowski
AdaGlimpse: Active Visual Exploration with Arbitrary Glimpse Position and Scale
Abstract: Active Visual Exploration (AVE) is a task that involves dynamically selecting observations (glimpses), which is critical to facilitate comprehension and navigation within an environment. While modern AVE methods have demonstrated impressive performance, they are constrained to fixed-scale glimpses from rigid grids. In contrast, existing mobile platforms equipped with optical zoom capabilities can capture glimpses of arbitrary positions and scales. To address this gap between software and hardware capabilities, we introduce AdaGlimpse. It uses Soft Actor-Critic, a reinforcement learning algorithm tailored for exploration tasks, to select glimpses of arbitrary position and scale. This approach enables our model to rapidly establish a general awareness of the environment before zooming in for detailed analysis. Experimental results demonstrate that AdaGlimpse surpasses previous methods across various visual tasks while maintaining greater applicability in realistic AVE scenarios.
Type of Publication: publication
Type of Conference: The 18th European Conference on Computer Vision (ECCV) , Milano, 2024
Authors: Annapureddy, Ravinithesh; Fornaroli, Alessandro; Gatica-Perez, Daniel
Generative AI Literacy: Twelve Defining Competencies
Abstract:
This paper introduces a competency-based model for generative artificial intelligence (AI) literacy covering essential skills and knowledge areas necessary to interact with generative AI. The competencies range from foundational AI literacy to prompt engineering and programming skills, including ethical and legal considerations. These twelve competencies offer a framework for individuals, policymakers, government officials, and educators looking to navigate and take advantage of the potential of generative AI responsibly. Embedding these competencies into educational programs and professional training initiatives can equip individuals to become responsible and informed users and creators of generative AI. The competencies follow a logical progression and serve as a roadmap for individuals seeking to get familiar with generative AI and for researchers and policymakers to develop assessments, educational programs, guidelines, and regulations.
Type of Publication: publication
Authors: Annapureddy, Ravinithesh; Fornaroli, Alessandro; Gatica-Perez, Daniel
Trading Volume Maximization with Online Learning
Abstract: We explore brokerage between traders in an online learning framework. At any round t, two traders meet to exchange an asset, provided the exchange is mutually beneficial. The broker proposes a trading price, and each trader tries to sell their asset or buy the asset from the other party, depending on whether the price is higher or lower than their private valuations. A trade happens if one trader is willing to sell and the other is willing to buy at the proposed price.
Previous work provided guidance to a broker aiming at enhancing traders’ total earnings by maximizing the gain from trade, defined as the sum of the traders’ net utilities after each interaction. In contrast, we investigate how the broker should behave to maximize the trading volume, i.e., the total number of trades.
We model the traders’ valuations as an i.i.d. process with an unknown distribution. If the traders’ valuations are revealed after each interaction (full-feedback), and the traders’ valuations cumulative distribution function (cdf) is continuous, we provide an algorithm achieving logarithmic regret and show its optimality up to constant factors.
If only their willingness to sell or buy at the proposed price is revealed after each interaction (2-bitfeedback), we provide an algorithm achieving polylogarithmic regret when the traders’ valuations cdf is Lipschitz and show that this rate is near-optimal.
We complement our results by analyzing the implications of dropping the regularity assumptions on the unknown traders’ valuations cdf. If we drop the continuous cdf assumption, the regret rate degrades to Θ(√T) in the full-feedback case, where T is the time horizon. If we drop the Lipschitz cdf assumption, learning becomes impossible in the 2-bit feedback case.
Type of Publication: publication
Authors: Tommaso Cesari; Roberto Colomboni
A Contextual Online Learning Theory of Brokerage
Abstract: We study the role of contextual information in the online learning problem of brokerage between traders. At each round, two traders arrive with secret valuations about an asset they wish to trade. The broker suggests a trading price based on contextual data about the asset. Then, the traders decide to buy or sell depending on whether their valuations are higher or lower than the brokerage price. We assume the market value of traded assets is an unknown line ar function of a d-dimensional vector representing the contextual information available to the broker. Additionally, we model traders’ valuations as independent bounded zero-mean perturbations of the asset’s market value, allowing for potentially different unknown distributions across traders and time steps. Consistently with the existing online learning literature, we evaluate the performance of a learning algorithm with the regret with respect to the gain from trade. If the noise distributions admit densities bounded by someconstant L, then, for any time horizon T:
- If the agents’ valuations are revealed after each interact ion, we provide an algorithm achieving O(LdlnT) regret, and show a corresponding matching lower bound of Ω(LdlnT).
- If only their willingness to sell or buy at the proposed price is revealed after each interaction, we provide an algorithm achieving O(√LdTlnT) regret, and show that this rate is optimal (up to logarithmic factors), via a lower bound of Ω(√LdT).
Type of Publication: publication
Authors: François Bachoc; Tommaso Cesari; Roberto Colomboni
Fair Online Bilateral Trade
Abstract: In online bilateral trade, a platform posts prices to incoming pairs of buyers and sellers that have private valuations for a certain good. If the price is lower than the buyers’ valuation and higher than the sellers’ valuation, then a trade takes place. Previous work focused on the platform perspective, with the goal of setting prices maximizing the gain from trade (the sum of sellers’ and buyers’ utilities). Gain from trade is, however, potentially unfair to traders, as they may receive highly uneven shares of the total utility. In this work we enforce fairness by rewarding the platform with the fair gain from trade, defined as the minimum between sellers’ and buyers’ utilities. After showing that any no-regret learning algorithm designed to maximize the sum of the utilities may fail badly with fair gain from trade, we present our main contribution: a complete characterization of the regret regimes for fair gain from trade when, after each interaction, the platform only learns whether each trader accepted the current price. Specifically, we prove the following regret bounds: Θ(ln T ) in the deterministic setting, Ω(T)in the stochastic setting, and ̃Θ(T2~3)in the stochastic setting when sellers’ and buyers’ valuations are independent of each other. We conclude by providing tight regret bounds when, after each interaction, the platform is allowed to observe the true traders’ valuations.
Type of Publication: publication
Authors: François Bachoc; Nicolò Cesa-Bianchi; Tommaso Cesari; Roberto Colomboni
A deep cut into Split Federated Self-supervised Learning
Abstract: Collaborative self-supervised learning has recently become feasible in highly distributed environments by dividing the network layers between client devices and a central server. However, state-of-the-art methods, such as MocoSFL, are optimized for network division at the initial layers, which decreases the protection of the client data and increases communication overhead. In this paper, we demonstrate that splitting depth is crucial for maintaining privacy and communication efficiency in distributed training. We also show that MocoSFL suffers from a catastrophic quality deterioration for the minimal communication overhead. As a remedy, we introduce Momentum-Aligned contrastive Split Federated Learning (MonAcoSFL), which aligns online and momentum client models during training procedure. Consequently, we achieve state-of-the-art accuracy while significantly reducing the communication overhead, making MonAcoSFL more practical in real-world scenarios.
Type of Publication: publication
Title of Journal: International Conference on Learning Representations (ICLR), Vienna Austria, 7-11.05.2024
Authors: Marcin Przewięźlikowski; Marcin Osial; Bartosz Zieliński; Marek Śmieja
Access the open software and datasets produced by ELIAS!
ELIAS aims at establishing Europe as a leader in Artificial Intelligence (AI) research that drives sustainable innovation and economic development.
We will create a Network of Excellence connecting researchers in academia with practitioners in the industry to differentiate Europe as a region where AI research builds towards a sustainable long-term future for our planet, contributes to a cohesive society, and respects individual preferences and rights.