People

Isabel Valera
Saarland Informatics Campus
Building E1 1, R. 225
For administrative services, contact ml-office@lists.saarland-informatics-campus.de
To apply for PhD/PostDoc/HiWi/Thesis, see the information on the “Positions” page for the correct e-mail to use.
Otherwise, contact ivalera@cs.uni-saarland.de.
About me
I am a full Professor on Machine Learning at the Department of Computer Science of Saarland University (Saarbrücken, Germany), and Adjunct Faculty at MPI for Software Systems in Saarbrücken (Saarbrücken, Germany).
I am a fellow of the European Laboratory for Learning and Intelligent Systems (ELLIS), where I am part of the Robust Machine Learning Program and of the Saarbrücken Artificial Intelligence & Machine learning (Sam) Unit.
Prior to this, I was an independent group leader at the MPI for Intelligent Systems in Tübingen (Germany) until the end of the year. I have held a German Humboldt Post-Doctoral Fellowship, and a “Minerva fast track” fellowship from the Max Planck Society. I obtained my PhD in 2014 and MSc degree in 2012 from the University Carlos III in Madrid (Spain), and worked as postdoctoral researcher at the MPI for Software Systems (Germany) and at the University of Cambridge (UK).
Publications
2026
Majumdar, Ayan; Kanubala, Deborah Dormah; Gupta, Kavya; Valera, Isabel
A Causal Framework to Measure and Mitigate Non-binary Treatment Discrimination Journal Article
In: CoRR, vol. abs/2503.22454, 2026.
@article{DBLP:journals/corr/abs-2503-22454,
title = {A Causal Framework to Measure and Mitigate Non-binary Treatment Discrimination},
author = {Ayan Majumdar and Deborah Dormah Kanubala and Kavya Gupta and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2503.22454},
doi = {10.48550/ARXIV.2503.22454},
year = {2026},
date = {2026-03-19},
urldate = {2026-03-19},
journal = {CoRR},
volume = {abs/2503.22454},
abstract = {Fairness studies of algorithmic decision-making systems often simplify complex decision processes, such as bail or lending decisions, into binary classification tasks (e.g., approve or not approve). However, these approaches overlook that such decisions are not inherently binary; they also involve non-binary treatment decisions (e.g., loan or bail terms) that can influence the downstream outcomes (e.g., loan repayment or reoffending). We argue that treatment decisions are integral to the decision-making process and, therefore, should be central to fairness analyses. Consequently, we propose a causal framework that extends and complements existing fairness notions by explicitly distinguishing between decision-subjects’ covariates and the treatment decisions. Our framework leverages path-specific counterfactual reasoning to: (i) measure treatment disparity and its downstream effects in historical data; and (ii) mitigate the impact of past unfair treatment decisions when automating decision-making. We use our framework to empirically analyze four widely used loan approval datasets to reveal potential disparity in non-binary treatment decisions and their discriminatory impact on outcomes, highlighting the need to incorporate treatment decisions in fairness assessments. Finally, by intervening in treatment decisions, we show that our framework effectively mitigates treatment discrimination from historical loan approval data to ensure fair risk score estimation and (non-binary) decision-making processes that benefit all stakeholders.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Mayank Jobanputra Yifan Wang, Ji-Ung Lee
Bridging Fairness and Explainability: Can Input-Based Explanations Promote Fairness in Hate Speech Detection? Journal Article
In: 2026.
@article{nokey,
title = {Bridging Fairness and Explainability: Can Input-Based Explanations Promote Fairness in Hate Speech Detection?},
author = {Yifan Wang, Mayank Jobanputra, Ji-Ung Lee, Soyoung Oh, Isabel Valera, Vera Demberg},
doi = { https://doi.org/10.48550/arXiv.2509.22291},
year = {2026},
date = {2026-02-11},
urldate = {2026-02-11},
abstract = {Natural language processing (NLP) models often replicate or amplify social bias from training data, raising concerns about fairness. At the same time, their black-box nature makes it difficult for users to recognize biased predictions and for developers to effectively mitigate them. While some studies suggest that input-based explanations can help detect and mitigate bias, others question their reliability in ensuring fairness. Existing research on explainability in fair NLP has been predominantly qualitative, with limited large-scale quantitative analysis. In this work, we conduct the first systematic study of the relationship between explainability and fairness in hate speech detection, focusing on both encoder- and decoder-only models. We examine three key dimensions: (1) identifying biased predictions, (2) selecting fair models, and (3) mitigating bias during model training. Our findings show that input-based explanations can effectively detect biased predictions and serve as useful supervision for reducing bias during training, but they are unreliable for selecting fair models among this http URL http://candidates.our/ code is available at this URL https://github.com/Ewanwong/fairness_x_explainability},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Uth, Richard; Niemitz, Nelli; Valera, Isabel; Langer, Markus
Personalizing explanations in AI-based decisions: The effects of personalization and (Mis)aligning with individual preferences Journal Article
In: Computers in Human Behavior, 2026.
@article{Uth2025PersonalizingEI,
title = {Personalizing explanations in AI-based decisions: The effects of personalization and (Mis)aligning with individual preferences},
author = {Richard Uth and Nelli Niemitz and Isabel Valera and Markus Langer},
url = {https://api.semanticscholar.org/CorpusID:283171660},
year = {2026},
date = {2026-02-03},
urldate = {2025-01-01},
journal = {Computers in Human Behavior},
abstract = {The increasing reliance on AI-based decision-making in high-stakes contexts underscores the need for transparency and justice. Here, negative outcomes drive individuals affected by AI-based decisions to seek actionable explanations that enable them to realize what they can do to achieve a better future outcome. However, actionability is subjective, varying across individuals and contexts. Personalization of explanations has been proposed to address this variability, but insights on personalized explanation processes, their potential, and challenges are scarce. This paper investigates the impact of personalization and (mis)alignment with individual needs and preferences in explanations for AI-based decisions through an experimental online study simulating denied loan applications. In a within-participants design (N=255), participants ranked the actionability of decision-relevant features and experienced five explanation conditions: personalized directive explanations based on the most, second most, or least actionable feature (as ranked by participants); a non-personalized directive explanation highlighting a random feature; and no explanation. In line with justice theory, our results show that any explanation was better than none, and that personalized explanations led to more favorable reactions than non-personalized explanations, enhancing perceptions of justice and attractiveness of the bank. Closer alignment with preferences had only small positive effects, mainly for attractiveness. These findings highlight that even simple ranking-based approaches can make explanations more effective and accessible without requiring technical expertise while cautioning against offering superficial control. This study provides insights into the effects of ranking-based personalization, informing the design of explainability tailored to diverse user needs and addressing ethical and practical considerations in personalization.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Vo, Huyen; Valera, Isabel
Hellinger Multimodal Variational Autoencoders Miscellaneous
2026.
@misc{vo2026hellingermultimodalvariationalautoencoders,
title = {Hellinger Multimodal Variational Autoencoders},
author = {Huyen Vo and Isabel Valera},
url = {https://arxiv.org/abs/2601.06572},
year = {2026},
date = {2026-01-01},
urldate = {2026-01-01},
abstract = {Multimodal variational autoencoders (VAEs) are widely used for weakly supervised generative learning with multiple modalities. Predominant methods aggregate unimodal inference distributions using either a product of experts (PoE), a mixture of experts (MoE), or their combinations to approximate the joint posterior. In this work, we revisit multimodal inference through the lens of probabilistic opinion pooling, an optimization-based approach. We start from Hölder pooling with , which corresponds to the unique symmetric member of the family, and derive a moment-matching approximation, termed Hellinger. We then leverage such an approximation to propose HELVAE, a multimodal VAE that avoids sub-sampling, yielding an efficient yet effective model that: (i) learns more expressive latent representations as additional modalities are observed; and (ii) empirically achieves better trade-offs between generative coherence and quality, outperforming state-of-the-art multimodal VAE models.},
keywords = {},
pubstate = {published},
tppubtype = {misc}
}
Müller, Nicola J.; Oster, Moritz; Valera, Isabel; Hoffmann, Jörg; Gros, Timo P.
Per-Domain Generalizing Policies: On Learning Efficient and Robust Q-Value Functions (Extended Version with Technical Appendix) Journal Article
In: CoRR, vol. abs/2603.17544, 2026.
@article{DBLP:journals/corr/abs-2603-17544,
title = {Per-Domain Generalizing Policies: On Learning Efficient and Robust Q-Value Functions (Extended Version with Technical Appendix)},
author = {Nicola J. Müller and Moritz Oster and Isabel Valera and Jörg Hoffmann and Timo P. Gros},
url = {https://doi.org/10.48550/arXiv.2603.17544},
doi = {10.48550/ARXIV.2603.17544},
year = {2026},
date = {2026-01-01},
urldate = {2026-01-01},
journal = {CoRR},
volume = {abs/2603.17544},
abstract = {Learning per-domain generalizing policies is a key challenge in learning for planning. Standard approaches learn state-value functions represented as graph neural networks using supervised learning on optimal plans generated by a teacher planner. In this work, we advocate for learning Q-value functions instead. Such policies are drastically cheaper to evaluate for a given state, as they need to process only the current state rather than every successor. Surprisingly, vanilla supervised learning of Q-values performs poorly as it does not learn to distinguish between the actions taken and those not taken by the teacher. We address this by using regularization terms that enforce this distinction, resulting in Q-value policies that consistently outperform state-value policies across a range of 10 domains and are competitive with the planner LAMA-first.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2025
Eisenhut, Jan; Fivser, Daniel; Valera, Isabel; Hoffmann, J¨org
On Picking Good Policies: Leveraging Action-Policy Testing in Policy Training Journal Article
In: Proceedings of the International Conference on Automated Planning and Scheduling, 2025.
@article{Eisenhut2025OnPG,
title = {On Picking Good Policies: Leveraging Action-Policy Testing in Policy Training},
author = {Jan Eisenhut and Daniel Fivser and Isabel Valera and J¨org Hoffmann},
url = {https://api.semanticscholar.org/CorpusID:281347744},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Proceedings of the International Conference on Automated Planning and Scheduling},
abstract = {Testing is a natural approach to assess the quality of learned action policies π. Prior work introduced policy testing in AI planning as searching for bugs in π, that is, states where π is sub-optimal with respect to a given testing objective. Beyond quality assurance, an obvious application of these methods is policy selection: given several π to choose from, we can use testing to select the "least buggy" one. Here, we integrate testing-based policy selection into the training process. This includes making more informed decisions when selecting the final policy after training, as well as choosing more promising intermediate policies during the training process. Our experiments with ASNets action policies show that integrating testing allows us to more reliably obtain good-quality policies.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Eniser, Hasan Ferit; Lin, Songtuan; Müller, Nicola; Isychev, Anastasia; Wüstholz, Valentin; Valera, Isabel; Hoffmann, J¨org; Christakis, Maria
Using Action-Policy Testing in RL to Reduce the Number of Bugs Journal Article
In: Proceedings of the International Symposium on Combinatorial Search, 2025.
@article{Eniser2025UsingAT,
title = {Using Action-Policy Testing in RL to Reduce the Number of Bugs},
author = {Hasan Ferit Eniser and Songtuan Lin and Nicola Müller and Anastasia Isychev and Valentin Wüstholz and Isabel Valera and J¨org Hoffmann and Maria Christakis},
url = {https://api.semanticscholar.org/CorpusID:280219509},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Proceedings of the International Symposium on Combinatorial Search},
abstract = {Reinforcement learning is becoming ever more prominent in solving combinatorial search problems, in particular ones where states are images. Prior work has devised action-policy testing methodology, that identifies so-called bug states where policy performance is sub-optimal. Here we show how to leverage this methodology during the RL process, using action-policy testing to find bugs and injecting those as alternate start states for the training runs. Running experiments across six 2D games, we find that our testing-guided training often achieves similar expected reward while reducing the number of bugs.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Javaloy, Adrián; Vergari, Antonio; Valera, Isabel
COPA: Comparing the Incomparable to Explore the Pareto Front Journal Article
In: CoRR, vol. abs/2503.14321, 2025.
@article{DBLP:journals/corr/abs-2503-14321,
title = {COPA: Comparing the Incomparable to Explore the Pareto Front},
author = {Adrián Javaloy and Antonio Vergari and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2503.14321},
doi = {10.48550/ARXIV.2503.14321},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2503.14321},
abstract = {In machine learning (ML), we often need to choose one among hundreds of trained ML models at hand, based on various objectives such as accuracy, robustness, fairness or scalability. However, it is often unclear how to compare, aggregate and, ultimately, trade-off these objectives, making it a time-consuming task that requires expert knowledge, as objectives may be measured in different units and scales. In this work, we investigate how objectives can be automatically normalized and aggregated to systematically help the user navigate their Pareto front. To this end, we make incomparable objectives comparable using their cumulative functions, approximated by their relative rankings. As a result, our proposed approach, COPA, can aggregate them while matching user-specific preferences, allowing practitioners to meaningfully navigate and search for models in the Pareto front. We demonstrate the potential impact of COPA in both model selection and benchmarking tasks across diverse ML areas such as fair ML, domain generalization, AutoML and foundation models, where classical ways to normalize and aggregate objectives fall short.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Almodóvar, Alejandro; Javaloy, Adrián; Parras, Juan; Zazo, Santiago; Valera, Isabel
DeCaFlow: A Deconfounding Causal Generative Model Journal Article
In: CoRR, vol. abs/2503.15114, 2025.
@article{DBLP:journals/corr/abs-2503-15114,
title = {DeCaFlow: A Deconfounding Causal Generative Model},
author = {Alejandro Almodóvar and Adrián Javaloy and Juan Parras and Santiago Zazo and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2503.15114},
doi = {10.48550/ARXIV.2503.15114},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2503.15114},
abstract = {We introduce DeCaFlow, a deconfounding causal generative model. Training once per dataset using just observational data and the underlying causal graph, DeCaFlow enables accurate causal inference on continuous variables under the presence of hidden confounders. Specifically, we extend previous results on causal estimation under hidden confounding to show that a single instance of DeCaFlow provides correct estimates for all causal queries identifiable with do-calculus, leveraging proxy variables to adjust for the causal effects when do-calculus alone is insufficient. Moreover, we show that counterfactual queries are identifiable as long as their interventional counterparts are identifiable, and thus are also correctly estimated by DeCaFlow. Our empirical results on diverse settings (including the Ecoli70 dataset, with 3 independent hidden confounders, tens of observed variables and hundreds of causal queries) show that DeCaFlow outperforms existing approaches, while demonstrating its out-of-the-box applicability to any given causal graph},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Cinquini, Martina; Beretta, Isacco; Ruggieri, Salvatore; Valera, Isabel
A Practical Approach to Causal Inference over Time Proceedings Article
In: Walsh, Toby; Shah, Julie; Kolter, Zico (Ed.): AAAI-25, Sponsored by the Association for the Advancement of Artificial Intelligence, February 25 - March 4, 2025, Philadelphia, PA, USA, pp. 14832–14839, AAAI Press, 2025.
@inproceedings{DBLP:conf/aaai/CinquiniBRV25,
title = {A Practical Approach to Causal Inference over Time},
author = {Martina Cinquini and Isacco Beretta and Salvatore Ruggieri and Isabel Valera},
editor = {Toby Walsh and Julie Shah and Zico Kolter},
url = {https://doi.org/10.1609/aaai.v39i14.33626},
doi = {10.1609/AAAI.V39I14.33626},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
booktitle = {AAAI-25, Sponsored by the Association for the Advancement of Artificial
Intelligence, February 25 - March 4, 2025, Philadelphia, PA, USA},
pages = {14832–14839},
publisher = {AAAI Press},
abstract = {In this paper, we focus on estimating the causal effect of an intervention over time on a dynamical system. To that end, we formally define causal interventions and their effects over time on discrete-time stochastic processes (DSPs). Then, we show under which conditions the equilibrium states of a DSP, both before and after a causal intervention, can be captured by a structural causal model (SCM). With such an equivalence at hand, we provide an explicit mapping from vector autoregressive models (VARs), broadly applied in econometrics, to linear, but potentially cyclic and/or affected by unmeasured confounders, SCMs. The resulting causal VAR framework allows us to perform causal inference over time from observational time series data. Our experiments on synthetic and real-world datasets show that the proposed framework achieves strong performance in terms of observational forecasting while enabling accurate estimation of the causal effect of interventions on dynamical systems. We demonstrate, through a case study, the potential practical questions that can be addressed using the proposed causal VAR framework.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Peis, Ignacio; Koyuncu, Batuhan; Valera, Isabel; Frellsen, Jes
Hyper-Transforming Latent Diffusion Models Journal Article
In: CoRR, vol. abs/2504.16580, 2025.
@article{DBLP:journals/corr/abs-2504-16580,
title = {Hyper-Transforming Latent Diffusion Models},
author = {Ignacio Peis and Batuhan Koyuncu and Isabel Valera and Jes Frellsen},
url = {https://doi.org/10.48550/arXiv.2504.16580},
doi = {10.48550/ARXIV.2504.16580},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2504.16580},
abstract = {We introduce a novel generative framework for functions by integrating Implicit Neural Representations (INRs) and Transformer-based hypernetworks into latent variable models. Unlike prior approaches that rely on MLP-based hypernetworks with scalability limitations, our method employs a Transformer-based decoder to generate INR parameters from latent variables, addressing both representation capacity and computational efficiency. Our framework extends latent diffusion models (LDMs) to INR generation by replacing standard decoders with a Transformer-based hypernetwork, which can be trained either from scratch or via hyper-transforming: a strategy that fine-tunes only the decoder while freezing the pre-trained latent space. This enables efficient adaptation of existing generative models to INR-based representations without requiring full retraining. We validate our approach across multiple modalities, demonstrating improved scalability, expressiveness, and generalization over existing INR-based generative models. Our findings establish a unified and flexible framework for learning structured function representations.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Koyuncu, Batuhan; DeVries, Rachael; Winther, Ole; Valera, Isabel
Temporal Variational Implicit Neural Representations Journal Article
In: CoRR, vol. abs/2506.01544, 2025.
@article{DBLP:journals/corr/abs-2506-01544,
title = {Temporal Variational Implicit Neural Representations},
author = {Batuhan Koyuncu and Rachael DeVries and Ole Winther and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2506.01544},
doi = {10.48550/ARXIV.2506.01544},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2506.01544},
abstract = {We introduce Temporal Variational Implicit Neural Representations (TV-INRs), a probabilistic framework for modeling irregular multivariate time series that enables efficient individualized imputation and forecasting. By integrating implicit neural representations with latent variable models, TV-INRs learn distributions over time-continuous generator functions conditioned on signal-specific covariates. Unlike existing approaches that require extensive training, fine-tuning or meta-learning, our method achieves accurate individualized predictions through a single forward pass. Our experiments demonstrate that with a single TV-INRs instance, we can accurately solve diverse imputation and forecasting tasks, offering a computationally efficient and scalable solution for real-world applications. TV-INRs excel especially in low-data regimes, where it outperforms existing methods by an order of magnitude in mean squared error for imputation task.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Kalampalikis, Nektarios; Gupta, Kavya; Vitanov, Georgi; Valera, Isabel
Towards Reasonable Concept Bottleneck Models Journal Article
In: CoRR, vol. abs/2506.05014, 2025.
@article{DBLP:journals/corr/abs-2506-05014,
title = {Towards Reasonable Concept Bottleneck Models},
author = {Nektarios Kalampalikis and Kavya Gupta and Georgi Vitanov and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2506.05014},
doi = {10.48550/ARXIV.2506.05014},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2506.05014},
abstract = {In this paper, we propose textbf{C}oncept textbf{REA}soning textbf{M}odels (CREAM), a novel family of Concept Bottleneck Models (CBMs) that: (i) explicitly encodes concept-concept ({texttt{C-C}}) and concept-task ({texttt{C$rightarrow$Y}}) relationships to enforce a desired model reasoning; and (ii) use a regularized side-channel to achieve competitive task performance, while keeping high concept importance. Specifically, CREAM architecturally embeds (bi)directed concept-concept, and concept to task relationships specified by a human expert, while severing undesired information flows (e.g., to handle mutually exclusive concepts). Moreover, CREAM integrates a black-box side-channel that is regularized to encourage task predictions to be grounded in the relevant concepts, thereby utilizing the side-channel only when necessary to enhance performance. Our experiments show that: (i) CREAM mainly relies on concepts while achieving task performance on par with black-box models; and (ii) the embedded {texttt{C-C}} and {texttt{C$rightarrow$Y}} relationships ease model interventions and mitigate concept leakage.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Azime, Israel Abebe; Kanubala, Deborah Dormah; Afonja, Tejumade; Fritz, Mario; Valera, Isabel; Klakow, Dietrich; Slusallek, Philipp
Accept or Deny? Evaluating LLM Fairness and Performance in Loan Approval across Table-to-Text Serialization Approaches Journal Article
In: CoRR, vol. abs/2508.21512, 2025.
@article{DBLP:journals/corr/abs-2508-21512,
title = {Accept or Deny? Evaluating LLM Fairness and Performance in Loan Approval across Table-to-Text Serialization Approaches},
author = {Israel Abebe Azime and Deborah Dormah Kanubala and Tejumade Afonja and Mario Fritz and Isabel Valera and Dietrich Klakow and Philipp Slusallek},
url = {https://doi.org/10.48550/arXiv.2508.21512},
doi = {10.48550/ARXIV.2508.21512},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2508.21512},
abstract = {Large Language Models (LLMs) are increasingly employed in high-stakes decision-making tasks, such as loan approvals. While their applications expand across domains, LLMs struggle to process tabular data, ensuring fairness and delivering reliable predictions. In this work, we assess the performance and fairness of LLMs on serialized loan approval datasets from three geographically distinct regions: Ghana, Germany, and the United States. Our evaluation focuses on the model’s zero-shot and in-context learning (ICL) capabilities. Our results reveal that the choice of serialization format significantly affects both performance and fairness in LLMs, with certain formats such as GReaT and LIFT yielding higher F1 scores but exacerbating fairness disparities. Notably, while ICL improved model performance by 4.9-59.6% relative to zero-shot baselines, its effect on fairness varied considerably across datasets. Our work underscores the importance of effective tabular data representation methods and fairness-aware models to improve the reliability of LLMs in financial decision-making.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Gros, Timo P.; Müller, Nicola J.; Fiser, Daniel; Valera, Isabel; Wolf, Verena; Hoffmann, Jörg
Per-Domain Generalizing Policies: On Validation Instances and Scaling Behavior Journal Article
In: CoRR, vol. abs/2505.00439, 2025.
@article{DBLP:journals/corr/abs-2505-00439,
title = {Per-Domain Generalizing Policies: On Validation Instances and Scaling Behavior},
author = {Timo P. Gros and Nicola J. Müller and Daniel Fiser and Isabel Valera and Verena Wolf and Jörg Hoffmann},
url = {https://doi.org/10.48550/arXiv.2505.00439},
doi = {10.48550/ARXIV.2505.00439},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {CoRR},
volume = {abs/2505.00439},
abstract = {Recent work has shown that successful per-domain generalizing action policies can be learned. Scaling behavior, from small training instances to large test instances, is the key objective; and the use of validation instances larger than training instances is one key to achieve it. Prior work has used fixed validation sets. Here, we introduce a method generating the validation set dynamically, on the fly, increasing instance size so long as informative and this http://feasible.we/ also introduce refined methodology for evaluating scaling behavior, generating test instances systematically to guarantee a given confidence in coverage performance for each instance size. In experiments, dynamic validation improves scaling behavior of GNN policies in all 9 domains used.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Kanubala, Deborah Dormah; Valera, Isabel
On the Misalignment Between Legal Notions and Statistical Metrics of Intersectional Fairness Journal Article
In: Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, 2025.
@article{Kanubala2025OnTM,
title = {On the Misalignment Between Legal Notions and Statistical Metrics of Intersectional Fairness},
author = {Deborah Dormah Kanubala and Isabel Valera},
url = {https://api.semanticscholar.org/CorpusID:282175621},
year = {2025},
date = {2025-01-01},
urldate = {2025-01-01},
journal = {Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society},
abstract = {Intersectional (un)fairness, as conceptualized in legal and social theory, emphasizes the non-additive and structurally complex nature of discrimination against individuals at the intersection of multiple sensitive attributes (such as race, gender, etc). Recent works have proposed statistical metrics for intersectional fairness by estimating disparities across groups of individuals sharing two or more sensitive attributes. However, it is unclear if these metrics detect uniquely intersectional discrimination. We therefore pose the following question, Do current statistical intersectional metrics detect the non-additive discrimination highlighted by intersectionality theory? More specifically, to answer this, we run controlled synthetic data experiments that explicitly allow us to control for single, multiple, intersectional, and compounded forms of discrimination. Our analyses show that current statistical metrics for intersectional fairness behave more like multi-attribute disparity measures. Specifically, they respond more strongly to additive or compounded biases than to non-additive interaction effects. While they effectively capture disparities across multiple sensitive attributes, they often fail to detect uniquely intersectional discrimination. These findings reveal a fundamental misalignment between existing intersectional fairness metrics and the legal and theoretical foundations of intersectionality. We argue that if intersectional fairness metrics are to be deemed truly intersectional, they must be explicitly designed to account for the structural, non-additive nature of intersectional discrimination.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2024
Beretta, Isacco; Cinquini, Martina; Valera, Isabel
Enhancing Fairness Through Time-aware Recourse: a Pathway to Realistic Algorithmic Recommendations Proceedings Article
In: Cerrato, Mattia; Coronel, Alesia Vallenas; Ahrweiler, Petra; Loi, Michele; Pechenizkiy, Mykola; Tamò-Larrieux, Aurelia (Ed.): Proceedings of the 3rd European Workshop on Algorithmic Fairness, Mainz, Germany, July 1st to 3rd, 2024, CEUR-WS.org, 2024.
@inproceedings{DBLP:conf/ewaf/BerettaCV24,
title = {Enhancing Fairness Through Time-aware Recourse: a Pathway to Realistic Algorithmic Recommendations},
author = {Isacco Beretta and Martina Cinquini and Isabel Valera},
editor = {Mattia Cerrato and Alesia Vallenas Coronel and Petra Ahrweiler and Michele Loi and Mykola Pechenizkiy and Aurelia Tamò-Larrieux},
url = {https://ceur-ws.org/Vol-3908/paper_10.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {Proceedings of the 3rd European Workshop on Algorithmic Fairness,
Mainz, Germany, July 1st to 3rd, 2024},
volume = {3908},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
abstract = {Algorithmic Recourse (AR) addresses adverse outcomes in automated decision-making by offering actionable recommendations. However, current state-of-the-art methods overlook the interdependence of features and do not consider the temporal dimension. To fill this gap, time-car emerges as a pioneering approach that integrates temporal information. Building upon this formulation, this work investigates the context of fairness, specifically focusing on the implications for marginalized demographic groups. Since long wait times can significantly impact communities’ financial, educational, and personal lives, exploring how time-related factors affect the fair treatment of these groups is crucial to suggest potential
solutions to reduce the negative effects on minority populations. Our findings set the stage for more equitable AR techniques sensitive to individual needs, ultimately fostering fairer suggestions.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
solutions to reduce the negative effects on minority populations. Our findings set the stage for more equitable AR techniques sensitive to individual needs, ultimately fostering fairer suggestions.
Kanubala, Deborah Dormah; Valera, Isabel; Gupta, Kavya
Fairness Beyond Binary Decisions: a Case Study on German Credit Proceedings Article
In: Cerrato, Mattia; Coronel, Alesia Vallenas; Ahrweiler, Petra; Loi, Michele; Pechenizkiy, Mykola; Tamò-Larrieux, Aurelia (Ed.): Proceedings of the 3rd European Workshop on Algorithmic Fairness, Mainz, Germany, July 1st to 3rd, 2024, CEUR-WS.org, 2024.
@inproceedings{DBLP:conf/ewaf/KanubalaVG24,
title = {Fairness Beyond Binary Decisions: a Case Study on German Credit},
author = {Deborah Dormah Kanubala and Isabel Valera and Kavya Gupta},
editor = {Mattia Cerrato and Alesia Vallenas Coronel and Petra Ahrweiler and Michele Loi and Mykola Pechenizkiy and Aurelia Tamò-Larrieux},
url = {https://ceur-ws.org/Vol-3908/paper_15.pdf},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {Proceedings of the 3rd European Workshop on Algorithmic Fairness,
Mainz, Germany, July 1st to 3rd, 2024},
volume = {3908},
publisher = {CEUR-WS.org},
series = {CEUR Workshop Proceedings},
abstract = {Data-driven approaches are increasingly used to (partially) automate decision-making in credit scoring
by predicting whether an applicant is “creditworthy or not” based on a set of features about the applicant,
such as age and income, along with what we refer here to as treatment decisions, e.g., loan amount and
duration. Existing data-driven approaches for automating and evaluating the accuracy and fairness of
such credit decisions ignore that treatment decisions (here, loan terms) are part of the decision and
thus may be subject to discrimination. This discrimination can propagate to the final outcome (repaid
or not) of positive decisions (granted loans). In this extended abstract, we rely on causal reasoning
and a broadly studied fair machine-learning dataset, the German credit, to i) show that the current fair
data-driven approach neglects discrimination in treatment decisions (i.e., loan terms) and its downstream
consequences on the decision outcome (i.e., ability to repay); and ii) argue for the need to move beyond
binary decisions in fair data-driven decision-making in consequential settings like credit scoring},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
by predicting whether an applicant is “creditworthy or not” based on a set of features about the applicant,
such as age and income, along with what we refer here to as treatment decisions, e.g., loan amount and
duration. Existing data-driven approaches for automating and evaluating the accuracy and fairness of
such credit decisions ignore that treatment decisions (here, loan terms) are part of the decision and
thus may be subject to discrimination. This discrimination can propagate to the final outcome (repaid
or not) of positive decisions (granted loans). In this extended abstract, we rely on causal reasoning
and a broadly studied fair machine-learning dataset, the German credit, to i) show that the current fair
data-driven approach neglects discrimination in treatment decisions (i.e., loan terms) and its downstream
consequences on the decision outcome (i.e., ability to repay); and ii) argue for the need to move beyond
binary decisions in fair data-driven decision-making in consequential settings like credit scoring
Sánchez-Martin, Pablo; Utz, Sonja; Valera, Isabel
Exploring the Boundaries of Ambient Awareness in Twitter Journal Article
In: CoRR, vol. abs/2403.17776, 2024.
@article{DBLP:journals/corr/abs-2403-17776,
title = {Exploring the Boundaries of Ambient Awareness in Twitter},
author = {Pablo Sánchez-Martin and Sonja Utz and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2403.17776},
doi = {10.48550/ARXIV.2403.17776},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
journal = {CoRR},
volume = {abs/2403.17776},
abstract = {Ambient awareness refers to the ability of social media users to obtain knowledge about who knows what (i.e., users' expertise) in their network, by simply being exposed to other users' content (e.g, tweets on Twitter). Previous work, based on user surveys, reveals that individuals self-report ambient awareness only for parts of their networks. However, it is unclear whether it is their limited cognitive capacity or the limited exposure to diagnostic tweets (i.e., online content) that prevents people from developing ambient awareness for their complete network. In this work, we focus on in-wall ambient awareness (IWAA) in Twitter and conduct a two-step data-driven analysis, that allows us to explore to which extent IWAA is likely, or even possible. First, we rely on reactions (e.g., likes), as strong evidence of users being aware of experts in Twitter. Unfortunately, such strong evidence can be only measured for active users, which represent the minority in the network. Thus to study the boundaries of IWAA to a larger extent, in the second part of our analysis, we instead focus on the passive exposure to content generated by other users -- which we refer to as in-wall visibility. This analysis shows that (in line with citet{levordashka2016ambient}) only for a subset of users IWAA is plausible, while for the majority it is unlikely, if even possible, to develop IWAA. We hope that our methodology paves the way for the emergence of data-driven approaches for the study of ambient awareness.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Sánchez-Martin, Pablo; Khan, Kinaan Aamir; Valera, Isabel
Improving the interpretability of GNN predictions through conformal-based graph sparsification Journal Article
In: CoRR, vol. abs/2404.12356, 2024.
@article{DBLP:journals/corr/abs-2404-12356,
title = {Improving the interpretability of GNN predictions through conformal-based graph sparsification},
author = {Pablo Sánchez-Martin and Kinaan Aamir Khan and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2404.12356},
doi = {10.48550/ARXIV.2404.12356},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
journal = {CoRR},
volume = {abs/2404.12356},
abstract = {Graph Neural Networks (GNNs) have achieved state-of-the-art performance in solving graph classification tasks. However, most GNN architectures aggregate information from all nodes and edges in a graph, regardless of their relevance to the task at hand, thus hindering the interpretability of their predictions. In contrast to prior work, in this paper we propose a GNN emph{training} approach that jointly i) finds the most predictive subgraph by removing edges and/or nodes -- -emph{without making assumptions about the subgraph structure} -- while ii) optimizing the performance of the graph classification task. To that end, we rely on reinforcement learning to solve the resulting bi-level optimization with a reward function based on conformal predictions to account for the current in-training uncertainty of the classifier. Our empirical results on nine different graph classification datasets show that our method competes in performance with baselines while relying on significantly sparser subgraphs, leading to more interpretable GNN-based predictions.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Rateike, Miriam; Valera, Isabel; Forré, Patrick
Designing Long-term Group Fair Policies in Dynamical Systems Proceedings Article
In: The 2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024, Rio de Janeiro, Brazil, June 3-6, 2024, pp. 20–50, ACM, 2024.
@inproceedings{DBLP:conf/fat/RateikeVF24,
title = {Designing Long-term Group Fair Policies in Dynamical Systems},
author = {Miriam Rateike and Isabel Valera and Patrick Forré},
url = {https://doi.org/10.1145/3630106.3658538},
doi = {10.1145/3630106.3658538},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {The 2024 ACM Conference on Fairness, Accountability, and Transparency,
FAccT 2024, Rio de Janeiro, Brazil, June 3-6, 2024},
pages = {20–50},
publisher = {ACM},
abstract = {Neglecting the effect that decisions have on individuals (and thus, on the underlying data distribution) when designing algorithmic decision-making policies may increase inequalities and unfairness in the long term—even if fairness considerations were taken into account in the policy design process. In this paper, we propose a novel framework for studying long-term group fairness in dynamical systems, in which current decisions may affect an individual’s features in the next step, and thus, future decisions. Specifically, our framework allows us to identify a time-independent policy that converges, if deployed, to the targeted fair stationary state of the system in the long-term, independently of the initial data distribution. We model the system dynamics with a time-homogeneous Markov chain and optimize the policy leveraging the Markov Chain Convergence Theorem to ensure unique convergence. Our framework enables the utilization of historical temporal data to tackle challenges associated with delayed feedback when learning long-term fair policies in practice. Importantly, our framework shows that interventions on the data distribution (e.g., subsidies) can be used to achieve policy learning that is both short- and long-term fair. We provide examples of different targeted fair states of the system, encompassing a range of long-term goals for society and policymakers. In semi-synthetic simulations based on real-world datasets, we show how our approach facilitates identifying effective interventions for long-term fairness.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Majumdar, Ayan; Valera, Isabel
CARMA: A practical framework to generate recommendations for causal algorithmic recourse at scale Proceedings Article
In: The 2024 ACM Conference on Fairness, Accountability, and Transparency, FAccT 2024, Rio de Janeiro, Brazil, June 3-6, 2024, pp. 1745–1762, ACM, 2024.
@inproceedings{DBLP:conf/fat/MajumdarV24,
title = {CARMA: A practical framework to generate recommendations for causal algorithmic recourse at scale},
author = {Ayan Majumdar and Isabel Valera},
url = {https://doi.org/10.1145/3630106.3659003},
doi = {10.1145/3630106.3659003},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
booktitle = {The 2024 ACM Conference on Fairness, Accountability, and Transparency,
FAccT 2024, Rio de Janeiro, Brazil, June 3-6, 2024},
pages = {1745–1762},
publisher = {ACM},
abstract = {Algorithms are increasingly used to automate large-scale decision-making processes, e.g., online platforms that make instant decisions in lending, hiring, and education. When such automated systems yield unfavorable decisions, it is imperative to allow for recourse by accompanying the instantaneous negative decisions with recommendations that can help affected individuals to overturn them. However, the practical challenges of providing algorithmic recourse in large-scale settings are not negligible: giving recourse recommendations that are actionable requires not only causal knowledge of the relationships between applicant features but also solving a complex combinatorial optimization problem for each rejected applicant. In this work, we introduce CARMA, a novel framework to generate causal recourse recommendations at scale. For practical settings with limited causal information, CARMA leverages pre-trained state-of-the-art causal generative models to find recourse recommendations. More importantly, CARMA addresses the scalability of finding these recommendations by casting the complex recourse optimization problem as a prediction task. By training a novel neural-network-based framework, CARMA efficiently solves the prediction task without requiring supervision for optimal recourse actions. Our extensive evaluations show that post-training, running inference on CARMA reliably amortizes causal recourse, generating optimal and instantaneous recommendations. CARMA exhibits flexibility, as its optimization is versatile with respect to the algorithmic decision-making and pre-trained causal generative models, provided their differentiability is ensured. Furthermore, we showcase CARMA in a case study, illustrating its ability to tailor causal recourse recommendations by readily incorporating population-level feature preferences based on factors such as difficulty or time needed.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Wu, Nan; Valera, Isabel; Sinz, Fabian H.; Ecker, Alexander S.; Euler, Thomas; Qiu, Yongrong
Probabilistic neural transfer function estimation with Bayesian system identification Journal Article
In: PLoS Comput. Biol., vol. 20, no. 7, pp. 1012354, 2024.
@article{DBLP:journals/ploscb/WuVSEEQ24,
title = {Probabilistic neural transfer function estimation with Bayesian system identification},
author = {Nan Wu and Isabel Valera and Fabian H. Sinz and Alexander S. Ecker and Thomas Euler and Yongrong Qiu},
url = {https://doi.org/10.1371/journal.pcbi.1012354},
doi = {10.1371/JOURNAL.PCBI.1012354},
year = {2024},
date = {2024-01-01},
urldate = {2024-01-01},
journal = {PLoS Comput. Biol.},
volume = {20},
number = {7},
pages = {1012354},
abstract = {Neural population responses in sensory systems are driven by external physical stimuli. This stimulus-response relationship is typically characterized by receptive fields, which have been estimated by neural system identification approaches. Such models usually require a large amount of training data, yet, the recording time for animal experiments is limited, giving rise to epistemic uncertainty for the learned neural transfer functions. While deep neural network models have demonstrated excellent power on neural prediction, they usually do not provide the uncertainty of the resulting neural representations and derived statistics, such as most exciting inputs (MEIs), from in silico experiments. Here, we present a Bayesian system identification approach to predict neural responses to visual stimuli, and explore whether explicitly modeling network weight variability can be beneficial for identifying neural response properties. To this end, we use variational inference to estimate the posterior distribution of each model weight given the training data. Tests with different neural datasets demonstrate that this method can achieve higher or comparable performance on neural prediction, with a much higher data efficiency compared to Monte Carlo dropout methods and traditional models using point estimates of the model parameters. At the same time, our variational method provides us with an effectively infinite ensemble, avoiding the idiosyncrasy of any single model, to generate MEIs. This allows us to estimate the uncertainty of stimulus-response function, which we have found to be negatively correlated with the predictive performance at model level and may serve to evaluate models. Furthermore, our approach enables us to identify response properties with credible intervals and to determine whether the inferred features are meaningful by performing statistical tests on MEIs. Finally, in silico experiments show that our model generates stimuli driving neuronal activity significantly better than traditional models in the limited-data regime.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2023
Javaloy, Adrián; Sánchez-Mart'ın, Pablo; Valera, Isabel
Causal normalizing flows: from theory to practice Journal Article
In: CoRR, vol. abs/2306.05415, 2023.
@article{DBLP:journals/corr/abs-2306-05415,
title = {Causal normalizing flows: from theory to practice},
author = {Adrián Javaloy and Pablo Sánchez-Mart'ın and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2306.05415},
doi = {10.48550/arXiv.2306.05415},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {CoRR},
volume = {abs/2306.05415},
abstract = {In this work, we deepen on the use of normalizing flows for causal reasoning. Specifically, we first leverage recent results on non-linear ICA to show that causal models are identifiable from observational data given a causal ordering, and thus can be recovered using autoregressive normalizing flows (NFs). Second, we analyze different design and learning choices for causal normalizing flows to capture the underlying causal data-generating process. Third, we describe how to implement the do-operator in causal NFs, and thus, how to answer interventional and counterfactual questions. Finally, in our experiments, we validate our design and training choices through a comprehensive ablation study; compare causal NFs to other approaches for approximating causal models; and empirically demonstrate that causal NFs can be used to address real-world problems—where the presence of mixed discrete-continuous data and partial knowledge on the causal graph is the norm. The code for this work can be found at https://github.com/psanch21/causal-flows.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Koyuncu, Batuhan; Sánchez-Mart'ın, Pablo; Peis, Ignacio; Olmos, Pablo M.; Valera, Isabel
Variational Mixture of HyperGenerators for Learning Distributions Over Functions Journal Article
In: CoRR, vol. abs/2302.06223, 2023.
@article{DBLP:journals/corr/abs-2302-06223,
title = {Variational Mixture of HyperGenerators for Learning Distributions Over Functions},
author = {Batuhan Koyuncu and Pablo Sánchez-Mart'ın and Ignacio Peis and Pablo M. Olmos and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2302.06223},
doi = {10.48550/arXiv.2302.06223},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {CoRR},
volume = {abs/2302.06223},
abstract = {Recent approaches build on implicit neural representations (INRs) to propose generative models over function spaces. However, they are computationally costly when dealing with inference tasks, such as missing data imputation, or directly cannot tackle them. In this work, we propose a novel deep generative model, named VAMoH. VAMoH combines the capabilities of modeling continuous functions using INRs and the inference capabilities of Variational Autoencoders (VAEs). In addition, VAMoH relies on a normalizing flow to define the prior, and a mixture of hypernetworks to parametrize the data log-likelihood. This gives VAMoH a high expressive capability and interpretability. Through experiments on a diverse range of data types, such as images, voxels, and climate data, we show that VAMoH can effectively learn rich distributions over continuous functions. Furthermore, it can perform inference-related tasks, such as conditional super-resolution generation and in-painting, as well or better than previous approaches, while being less computationally demanding.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Javaloy, Adrián; Sánchez-Mart'ın, Pablo; Levi, Amit; Valera, Isabel
Learnable Graph Convolutional Attention Networks Proceedings Article
In: The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023, OpenReview.net, 2023.
@inproceedings{DBLP:conf/iclr/JavaloySLV23,
title = {Learnable Graph Convolutional Attention Networks},
author = {Adrián Javaloy and Pablo Sánchez-Mart'ın and Amit Levi and Isabel Valera},
url = {https://openreview.net/pdf?id=WsUMeHPo-2},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {The Eleventh International Conference on Learning Representations,
ICLR 2023, Kigali, Rwanda, May 1-5, 2023},
publisher = {OpenReview.net},
abstract = {Existing Graph Neural Networks (GNNs) compute the message exchange between nodes by either aggregating uniformly (convolving) the features of all the neighbor- ing nodes, or by applying a non-uniform score (attending) to the features. Recent works have shown the strengths and weaknesses of the resulting GNN architectures, respectively, GCNs and GATs. In this work, we aim at exploiting the strengths of both approaches to their full extent. To this end, we first introduce the graph convolutional attention layer (CAT), which relies on convolutions to compute the attention scores. Unfortunately, as in the case of GCNs and GATs, we show that there exists no clear winner between the three—neither theoretically nor in practice—as their performance directly depends on the nature of the data (i.e., of the graph and features). This result brings us to the main contribution of our work, the learnable graph convolutional attention network (L-CAT): a GNN architecture that automatically interpolates between GCN, GAT and CAT in each layer, by adding only two scalar parameters. Our results demonstrate that L-CAT is able to efficiently combine different GNN layers along the network, outperforming competing methods in a wide range of datasets, and resulting in a more robust model that reduces the need of cross-validating.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Karimi, Amir-Hossein; Barthe, Gilles; Schölkopf, Bernhard; Valera, Isabel
A Survey of Algorithmic Recourse: Contrastive Explanations and Consequential Recommendations Journal Article
In: ACM Comput. Surv., vol. 55, no. 5, pp. 95:1–95:29, 2023.
@article{DBLP:journals/csur/KarimiBSV23,
title = {A Survey of Algorithmic Recourse: Contrastive Explanations and Consequential Recommendations},
author = {Amir-Hossein Karimi and Gilles Barthe and Bernhard Schölkopf and Isabel Valera},
url = {https://doi.org/10.1145/3527848},
doi = {10.1145/3527848},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
journal = {ACM Comput. Surv.},
volume = {55},
number = {5},
pages = {95:1–95:29},
abstract = {Machine learning is increasingly used to inform decision making in sensitive situations where decisions have consequential effects on individuals’ lives. In these settings, in addition to requiring models to be accurate and robust, socially relevant values such as fairness, privacy, accountability, and explainability play an important role in the adoption and impact of said technologies. In this work, we focus on algorithmic recourse, which is concerned with providing explanations and recommendations to individuals who are unfavorably treated by automated decision-making systems. We first perform an extensive literature review, and align the efforts of many authors by presenting unified definitions, formulations, and solutions to recourse. Then, we provide an overview of the prospective research directions toward which the community may engage, challenging existing assumptions and making explicit connections to other ethical challenges such as security, privacy, and fairness.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Langer, Markus; Valera, Isabel
Leveraging Actionable Explanations to Improve People's Reactions to AI-Based Decisions Proceedings Article
In: Steffen, Bernhard (Ed.): Bridging the Gap Between AI and Reality - First International Conference, AISoLA 2023, Crete, Greece, October 23-28, 2023, Selected Papers, pp. 293–306, Springer, 2023.
@inproceedings{DBLP:conf/vecos/LangerV23,
title = {Leveraging Actionable Explanations to Improve People's Reactions to AI-Based Decisions},
author = {Markus Langer and Isabel Valera},
editor = {Bernhard Steffen},
url = {https://doi.org/10.1007/978-3-031-73741-1_18},
doi = {10.1007/978-3-031-73741-1_18},
year = {2023},
date = {2023-01-01},
urldate = {2023-01-01},
booktitle = {Bridging the Gap Between AI and Reality - First International Conference,
AISoLA 2023, Crete, Greece, October 23-28, 2023, Selected Papers},
volume = {14129},
pages = {293–306},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
abstract = {This paper explores the role of explanations in mitigating negative reactions among people affected by AI-based decisions. While existing research focuses primarily on user perspectives, this study addresses the unique needs of people affected by AI-based decisions. Drawing on justice theory and the algorithmic recourse literature, we propose that actionability is a primary need of people affected by AI-based decisions. Thus, we expected that more actionable explanations – that is, explanations that guide people on how to address negative outcomes – would elicit more favorable reactions than feature relevance explanations or no explanations. In a within-participants experiment, participants (N = 138) imagined being loan applicants and were informed that their loan application had been rejected by AI-based systems at five different banks. Participants received either no explanation, feature relevance explanations, or actionable explanations for this decision. Additionally, we varied the degree of actionability of the features mentioned in the explanations to explore whether features that are more actionable (i.e., reduce the amount of loan) lead to additional positive effects on people’s reactions compared to less actionable features (i.e., increase your income). We found that providing any explanation led to more favorable reactions, and that actionable explanations led to more favorable reactions than feature relevance explanations. However, focusing on the supposedly more actionable feature led to comparably more negative effects possibly due to our specific context of application. We discuss the crucial role that perceived actionability may play for people affected by AI-based decisions as well as the nuanced effects that focusing on different features in explanations may have.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2022
Javaloy, Adrián; Meghdadi, Maryam; Valera, Isabel
Mitigating Modality Collapse in Multimodal VAEs via Impartial Optimization Journal Article
In: CoRR, vol. abs/2206.04496, 2022.
@article{DBLP:journals/corr/abs-2206-04496,
title = {Mitigating Modality Collapse in Multimodal VAEs via Impartial Optimization},
author = {Adrián Javaloy and Maryam Meghdadi and Isabel Valera},
url = {https://doi.org/10.48550/arXiv.2206.04496},
doi = {10.48550/arXiv.2206.04496},
year = {2022},
date = {2022-01-01},
urldate = {2022-01-01},
journal = {CoRR},
volume = {abs/2206.04496},
abstract = {A number of variational autoencoders (VAEs) have recently emerged with the aim of modeling multimodal data, e.g., to jointly model images and their corresponding captions. Still, multimodal VAEs tend to focus solely on a subset of the modalities, e.g., by fitting the image while neglecting the caption. We refer to this limitation as modality collapse. In this work, we argue that this effect is a consequence of conflicting gradients during multimodal VAE training. We show how to detect the sub-graphs in the computational graphs where gradients conflict (impartiality blocks), as well as how to leverage existing gradient-conflict solutions from multitask learning to mitigate modality collapse. That is, to ensure impartial optimization across modalities. We apply our training framework to several multimodal VAE models, losses and datasets from the literature, and empirically show that our framework significantly improves the reconstruction performance, conditional generation, and coherence of the latent space across modalities.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Kügelgen, Julius; Karimi, Amir-Hossein; Bhatt, Umang; Valera, Isabel; Weller, Adrian; Schölkopf, Bernhard
On the Fairness of Causal Algorithmic Recourse Proceedings Article
In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, pp. 9584–9594, AAAI Press, 2022.
@inproceedings{DBLP:conf/aaai/KugelgenKBVWS22,
title = {On the Fairness of Causal Algorithmic Recourse},
author = {Julius Kügelgen and Amir-Hossein Karimi and Umang Bhatt and Isabel Valera and Adrian Weller and Bernhard Schölkopf},
url = {https://ojs.aaai.org/index.php/AAAI/article/view/21192},
year = {2022},
date = {2022-01-01},
urldate = {2022-01-01},
booktitle = {Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI
2022, Thirty-Fourth Conference on Innovative Applications of Artificial
Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances
in Artificial Intelligence, EAAI 2022 Virtual Event, February 22
- March 1, 2022},
pages = {9584–9594},
publisher = {AAAI Press},
abstract = {Algorithmic fairness is typically studied from the perspective of predictions. Instead, here we investigate fairness from the perspective of recourse actions suggested to individuals to remedy an unfavourable classification. We propose two new fair-ness criteria at the group and individual level, which—unlike prior work on equalising the average group-wise distance from the decision boundary—explicitly account for causal relationships between features, thereby capturing downstream effects of recourse actions performed in the physical world. We explore how our criteria relate to others, such as counterfactual fairness, and show that fairness of recourse is complementary to fairness of prediction. We study theoretically and empirically how to enforce fair causal recourse by altering the classifier and perform a case study on the Adult dataset. Finally, we discuss whether fairness violations in the data generating process revealed by our criteria may be better addressed by societal interventions as opposed to constraints on the classifier.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Javaloy, Adrián; Valera, Isabel
RotoGrad: Gradient Homogenization in Multitask Learning Proceedings Article
In: The Tenth International Conference on Learning Representations, ICLR 2022, Virtual Event, April 25-29, 2022, OpenReview.net, 2022.
@inproceedings{DBLP:conf/iclr/JavaloyV22,
title = {RotoGrad: Gradient Homogenization in Multitask Learning},
author = {Adrián Javaloy and Isabel Valera},
url = {https://openreview.net/forum?id=T8wHz4rnuGL},
year = {2022},
date = {2022-01-01},
urldate = {2022-01-01},
booktitle = {The Tenth International Conference on Learning Representations, ICLR
2022, Virtual Event, April 25-29, 2022},
publisher = {OpenReview.net},
abstract = {Multitask learning is being increasingly adopted in applications domains like computer vision and reinforcement learning. However, optimally exploiting its advantages remains a major challenge due to the effect of negative transfer. Previous works have tracked down this issue to the disparities in gradient magnitudes and directions across tasks, when optimizing the shared network parameters. While recent work has acknowledged that negative transfer is a two-fold problem, existing approaches fall short as they only focus on either homogenizing the gradient magnitude across tasks; or greedily change the gradient directions, overlooking future conflicts. In this work, we introduce RotoGrad, an algorithm that tackles negative transfer as a whole: it jointly homogenizes gradient magnitudes and directions, while ensuring training convergence. We show that RotoGrad outperforms competing methods in complex problems, including multi-label classification in CelebA and computer vision tasks in the NYUv2 dataset. A Pytorch implementation can be found in https://github.com/adrianjav/rotograd.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Camps-Valls, Gustau; Ruiz, Francisco J. R.; Valera, Isabel (Ed.)
PMLR, vol. 151, 2022.
@proceedings{DBLP:conf/aistats/2022,
title = {International Conference on Artificial Intelligence and Statistics,
AISTATS 2022, 28-30 March 2022, Virtual Event},
editor = {Gustau Camps-Valls and Francisco J. R. Ruiz and Isabel Valera},
url = {http://proceedings.mlr.press/v151/},
year = {2022},
date = {2022-01-01},
volume = {151},
publisher = {PMLR},
series = {Proceedings of Machine Learning Research},
keywords = {},
pubstate = {published},
tppubtype = {proceedings}
}
Sánchez-Mart'ın, Pablo; Rateike, Miriam; Valera, Isabel
VACA: Designing Variational Graph Autoencoders for Causal Queries Proceedings Article
In: Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI 2022, Thirty-Fourth Conference on Innovative Applications of Artificial Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances in Artificial Intelligence, EAAI 2022 Virtual Event, February 22 - March 1, 2022, pp. 8159–8168, AAAI Press, 2022.
@inproceedings{DBLP:conf/aaai/Sanchez-MartinR22b,
title = {VACA: Designing Variational Graph Autoencoders for Causal Queries},
author = {Pablo Sánchez-Mart'ın and Miriam Rateike and Isabel Valera},
url = {https://ojs.aaai.org/index.php/AAAI/article/view/20789},
year = {2022},
date = {2022-01-01},
urldate = {2022-01-01},
booktitle = {Thirty-Sixth AAAI Conference on Artificial Intelligence, AAAI
2022, Thirty-Fourth Conference on Innovative Applications of Artificial
Intelligence, IAAI 2022, The Twelveth Symposium on Educational Advances
in Artificial Intelligence, EAAI 2022 Virtual Event, February 22
- March 1, 2022},
pages = {8159–8168},
publisher = {AAAI Press},
abstract = {In this paper, we introduce VACA, a novel class of variational graph autoencoders for causal inference in the absence of hidden confounders, when only observational data and the causal graph are available. Without making any parametric assumptions, VACA mimics the necessary properties of a Structural Causal Model (SCM) to provide a flexible and practical framework for approximating interventions (do-operator) and abduction-action-prediction steps. As a result, and as shown by our empirical results, VACA accurately approximates the interventional and counterfactual distributions on diverse SCMs. Finally, we apply VACA to evaluate counterfactual fairness in fair classification problems, as well as to learn fair classifiers without compromising performance.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Rateike, Miriam; Majumdar, Ayan; Mineeva, Olga; Gummadi, Krishna P.; Valera, Isabel
Don't Throw it Away! The Utility of Unlabeled Data in Fair Decision Making Proceedings Article
In: FAccT '22: 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, June 21 - 24, 2022, pp. 1421–1433, ACM, 2022.
@inproceedings{DBLP:conf/fat/RateikeMMGV22,
title = {Don't Throw it Away! The Utility of Unlabeled Data in Fair Decision Making},
author = {Miriam Rateike and Ayan Majumdar and Olga Mineeva and Krishna P. Gummadi and Isabel Valera},
url = {https://doi.org/10.1145/3531146.3533199},
doi = {10.1145/3531146.3533199},
year = {2022},
date = {2022-01-01},
urldate = {2022-01-01},
booktitle = {FAccT '22: 2022 ACM Conference on Fairness, Accountability, and
Transparency, Seoul, Republic of Korea, June 21 - 24, 2022},
pages = {1421–1433},
publisher = {ACM},
abstract = {unbiased, i.e., equally distributed across socially salient groups. In many practical settings, the ground-truth cannot be directly observed, and instead, we have to rely on a biased proxy measure of the ground-truth, i.e., biased labels, in the data. In addition, data is often selectively labeled, i.e., even the biased labels are only observed for a small fraction of the data that received a positive decision. To overcome label and selection biases, recent work proposes to learn stochastic, exploring decision policies via i) online training of new policies at each time-step and ii) enforcing fairness as a constraint on performance. However, the existing approach uses only labeled data, disregarding a large amount of unlabeled data, and thereby suffers from high instability and variance in the learned decision policies at different times. In this paper, we propose a novel method based on a variational autoencoder for practical fair decision-making. Our method learns an unbiased data representation leveraging both labeled and unlabeled data and uses the representations to learn a policy in an online process. Using synthetic data, we empirically validate that our method converges to the optimal (fair) policy according to the ground-truth with low variance. In real-world experiments, we further show that our training approach not only offers a more stable learning process but also yields policies with higher fairness as well as utility than previous approaches.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
2021
Javaloy, Adrián; Meghdadi, Maryam; Valera, Isabel
Boosting heterogeneous VAEs via multi-objective optimization Workshop
2021.
@workshop{nokey,
title = {Boosting heterogeneous VAEs via multi-objective optimization},
author = {Adrián Javaloy and Maryam Meghdadi and Isabel Valera},
url = {http://adrian.javaloy.com/publication/mo-vae/},
year = {2021},
date = {2021-12-01},
urldate = {2021-12-01},
abstract = {A number of variational autoencoders (VAEs) have recently emerged with the aim of modeling multimodal data, e.g., to jointly model images and their corresponding captions. Still, multimodal VAEs tend to focus solely on a subset of the modalities, e.g., by fitting the image while neglecting the caption. We refer to this limitation as modality collapse. In this work, we argue that this effect is a consequence of conflicting gradients during multimodal VAE training. We show how to detect the sub-graphs in the computational graphs where gradients conflict (impartiality blocks), as well as how to leverage existing gradient-conflict solutions from multitask learning to mitigate modality collapse. That is, to ensure impartial optimization across modalities. We apply our training framework to several multimodal VAE models, losses and datasets from the literature, and empirically show that our framework significantly improves the reconstruction performance, conditional generation, and coherence of the latent space across modalities.},
keywords = {},
pubstate = {published},
tppubtype = {workshop}
}
Mohammadi, Kiarash; Karimi, Amir-Hossein; Barthe, Gilles; Valera, Isabel
Scaling Guarantees for Nearest Counterfactual Explanations Proceedings Article
In: Fourcade, Marion; Kuipers, Benjamin; Lazar, Seth; Mulligan, Deirdre K. (Ed.): AIES '21: AAAI/ACM Conference on AI, Ethics, and Society, Virtual Event, USA, May 19-21, 2021, pp. 177–187, ACM, 2021.
@inproceedings{DBLP:conf/aies/MohammadiKBV21,
title = {Scaling Guarantees for Nearest Counterfactual Explanations},
author = {Kiarash Mohammadi and Amir-Hossein Karimi and Gilles Barthe and Isabel Valera},
editor = {Marion Fourcade and Benjamin Kuipers and Seth Lazar and Deirdre K. Mulligan},
url = {https://doi.org/10.1145/3461702.3462514},
doi = {10.1145/3461702.3462514},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
booktitle = {AIES '21: AAAI/ACM Conference on AI, Ethics, and Society, Virtual
Event, USA, May 19-21, 2021},
pages = {177--187},
publisher = {ACM},
abstract = {Counterfactual explanations (CFE) are being widely used to explain algorithmic decisions, especially in consequential decision-making contexts (e.g., loan approval or pretrial bail). In this context, CFEs aim to provide individuals affected by an algorithmic decision with the most similar individual (i.e., nearest individual) with a different outcome. However, while an increasing number of works propose algorithms to compute CFEs, such approaches either lack in optimality of distance (i.e., they do not return the nearest individual) and perfect coverage (i.e., they do not provide a CFE for all individuals); or they do not scale to complex models such as neural networks. In this work, we provide a framework based on Mixed-Integer Programming (MIP) to compute nearest counterfactual explanations for the outcomes of neural networks, with both provable guarantees and runtimes comparable to gradient-based approaches. Our experiments on the Adult, COMPAS, and Credit datasets show that, in contrast with previous methods, our approach allows for efficiently computing diverse CFEs with both distance guarantees and perfect coverage.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Karimi, Amir-Hossein; Schölkopf, Bernhard; Valera, Isabel
Algorithmic Recourse: from Counterfactual Explanations to Interventions Proceedings Article
In: Elish, Madeleine Clare; Isaac, William; Zemel, Richard S. (Ed.): FAccT '21: 2021 ACM Conference on Fairness, Accountability, and Transparency, Virtual Event / Toronto, Canada, March 3-10, 2021, pp. 353–362, ACM, 2021.
@inproceedings{DBLP:conf/fat/KarimiSV21,
title = {Algorithmic Recourse: from Counterfactual Explanations to Interventions},
author = {Amir-Hossein Karimi and Bernhard Schölkopf and Isabel Valera},
editor = {Madeleine Clare Elish and William Isaac and Richard S. Zemel},
url = {https://doi.org/10.1145/3442188.3445899},
doi = {10.1145/3442188.3445899},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
booktitle = {FAccT '21: 2021 ACM Conference on Fairness, Accountability, and
Transparency, Virtual Event / Toronto, Canada, March 3-10, 2021},
pages = {353--362},
publisher = {ACM},
abstract = {As machine learning is increasingly used to inform consequential decision-making (e.g., pre-trial bail and loan approval), it becomes important to explain how the system arrived at its decision, and also suggest actions to achieve a favorable decision. Counterfactual explanations -"how the world would have (had) to be different for a desirable outcome to occur"- aim to satisfy these criteria. Existing works have primarily focused on designing algorithms to obtain counterfactual explanations for a wide range of settings. However, it has largely been overlooked that ultimately, one of the main objectives is to allow people to act rather than just understand. In layman's terms, counterfactual explanations inform an individual where they need to get to, but not how to get there. In this work, we rely on causal reasoning to caution against the use of counterfactual explanations as a recommendable set of actions for recourse. Instead, we propose a shift of paradigm from recourse via nearest counterfactual explanations to recourse through minimal interventions, shifting the focus from explanations to interventions.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Schöffer, Jakob; Kuehl, Niklas; Valera, Isabel
A Ranking Approach to Fair Classification Proceedings Article
In: COMPASS '21: ACM SIGCAS Conference on Computing and Sustainable Societies, Virtual Event, Australia, 28 June 2021 - 2 July 2021, pp. 115–125, ACM, 2021.
@inproceedings{DBLP:conf/dev/SchofferKV21,
title = {A Ranking Approach to Fair Classification},
author = {Jakob Schöffer and Niklas Kuehl and Isabel Valera},
url = {https://doi.org/10.1145/3460112.3471950},
doi = {10.1145/3460112.3471950},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
booktitle = {COMPASS '21: ACM SIGCAS Conference on Computing and Sustainable
Societies, Virtual Event, Australia, 28 June 2021 - 2 July 2021},
pages = {115--125},
publisher = {ACM},
abstract = {Algorithmic decision systems are increasingly used in areas such as hiring, school admission, or loan approval. Typically, these systems rely on labeled data for training a classification model. However, in many scenarios, ground-truth labels are unavailable, and instead we have only access to imperfect labels as the result of (potentially biased) human-made decisions. Despite being imperfect, historical decisions often contain some useful information on the unobserved true labels. In this paper, we focus on scenarios where only imperfect labels are available and propose a new fair ranking-based decision system based on monotonic relationships between legitimate features and the outcome. Our approach is both intuitive and easy to implement, and thus particularly suitable for adoption in real-world settings. More in detail, we introduce a distance-based decision criterion, which incorporates useful information from historical decisions and accounts for unwanted correlation between protected and legitimate features. Through extensive experiments on synthetic and real-world data, we show that our method is fair in the sense that a) it assigns the desirable outcome to the most qualified individuals, and b) it removes the effect of stereotypes in decision-making, thereby outperforming traditional classification algorithms. Additionally, we are able to show theoretically that our method is consistent with a prominent concept of individual fairness which states that “similar individuals should be treated similarly.”},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Hutter, Frank; Kersting, Kristian; Lijffijt, Jefrey; Valera, Isabel (Ed.)
Springer, vol. 12457, 2021, ISBN: 978-3-030-67657-5.
@proceedings{DBLP:conf/pkdd/2020-1,
title = {Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2020, Ghent, Belgium, September 14-18, 2020, Proceedings, Part I},
editor = {Frank Hutter and Kristian Kersting and Jefrey Lijffijt and Isabel Valera},
url = {https://doi.org/10.1007/978-3-030-67658-2},
doi = {10.1007/978-3-030-67658-2},
isbn = {978-3-030-67657-5},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
volume = {12457},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
abstract = {Pattern Mining; clustering; privacy and fairness; (social) network analysis and computational social science; dimensionality reduction and autoencoders; domain adaptation; sketching, sampling, and binary projections; graphical models and causality; (spatio-) temporal data and recurrent neural networks; collaborative filtering and matrix completion.},
keywords = {},
pubstate = {published},
tppubtype = {proceedings}
}
Hutter, Frank; Kersting, Kristian; Lijffijt, Jefrey; Valera, Isabel (Ed.)
Springer, vol. 12458, 2021, ISBN: 978-3-030-67660-5.
@proceedings{DBLP:conf/pkdd/2020-2,
title = {Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2020, Ghent, Belgium, September 14-18, 2020, Proceedings, Part II},
editor = {Frank Hutter and Kristian Kersting and Jefrey Lijffijt and Isabel Valera},
url = {https://doi.org/10.1007/978-3-030-67661-2},
doi = {10.1007/978-3-030-67661-2},
isbn = {978-3-030-67660-5},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
volume = {12458},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
abstract = {Deep learning optimization and theory;active learning; adversarial learning; federated learning; Kernel methods and online learning; partial label learning; reinforcement learning; transfer and multi-task learning; Bayesian optimization and few-shot learning.},
keywords = {},
pubstate = {published},
tppubtype = {proceedings}
}
Hutter, Frank; Kersting, Kristian; Lijffijt, Jefrey; Valera, Isabel (Ed.)
Springer, vol. 12459, 2021, ISBN: 978-3-030-67663-6.
@proceedings{DBLP:conf/pkdd/2020-3,
title = {Machine Learning and Knowledge Discovery in Databases - European Conference, ECML PKDD 2020, Ghent, Belgium, September 14-18, 2020, Proceedings, Part III},
editor = {Frank Hutter and Kristian Kersting and Jefrey Lijffijt and Isabel Valera},
url = {https://doi.org/10.1007/978-3-030-67664-3},
doi = {10.1007/978-3-030-67664-3},
isbn = {978-3-030-67663-6},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
volume = {12459},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
abstract = {Combinatorial optimization; large-scale optimization and differential privacy; boosting and ensemble methods; Bayesian methods; architecture of neural networks; graph neural networks; Gaussian processes; computer vision and image processing; natural language processing; bioinformatics.},
keywords = {},
pubstate = {published},
tppubtype = {proceedings}
}
Javaloy, Adrián; Valera, Isabel
Rotograd: Dynamic Gradient Homogenization for Multi-Task Learning Journal Article
In: CoRR, vol. abs/2103.02631, 2021.
@article{DBLP:journals/corr/abs-2103-02631,
title = {Rotograd: Dynamic Gradient Homogenization for Multi-Task Learning},
author = {Adrián Javaloy and Isabel Valera},
url = {https://arxiv.org/abs/2103.02631},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
journal = {CoRR},
volume = {abs/2103.02631},
abstract = {Multitask learning is being increasingly adopted in applications domains like computer vision and reinforcement learning. However, optimally exploiting its advantages remains a major challenge due to the effect of negative transfer. Previous works have tracked down this issue to the disparities in gradient magnitudes and directions across tasks, when optimizing the shared network parameters. While recent work has acknowledged that negative transfer is a two-fold problem, existing approaches fall short as they only focus on either homogenizing the gradient magnitude across tasks; or greedily change the gradient directions, overlooking future conflicts. In this work, we introduce RotoGrad, an algorithm that tackles negative transfer as a whole: it jointly homogenizes gradient magnitudes and directions, while ensuring training convergence. We show that RotoGrad outperforms competing methods in complex problems, including multi-label classification in CelebA and computer vision tasks in the NYUv2 dataset. A Pytorch implementation can be found in this https://github.com/adrianjav/rotograd},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Sánchez-Mart'ın, Pablo; Rateike, Miriam; Valera, Isabel
VACA: Design of Variational Graph Autoencoders for Interventional and Counterfactual Queries Journal Article
In: CoRR, vol. abs/2110.14690, 2021.
@article{DBLP:journals/corr/abs-2110-14690,
title = {VACA: Design of Variational Graph Autoencoders for Interventional and Counterfactual Queries},
author = {Pablo Sánchez-Mart'ın and Miriam Rateike and Isabel Valera},
url = {https://arxiv.org/abs/2110.14690},
year = {2021},
date = {2021-01-01},
urldate = {2021-01-01},
journal = {CoRR},
volume = {abs/2110.14690},
abstract = {In this paper, we introduce VACA, a novel class of variational graph autoencoders for causal inference in the absence of hidden confounders, when only observational data and the causal graph are available. Without making any parametric assumptions, VACA mimics the necessary properties of a Structural Causal Model (SCM) to provide a flexible and practical framework for approximating interventions (do-operator) and abduction-action-prediction steps. As a result, and as shown by our empirical results, VACA accurately approximates the interventional and counterfactual distributions on diverse SCMs. Finally, we apply VACA to evaluate counterfactual fairness in fair classification problems, as well as to learn fair classifiers without compromising performance.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
2020
Javaloy, Adrián; Valera, Isabel
Lipschitz standardization for robust multivariate learning Journal Article
In: CoRR, vol. abs/2002.11369, 2020.
@article{DBLP:journals/corr/abs-2002-11369,
title = {Lipschitz standardization for robust multivariate learning},
author = {Adrián Javaloy and Isabel Valera},
url = {https://arxiv.org/abs/2002.11369},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
journal = {CoRR},
volume = {abs/2002.11369},
abstract = {Probabilistic learning is increasingly being tackled as an optimization problem, with gradient-based approaches as predominant methods. When modelling multivariate likelihoods, a usual but undesirable outcome is that the learned model fits only a subset of the observed variables, overlooking the rest. In this work, we study this problem through the lens of multitask learning (MTL), where similar effects have been broadly studied. While MTL solutions do not directly apply in the probabilistic setting (as they cannot handle the likelihood constraints) we show that similar ideas may be leveraged during data preprocessing. First, we show that data standardization often helps under common continuous likelihoods, but it is not enough in the general case, specially under mixed continuous and discrete likelihood models. In order for balance multivariate learning, we then propose a novel data preprocessing, Lipschitz standardization, which balances the local Lipschitz smoothness across variables. Our experiments on real-world datasets show that Lipschitz standardization leads to more accurate multivariate models than the ones learned using existing data preprocessing techniques. The models and datasets employed in the experiments can be found in this URL https://github.com/adrianjav/lipschitz-standardization},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Karimi, Amir-Hossein; Barthe, Gilles; Balle, Borja; Valera, Isabel
Model-Agnostic Counterfactual Explanations for Consequential Decisions Proceedings Article
In: Chiappa, Silvia; Calandra, Roberto (Ed.): The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy], pp. 895–905, PMLR, 2020.
@inproceedings{DBLP:conf/aistats/KarimiBBV20,
title = {Model-Agnostic Counterfactual Explanations for Consequential Decisions},
author = {Amir-Hossein Karimi and Gilles Barthe and Borja Balle and Isabel Valera},
editor = {Silvia Chiappa and Roberto Calandra},
url = {http://proceedings.mlr.press/v108/karimi20a.html},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
booktitle = {The 23rd International Conference on Artificial Intelligence and Statistics,
AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy]},
volume = {108},
pages = {895--905},
publisher = {PMLR},
series = {Proceedings of Machine Learning Research},
abstract = {Predictive models are being increasingly used to support consequential decision making at the individual level in contexts such as pretrial bail and loan approval. As a result, there is increasing social and legal pressure to provide explanations that help the affected individuals not only to understand why a prediction was output, but also how to act to obtain a desired outcome. To this end, several works have proposed optimization-based methods to generate nearest counterfactual explanations. However, these methods are often restricted to a particular subset of models (e.g., decision trees or linear models) and differentiable distance functions. In contrast, we build on standard theory and tools from formal verification and propose a novel algorithm that solves a sequence of satisfiability problems, where both the distance function (objective) and predictive model (constraints) are represented as logic formulae. As shown by our experiments on real-world data, our algorithm is: i) model-agnostic ({non-}linear, {non-}differentiable, {non-}convex); ii) data-type-agnostic (heterogeneous features); iii) distance-agnostic (l0, l1, l8, and combinations thereof); iv) able to generate plausible and diverse counterfactuals for any sample (i.e., 100% coverage); and v) at provably optimal distances.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Karimi, Amir-Hossein; Kügelgen, Bodo Julius; Schölkopf, Bernhard; Valera, Isabel
Algorithmic recourse under imperfect causal knowledge: a probabilistic approach Proceedings Article
In: Larochelle, Hugo; Ranzato, Marc'Aurelio; Hadsell, Raia; Balcan, Maria-Florina; Lin, Hsuan-Tien (Ed.): Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems 2020, NeurIPS 2020, December 6-12, 2020, virtual, 2020.
@inproceedings{DBLP:conf/nips/KarimiKSV20,
title = {Algorithmic recourse under imperfect causal knowledge: a probabilistic approach},
author = {Amir-Hossein Karimi and Bodo Julius Kügelgen and Bernhard Schölkopf and Isabel Valera},
editor = {Hugo Larochelle and Marc'Aurelio Ranzato and Raia Hadsell and Maria-Florina Balcan and Hsuan-Tien Lin},
url = {https://proceedings.neurips.cc/paper/2020/hash/02a3c7fb3f489288ae6942498498db20-Abstract.html},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
booktitle = {Advances in Neural Information Processing Systems 33: Annual Conference
on Neural Information Processing Systems 2020, NeurIPS 2020, December
6-12, 2020, virtual},
abstract = {Recent work has discussed the limitations of counterfactual explanations to recommend actions for algorithmic recourse, and argued for the need of taking causal relationships between features into consideration. Unfortunately, in practice, the true underlying structural causal model is generally unknown. In this work, we first show that it is impossible to guarantee recourse without access to the true structural equations. To address this limitation, we propose two probabilistic approaches to select optimal actions that achieve recourse with high probability given limited causal knowledge (e.g., only the causal graph). The first captures uncertainty over structural equations under additive Gaussian noise, and uses Bayesian model averaging to estimate the counterfactual distribution. The second removes any assumptions on the structural equations by instead computing the average effect of recourse actions on individuals similar to the person who seeks recourse, leading to a novel subpopulation-based interventional notion of recourse. We then derive a gradient-based procedure for selecting optimal recourse actions, and empirically show that the proposed approaches lead to more reliable recommendations under imperfect causal knowledge than non-probabilistic baselines.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Karimi, Amir-Hossein; Barthe, Gilles; Schölkopf, Bernhard; Valera, Isabel
A survey of algorithmic recourse: definitions, formulations, solutions, and prospects Journal Article
In: CoRR, vol. abs/2010.04050, 2020.
@article{DBLP:journals/corr/abs-2010-04050,
title = {A survey of algorithmic recourse: definitions, formulations, solutions, and prospects},
author = {Amir-Hossein Karimi and Gilles Barthe and Bernhard Schölkopf and Isabel Valera},
url = {https://arxiv.org/abs/2010.04050},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
journal = {CoRR},
volume = {abs/2010.04050},
abstract = {Machine learning is increasingly used to inform decision-making in sensitive situations where decisions have consequential effects on individuals' lives. In these settings, in addition to requiring models to be accurate and robust, socially relevant values such as fairness, privacy, accountability, and explainability play an important role for the adoption and impact of said technologies. In this work, we focus on algorithmic recourse, which is concerned with providing explanations and recommendations to individuals who are unfavourably treated by automated decision-making systems. We first perform an extensive literature review, and align the efforts of many authors by presenting unified definitions, formulations, and solutions to recourse. Then, we provide an overview of the prospective research directions towards which the community may engage, challenging existing assumptions and making explicit connections to other ethical challenges such as security, privacy, and fairness.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Valera, Isabel; Pradier, Melanie F.; Lomeli, Maria; Ghahramani, Zoubin
General Latent Feature Models for Heterogeneous Datasets Journal Article
In: J. Mach. Learn. Res., vol. 21, pp. 100:1–100:49, 2020.
@article{DBLP:journals/jmlr/ValeraPLG20,
title = {General Latent Feature Models for Heterogeneous Datasets},
author = {Isabel Valera and Melanie F. Pradier and Maria Lomeli and Zoubin Ghahramani},
url = {http://jmlr.org/papers/v21/17-328.html},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
journal = {J. Mach. Learn. Res.},
volume = {21},
pages = {100:1--100:49},
abstract = {Latent variable models allow capturing the hidden structure underlying the data. In particular, feature allocation models represent each observation by a linear combination of latent variables. These models are often used to make predictions either for new observations or for missing information in the original data, as well as to perform exploratory data analysis. Although there is an extensive literature on latent feature allocation models for homogeneous datasets, where all the attributes that describe each object are of the same (continuous or discrete) type, there is no general framework for practical latent feature modeling for heterogeneous datasets. In this paper, we introduce a general Bayesian nonparametric latent feature allocation model suitable for heterogeneous datasets, where the attributes describing each object can be arbitrary combinations of real-valued, positive real-valued, categorical, ordinal and count variables. The proposed model presents several important properties. First, it is suitable for heterogeneous data while keeping the properties of conjugate models, which enables us to develop an inference algorithm that presents linear complexity with respect to the number of objects and attributes per MCMC iteration. Second, the Bayesian nonparametric component allows us to place a prior distribution on the number of features required to capture the latent structure in the data. Third, the latent features in the model are binary-valued, which facilitates the interpretability of the obtained latent features in exploratory data analysis. Finally, a software package, called GLFM toolbox, is made publicly available for other researchers to use and extend. It is available at https://ivaleram.github.io/GLFM/. We show the flexibility of the proposed model by solving both prediction and data analysis tasks on several real-world datasets.},
keywords = {},
pubstate = {published},
tppubtype = {article}
}
Kilbertus, Niki; Rodriguez, Manuel Gomez; Schölkopf, Bernhard; Muandet, Krikamol; Valera, Isabel
Fair Decisions Despite Imperfect Predictions Proceedings Article
In: Chiappa, Silvia; Calandra, Roberto (Ed.): The 23rd International Conference on Artificial Intelligence and Statistics, AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy], pp. 277–287, PMLR, 2020.
@inproceedings{DBLP:conf/aistats/KilbertusRSMV20,
title = {Fair Decisions Despite Imperfect Predictions},
author = {Niki Kilbertus and Manuel Gomez Rodriguez and Bernhard Schölkopf and Krikamol Muandet and Isabel Valera},
editor = {Silvia Chiappa and Roberto Calandra},
url = {http://proceedings.mlr.press/v108/kilbertus20a.html},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
booktitle = {The 23rd International Conference on Artificial Intelligence and Statistics,
AISTATS 2020, 26-28 August 2020, Online [Palermo, Sicily, Italy]},
volume = {108},
pages = {277--287},
publisher = {PMLR},
series = {Proceedings of Machine Learning Research},
abstract = {Consequential decisions are increasingly informed by sophisticated data-driven predictive models. However, consistently learning accurate predictive models requires access to ground truth labels. Unfortunately, in practice, labels may only exist conditional on certain decisions—if a loan is denied, there is not even an option for the individual to pay back the loan. In this paper, we show that, in this selective labels setting, learning to predict is suboptimal in terms of both fairness and utility. To avoid this undesirable behavior, we propose to directly learn stochastic decision policies that maximize utility under fairness constraints. In the context of fair machine learning, our results suggest the need for a paradigm shift from "learning to predict" to "learning to decide". Experiments on synthetic and real-world data illustrate the favorable properties of learning to decide, in terms of both utility and fairness.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
Karimi, Amir-Hossein; Kügelgen, Julius; Schölkopf, Bernhard; Valera, Isabel
Towards Causal Algorithmic Recourse Proceedings Article
In: Holzinger, Andreas; Goebel, Randy; Fong, Ruth; Moon, Taesup; Müller, Klaus-Robert; Samek, Wojciech (Ed.): xxAI - Beyond Explainable AI - International Workshop, Held in Conjunction with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended Papers, pp. 139–166, Springer, 2020.
@inproceedings{DBLP:conf/icml/KarimiKSV20,
title = {Towards Causal Algorithmic Recourse},
author = {Amir-Hossein Karimi and Julius Kügelgen and Bernhard Schölkopf and Isabel Valera},
editor = {Andreas Holzinger and Randy Goebel and Ruth Fong and Taesup Moon and Klaus-Robert Müller and Wojciech Samek},
url = {https://doi.org/10.1007/978-3-031-04083-2_8},
doi = {10.1007/978-3-031-04083-2_8},
year = {2020},
date = {2020-01-01},
urldate = {2020-01-01},
booktitle = {xxAI - Beyond Explainable AI - International Workshop, Held in Conjunction
with ICML 2020, July 18, 2020, Vienna, Austria, Revised and Extended
Papers},
volume = {13200},
pages = {139–166},
publisher = {Springer},
series = {Lecture Notes in Computer Science},
abstract = {Algorithmic recourse is concerned with aiding individuals who are unfavorably treated by automated decision-making systems to overcome their hardship, by offering recommendations that would result in a more favorable prediction when acted upon. Such recourse actions are typically obtained through solving an optimization problem that minimizes changes to the individual’s feature vector, subject to various plausibility, diversity, and sparsity constraints. Whereas previous works offer solutions to the optimization problem in a variety of settings, they critically overlook real-world considerations pertaining to the environment in which recourse actions are performed.
The present work emphasizes that changes to a subset of the individual’s attributes may have consequential down-stream effects on other attributes, thus making recourse a fundamcausal problem. Here, we model such considerations using the framework of structural causal models, and highlight pitfalls of not considering causal relations through examples and theory. Such insights allow us to reformulate the optimization problem to directly optimize for minimally-costly recourse over a space of feasible actions (in the form of causal interventions) rather than optimizing for minimally-distant “counterfactual explanations”. We offer both the optimization formulations and solutions to deterministic and probabilistic recourse, on an individualized and sub-population level, overcoming the steep assumptive requirements of offering recourse in general settings. Finally, using synthetic and semi-synthetic experiments based on the German Credit dataset, we demonstrate how such methods can be applied in practice under minimal causal assumptions.},
keywords = {},
pubstate = {published},
tppubtype = {inproceedings}
}
The present work emphasizes that changes to a subset of the individual’s attributes may have consequential down-stream effects on other attributes, thus making recourse a fundamcausal problem. Here, we model such considerations using the framework of structural causal models, and highlight pitfalls of not considering causal relations through examples and theory. Such insights allow us to reformulate the optimization problem to directly optimize for minimally-costly recourse over a space of feasible actions (in the form of causal interventions) rather than optimizing for minimally-distant “counterfactual explanations”. We offer both the optimization formulations and solutions to deterministic and probabilistic recourse, on an individualized and sub-population level, overcoming the steep assumptive requirements of offering recourse in general settings. Finally, using synthetic and semi-synthetic experiments based on the German Credit dataset, we demonstrate how such methods can be applied in practice under minimal causal assumptions.
