Avrim Blum’s research while affiliated with Toyota Technological Institute at Chicago and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (123)


On Learning Verifiers for Chain-of-Thought Reasoning
  • Preprint

May 2025

Maria-Florina Balcan

·

Avrim Blum

·

Zhiyuan Li

·

Dravyansh Sharma

Chain-of-Thought reasoning has emerged as a powerful approach for solving complex mathematical and logical problems. However, it can often veer off track through incorrect or unsubstantiated inferences. Formal mathematical reasoning, which can be checked with a formal verifier, is one approach to addressing this issue. However, currently LLMs are simply not good enough to solve complex problems in a formal way, and even just formalizing an informal problem statement can be challenging. Motivated by this fact, in this work we consider the problem of learning reliable verifiers for natural language Chain-of-Thought reasoning. That is, given a problem statement and step-by-step solution in natural language, the aim of the verifier is to output [Yes] if the reasoning steps in the solution are all valid, and [No] otherwise. In this work we give a formal PAC-learning framework for studying this problem. We propose and analyze several natural verification goals, at different levels of strength, in this framework. We provide sample complexity upper-bounds for learning verifiers satisfying these goals, as well as lower-bound and impossibility results for learning other natural verification objectives without additional assumptions.


Proofs as Explanations: Short Certificates for Reliable Predictions
  • Preprint
  • File available

April 2025

·

4 Reads

We consider a model for explainable AI in which an explanation for a prediction h(x)=y consists of a subset SS' of the training data (if it exists) such that all classifiers hHh' \in H that make at most b mistakes on SS' predict h(x)=yh'(x)=y. Such a set SS' serves as a proof that x indeed has label y under the assumption that (1) the target function hh^\star belongs to H, and (2) the set S contains at most b corrupted points. For example, if b=0 and H is the family of linear classifiers in Rd\mathbb{R}^d, and if x lies inside the convex hull of the positive data points in S (and hence every consistent linear classifier labels x as positive), then Carath\'eodory's theorem states that x lies inside the convex hull of d+1 of those points. So, a set SS' of size d+1 could be released as an explanation for a positive prediction, and would serve as a short proof of correctness of the prediction under the assumption of realizability. In this work, we consider this problem more generally, for general hypothesis classes H and general values b0b\geq 0. We define the notion of the robust hollow star number of H (which generalizes the standard hollow star number), and show that it precisely characterizes the worst-case size of the smallest certificate achievable, and analyze its size for natural classes. We also consider worst-case distributional bounds on certificate size, as well as distribution-dependent bounds that we show tightly control the sample size needed to get a certificate for any given test example. In particular, we define a notion of the certificate coefficient εx\varepsilon_x of an example x with respect to a data distribution D and target function hh^\star, and prove matching upper and lower bounds on sample size as a function of εx\varepsilon_x, b, and the VC dimension d of H.

Download

PAC Learning with Improvements

March 2025

·

2 Reads

Idan Attias

·

Avrim Blum

·

Keziah Naggita

·

[...]

·

Matthew Walter

One of the most basic lower bounds in machine learning is that in nearly any nontrivial setting, it takes at least\textit{at least} 1/ϵ1/\epsilon samples to learn to error ϵ\epsilon (and more, if the classifier being learned is complex). However, suppose that data points are agents who have the ability to improve by a small amount if doing so will allow them to receive a (desired) positive classification. In that case, we may actually be able to achieve zero\textit{zero} error by just being "close enough". For example, imagine a hiring test used to measure an agent's skill at some job such that for some threshold θ\theta, agents who score above θ\theta will be successful and those who score below θ\theta will not (i.e., learning a threshold on the line). Suppose also that by putting in effort, agents can improve their skill level by some small amount r. In that case, if we learn an approximation θ^\hat{\theta} of θ\theta such that θθ^θ+r\theta \leq \hat{\theta} \leq \theta + r and use it for hiring, we can actually achieve error zero, in the sense that (a) any agent classified as positive is truly qualified, and (b) any agent who truly is qualified can be classified as positive by putting in effort. Thus, the ability for agents to improve has the potential to allow for a goal one could not hope to achieve in standard models, namely zero error. In this paper, we explore this phenomenon more broadly, giving general results and examining under what conditions the ability of agents to improve can allow for a reduction in the sample complexity of learning, or alternatively, can make learning harder. We also examine both theoretically and empirically what kinds of improvement-aware algorithms can take into account agents who have the ability to improve to a limited extent when it is in their interest to do so.



Replicable Online Learning

November 2024

·

4 Reads

We investigate the concept of algorithmic replicability introduced by Impagliazzo et al. 2022, Ghazi et al. 2021, Ahn et al. 2024 in an online setting. In our model, the input sequence received by the online learner is generated from time-varying distributions chosen by an adversary (obliviously). Our objective is to design low-regret online algorithms that, with high probability, produce the exact same sequence of actions when run on two independently sampled input sequences generated as described above. We refer to such algorithms as adversarially replicable. Previous works (such as Esfandiari et al. 2022) explored replicability in the online setting under inputs generated independently from a fixed distribution; we term this notion as iid-replicability. Our model generalizes to capture both adversarial and iid input sequences, as well as their mixtures, which can be modeled by setting certain distributions as point-masses. We demonstrate adversarially replicable online learning algorithms for online linear optimization and the experts problem that achieve sub-linear regret. Additionally, we propose a general framework for converting an online learner into an adversarially replicable one within our setting, bounding the new regret in terms of the original algorithm's regret. We also present a nearly optimal (in terms of regret) iid-replicable online algorithm for the experts problem, highlighting the distinction between the iid and adversarial notions of replicability. Finally, we establish lower bounds on the regret (in terms of the replicability parameter and time) that any replicable online algorithm must incur.


Regularized Robustly Reliable Learners and Instance Targeted Attacks

October 2024

·

3 Reads

Instance-targeted data poisoning attacks, where an adversary corrupts a training set to induce errors on specific test points, have raised significant concerns. Balcan et al (2022) proposed an approach to addressing this challenge by defining a notion of robustly-reliable learners that provide per-instance guarantees of correctness under well-defined assumptions, even in the presence of data poisoning attacks. They then give a generic optimal (but computationally inefficient) robustly reliable learner as well as a computationally efficient algorithm for the case of linear separators over log-concave distributions. In this work, we address two challenges left open by Balcan et al (2022). The first is that the definition of robustly-reliable learners in Balcan et al (2022) becomes vacuous for highly-flexible hypothesis classes: if there are two classifiers h_0, h_1 \in H both with zero error on the training set such that h_0(x) \neq h_1(x), then a robustly-reliable learner must abstain on x. We address this problem by defining a modified notion of regularized robustly-reliable learners that allows for nontrivial statements in this case. The second is that the generic algorithm of Balcan et al (2022) requires re-running an ERM oracle (essentially, retraining the classifier) on each test point x, which is generally impractical even if ERM can be implemented efficiently. To tackle this problem, we show that at least in certain interesting cases we can design algorithms that can produce their outputs in time sublinear in training time, by using techniques from dynamic algorithm design.


Distributional Adversarial Loss

June 2024

·

3 Reads

A major challenge in defending against adversarial attacks is the enormous space of possible attacks that even a simple adversary might perform. To address this, prior work has proposed a variety of defenses that effectively reduce the size of this space. These include randomized smoothing methods that add noise to the input to take away some of the adversary's impact. Another approach is input discretization which limits the adversary's possible number of actions. Motivated by these two approaches, we introduce a new notion of adversarial loss which we call distributional adversarial loss, to unify these two forms of effectively weakening an adversary. In this notion, we assume for each original example, the allowed adversarial perturbation set is a family of distributions (e.g., induced by a smoothing procedure), and the adversarial loss over each example is the maximum loss over all the associated distributions. The goal is to minimize the overall adversarial loss. We show generalization guarantees for our notion of adversarial loss in terms of the VC-dimension of the hypothesis class and the size of the set of allowed adversarial distributions associated with each input. We also investigate the role of randomness in achieving robustness against adversarial attacks in the methods described above. We show a general derandomization technique that preserves the extent of a randomized classifier's robustness against adversarial attacks. We corroborate the procedure experimentally via derandomizing the Random Projection Filters framework of \cite{dong2023adversarial}. Our procedure also improves the robustness of the model against various adversarial attacks.


On the Vulnerability of Fairness Constrained Learning to Malicious Noise

July 2023

·

2 Reads

We consider the vulnerability of fairness-constrained learning to small amounts of malicious noise in the training data. Konstantinov and Lampert (2021) initiated the study of this question and presented negative results showing there exist data distributions where for several fairness constraints, any proper learner will exhibit high vulnerability when group sizes are imbalanced. Here, we present a more optimistic view, showing that if we allow randomized classifiers, then the landscape is much more nuanced. For example, for Demographic Parity we show we can incur only a Θ(α)\Theta(\alpha) loss in accuracy, where α\alpha is the malicious noise rate, matching the best possible even without fairness constraints. For Equal Opportunity, we show we can incur an O(α)O(\sqrt{\alpha}) loss, and give a matching Ω(α)\Omega(\sqrt{\alpha})lower bound. In contrast, Konstantinov and Lampert (2021) showed for proper learners the loss in accuracy for both notions is Ω(1)\Omega(1). The key technical novelty of our work is how randomization can bypass simple "tricks" an adversary can use to amplify his power. We also consider additional fairness notions including Equalized Odds and Calibration. For these fairness notions, the excess accuracy clusters into three natural regimes O(α)O(\alpha),O(α)O(\sqrt{\alpha}) and O(1). These results provide a more fine-grained view of the sensitivity of fairness-constrained learning to adversarial noise in training data.



Strategic Classification under Unknown Personalized Manipulation

May 2023

·

1 Read

We study the fundamental mistake bound and sample complexity in the strategic classification, where agents can strategically manipulate their feature vector up to an extent in order to be predicted as positive. For example, given a classifier determining college admission, student candidates may try to take easier classes to improve their GPA, retake SAT and change schools in an effort to fool the classifier. Ball manipulations are a widely studied class of manipulations in the literature, where agents can modify their feature vector within a bounded radius ball. Unlike most prior work, our work considers manipulations to be personalized, meaning that agents can have different levels of manipulation abilities (e.g., varying radii for ball manipulations), and unknown to the learner. We formalize the learning problem in an interaction model where the learner first deploys a classifier and the agent manipulates the feature vector within their manipulation set to game the deployed classifier. We investigate various scenarios in terms of the information available to the learner during the interaction, such as observing the original feature vector before or after deployment, observing the manipulated feature vector, or not seeing either the original or the manipulated feature vector. We begin by providing online mistake bounds and PAC sample complexity in these scenarios for ball manipulations. We also explore non-ball manipulations and show that, even in the simplest scenario where both the original and the manipulated feature vectors are revealed, the mistake bounds and sample complexity are lower bounded by Ω(H)\Omega(|\mathcal{H}|) when the target function belongs to a known class H\mathcal{H}.


Citations (59)


... In online strategic classification, a decision-maker makes decisions over a sequence of agents who may manipulate their features to receive favorable outcomes [Brückner and Scheffer, 2011, Hardt et al., 2016, Ahmadi et al., 2023. For example, in college admissions, when a decision-maker evaluates applicants, students may retake the SAT, switch schools, or enroll in easier classes to boost their GPAs in hopes of gaining admission. ...

Reference:

Should Decision-Makers Reveal Classifiers in Online Strategic Classification?
Fundamental Bounds on Online Strategic Classification
  • Citing Conference Paper
  • July 2023

... In NTMs, topics are parameterized as word distributions [15], [16], similar to x bow . Hence, we conjecture that topic representation is a decomposed variant of each x bow , and we can only fully observe the distribution of the decompositions as in Figure 1 with the same number of times we retrieve x bow . ...

Algorithms for Generalized Topic Modeling
  • Citing Article
  • April 2018

Proceedings of the AAAI Conference on Artificial Intelligence

... This paper is related to the algorithmic fairness literature, which studies the design and evaluation of algorithms aimed to mitigate bias and improve fairness in algorithmic decision-making (Dwork, Hardt, et al. 2012;Zemel et al. 2013;Hardt et al. 2016;Zafar, Valera, Gomez Rodriguez, et al. 2017;Zafar, Valera, Rodriguez, et al. 2017;Geyik et al. 2019;Blum et al. 2022). In this literature, two broad notions of fairness exist: individual fairness, which requires that similar individuals are treated similarly by the algorithm; and group fairness, which requires that some statistic of interest is on average equal across groups along the lines of protected attributes. 1 Within group fairness, different definitions of fairness exist, such as demographic (or statistical) parity, equal selection, equal false-positive rates, equal false-negative rates, equal odds, equal accuracy rates, and equal positive predictive values across groups (see Table 7 for precise definitions and Mitchell et al. (2021) for a review). ...

Multi Stage Screening: Enforcing Fairness and Maximizing Efficiency in a Pre-Existing Pipeline
  • Citing Conference Paper
  • June 2022

... Another line of work related to this paper is the study of the stochastic minimum vertex cover problem [8,13,14]. In this problem, we are given a graph G = (V, E) and an existence probability for each edge e ∈ E. Edges of G are realized (or exist) independently with these probabilities, forming the realized subgraph G. ...

Stochastic Vertex Cover with Few Queries
  • Citing Chapter
  • January 2022

... Strategic Learning was initially introduced by Hardt et al. [2016] and sparked a large area of research studying how agents respond to decision rules in learning systems [Braverman and Garg, 2020, Dong et al., 2018, Zhang et al., 2022, Lechner et al., 2023, Chen et al., 2020, Ahmadi et al., 2021, Sundaram et al., 2023 to only name a few. For a recent survey, please refer to [Podimata, 2025]. ...

The Strategic Perceptron
  • Citing Conference Paper
  • July 2021

... The Stackelberg security game [24] is one widely studied example. Other game-theoretic models include the hide-and-seek game [8], blotto games [4], auditting games [5] and catcher-evader games [19]. Most of these games study the optimal usage of security forces under different game structures. ...

From Battlefields to Elections: Winning Strategies of Blotto and Auditing Games
  • Citing Book
  • January 2018

... For example, in the context of kidney exchange, living donors are often incompatible with their intended recipients, and it becomes increasingly more common to coordinate and match among multiple donor-recipient pairs simultaneously. Blum et al. (2020) formulated a stochastic matching problem with random edge deletions to help reduce the number of pairwise compatibility tests while identifying almost as many compatible patient-donor pairs as exhaustive testing does. We refer to Roth et al. (2005) and papers thereafter (e.g., Ashlagi et al. 2012, Dickerson et al. 2012, Ding et al. 2018) for more discussion of applying maximum matchings to kidney exchange. ...

Ignorance Is Almost Bliss: Near-Optimal Stochastic Matching with Few Queries
  • Citing Article
  • January 2020

Operations Research

... Though the applied surveys are scientifically validated and widely distributed, the gathered data are self-reported and biased. Because of the biased data, ML may yield prediction models that are biased with inferior accuracy on real-world data [54,55]. Therefore, future research should focus on gathering data through more objective personality assessment methods. ...

Recovering from Biased Data: Can Fairness Constraints Improve Accuracy?
  • Citing Preprint
  • December 2019

... In a more challenging setting, where the reward functions and/or transition kernel are unknown, another line of recent research resorts to reinforcement learning methods for computing Stackelberg equilibria [29] [28]. Among these, [2] explores sample-efficient algorithms for finding Stackelberg equilibrium of large general-sum games where the leader and/or the follower action set is exponential in the natural representation of the problem. ...

Computing Stackelberg Equilibria of Large General-Sum Games
  • Citing Chapter
  • September 2019

Lecture Notes in Computer Science

... A key insight was developed in (Roberson 2006) where the unique equilibrium payoffs and the distributions of the deployments were characterized in some detail. Since then several contributions to the algorithms to compute the equilibrium have been developed in (Ahmadinejad et al. 2016), (Behnezhad et al. 2017), (Behnezhad et al. 2018a), (Behnezhad et al. 2018b), (Behnezhad et al. 2019), and (Behnezhad et al. 2023) among others. There are also some similarities of the Ballot Stuffing Game to the Auditing Games games (Behnezhad et al. 2019) and Blotto Game with Testing (Sonin 2024). ...

Optimal Strategies of Blotto Games: Beyond Convexity
  • Citing Conference Paper
  • June 2019