The Power of Localization for Efficiently Learning Linear Separators with Noise

Type: Article

Publication Date: 2017-01-09

Citations: 42

DOI: https://doi.org/10.1145/3006384

Abstract

We introduce a new approach for designing computationally efficient learning algorithms that are tolerant to noise, and we demonstrate its effectiveness by designing algorithms with improved noise tolerance guarantees for learning linear separators. We consider both the malicious noise model of Valiant [1985] and Kearns and Li [1988] and the adversarial label noise model of Kearns, Schapire, and Sellie [1994]. For malicious noise, where the adversary can corrupt both the label and the features, we provide a polynomial-time algorithm for learning linear separators in ℜ d under isotropic log-concave distributions that can tolerate a nearly information-theoretically optimal noise rate of η = Ω(ϵ), improving on the Ω (ϵ 3 /log 2 ( d/ϵ )) noise-tolerance of Klivans et al. [2009a]. In the case that the distribution is uniform over the unit ball, this improves on the Ω (ϵ/ d 1/4 ) noise-tolerance of Kalai et al. [2005] and the Ω (ϵ 2 /log(d/ϵ)) of Klivans et al. [2009a]. For the adversarial label noise model, where the distribution over the feature vectors is unchanged and the overall probability of a noisy label is constrained to be at most η, we also give a polynomial-time algorithm for learning linear separators in ℜ d under isotropic log-concave distributions that can handle a noise rate of η = Ω(ϵ). In the case of uniform distribution, this improves over the results of Kalai et al. [2005], which either required runtime super-exponential in 1/ϵ (ours is polynomial in 1/ϵ) or tolerated less noise. 1 Our algorithms are also efficient in the active learning setting, where learning algorithms only receive the classifications of examples when they ask for them. We show that, in this model, our algorithms achieve a label complexity whose dependence on the error parameter ϵ is polylogarithmic (and thus exponentially better than that of any passive algorithm). This provides the first polynomial-time active learning algorithm for learning linear separators in the presence of malicious noise or adversarial label noise. Our algorithms and analysis combine several ingredients including aggressive localization, minimization of a progressively rescaled hinge loss, and a novel localized and soft outlier removal procedure. We use localization techniques (previously used for obtaining better sample complexity results) to obtain better noise-tolerant polynomial-time algorithms.

Locations

  • Journal of the ACM - View
  • arXiv (Cornell University) - View - PDF
  • OPAL (Open@LaTrobe) (La Trobe University) - View - PDF

Similar Works

Action Title Year Authors
+ The Power of Localization for Efficiently Learning Linear Separators with Noise 2013 Pranjal Awasthi
Maria Florina Balcan
Philip M. Long
+ The Power of Localization for Efficiently Learning Linear Separators with Malicious Noise. 2013 Pranjal Awasthi
Maria-Florina Balcan
Philip M. Long
+ PDF Chat The power of localization for efficiently learning linear separators with noise 2014 Pranjal Awasthi
Maria Florina Balcan
Philip M. Long
+ Efficient Learning of Linear Separators under Bounded Noise 2015 Pranjal Awasthi
Maria-Florina Balcan
Nika Haghtalab
Ruth Urner
+ Active and passive learning of linear separators under log-concave distributions 2012 Maria Florina Balcan
Philip M. Long
+ On the Power of Localized Perceptron for Label-Optimal Learning of Halfspaces with Adversarial Noise 2020 Jie Shen
+ Robustly-reliable learners under poisoning attacks 2022 Maria-Florina Balcan
Avrim Blum
Steve Hanneke
Dravyansh Sharma
+ Statistical Active Learning Algorithms for Noise Tolerance and Differential Privacy 2013 Maria Florina Balcan
Vitaly Feldman
+ Learning without Interaction Requires Separation. 2018 Amit Daniely
Vitaly Feldman
+ PDF Chat Efficient PAC Learning of Halfspaces with Constant Malicious Noise Rate 2024 Jie Shen
Xiaoyu Li
+ Attribute-Efficient Learning of Halfspaces with Malicious Noise: Near-Optimal Label Complexity and Noise Tolerance 2020 Jie Shen
Chicheng Zhang
+ Learning Geometric Concepts with Nasty Noise 2017 Ilias Diakonikolas
Daniel M. Kane
Alistair Stewart
+ Learning Geometric Concepts with Nasty Noise 2017 Ilias Diakonikolas
Daniel M. Kane
Alistair Stewart
+ Statistical Active Learning Algorithms for Noise Tolerance and Differential Privacy 2013 Maria Florina Balcan
Vitaly Feldman
+ Sample-Optimal PAC Learning of Halfspaces with Malicious Noise 2021 Jie Shen
+ Noise-tolerant, Reliable Active Classification with Comparison Queries 2020 Max Hopkins
Daniel M. Kane
Shachar Lovett
Gaurav Mahajan
+ Efficient Testable Learning of Halfspaces with Adversarial Label Noise 2023 Ilias Diakonikolas
Daniel M. Kane
Vasilis Kontonis
Sihan Liu
Nikos Zarifis
+ PDF Chat Efficient Certificates of Anti-Concentration Beyond Gaussians 2024 Ainesh Bakshi
Pravesh K. Kothari
Goutham Rajendran
Madhur Tulsiani
Aravindan Vijayaraghavan
+ Classification Under Misspecification: Halfspaces, Generalized Linear Models, and Connections to Evolvability 2020 Sitan Chen
Frederic Koehler
Ankur Moitra
Morris Yau
+ Revisiting Perceptron: Efficient and Label-Optimal Learning of Halfspaces 2017 Songbai Yan
Chicheng Zhang