Skip to main navigation Skip to search Skip to main content

On the convergence of a family of robust losses for stochastic gradient descent

  • Bo Han
  • , Ivor W. Tsang*
  • , Ling Chen
  • *Corresponding author for this work

Research output: Chapter in book/report/conference proceedingConference proceedingpeer-review

19 Citations (Scopus)

Abstract

The convergence of Stochastic Gradient Descent (SGD) using convex loss functions has been widely studied. However, vanilla SGD methods using convex losses cannot perform well with noisy labels, which adversely affect the update of the primal variable in SGD methods. Unfortunately, noisy labels are ubiquitous in real world applications such as crowdsourcing. To handle noisy labels, in this paper, we present a family of robust losses for SGD methods. By employing our robust losses, SGD methods successfully reduce negative effects caused by noisy labels on each update of the primal variable. We not only reveal the convergence rate of SGD methods using robust losses, but also provide the robustness analysis on two representative robust losses. Comprehensive experimental results on six real-world datasets show that SGD methods using robust losses are obviously more robust than other baseline methods in most situations with fast convergence.

Original languageEnglish
Title of host publicationMachine Learning and Knowledge Discovery in Databases
Subtitle of host publicationEuropean Conference, ECML PKDD 2016, Riva del Garda, Italy, September 19-23, 2016, Proceedings, Part I
EditorsJilles Giuseppe, Niels Landwehr, Giuseppe Manco, Paolo Frasconi
PublisherSpringer Cham
Pages665-680
Number of pages16
ISBN (Electronic)9783319461281
ISBN (Print)9783319461274
DOIs
Publication statusPublished - 3 Sept 2016
Event15th European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases - Riva del Garda, Italy
Duration: 19 Sept 201623 Sept 2016
https://ecmlpkdd.org/2016/ (Conference website)
https://ecmlpkdd-storage.s3.eu-central-1.amazonaws.com/former-websites/2016/downloads/program_booklet.pdf (Conference booklet)
https://link.springer.com/book/10.1007/978-3-319-46128-1#accessibility-statement (Conference proceeding)

Publication series

NameLecture Notes in Computer Science
Volume9851
ISSN (Print)0302-9743
ISSN (Electronic)1611-3349
Name Lecture Notes in Artificial Intelligence
ISSN (Print)2945-9133
ISSN (Electronic)2945-9141
NameECML PKDD: Joint European Conference on Machine Learning and Knowledge Discovery in Databases

Conference

Conference15th European Conference on Machine Learning and Principles and Practice of Knowledge Discovery in Databases
Abbreviated titleECML PKDD 2016
Country/TerritoryItaly
CityRiva del Garda
Period19/09/1623/09/16
Internet address

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 9 - Industry, Innovation, and Infrastructure
    SDG 9 Industry, Innovation, and Infrastructure

Fingerprint

Dive into the research topics of 'On the convergence of a family of robust losses for stochastic gradient descent'. Together they form a unique fingerprint.

Cite this