Skip to main content
Download PDF
- Main
Bad Optimizations Make Good Learning
- Chen, Ziqi
- Advisor(s): Helmbold, David P.
Abstract
This thesis reports on experiments aimed at explaining why machine
learning algorithms using the greedy stochastic gradient descent
(SGD) algorithm sometimes generalize better than algorithms using other optimization techniques. We propose two hypothesis, namely the "canyon effect" and the ``classification insensitivity'', and illustrate them with two data sources. On these data sources, SGD generalizes more
accurately than SVMperf, which performs more intensive optimization, over a wide variety of choices of the regularization parameters. Finally, we report on some similar, but predictably less dramatic, effects on natural data.
Main Content
For improved accessibility of PDF content, download the file to your device.
Enter the password to open this PDF file:
File name:
-
File size:
-
Title:
-
Author:
-
Subject:
-
Keywords:
-
Creation Date:
-
Modification Date:
-
Creator:
-
PDF Producer:
-
PDF Version:
-
Page Count:
-
Page Size:
-
Fast Web View:
-
Preparing document for printing…
0%