Skip to content

Commit 0515d9d

Browse files
committed
Add reference
1 parent 3085a2f commit 0515d9d

File tree

1 file changed

+16
-1
lines changed

1 file changed

+16
-1
lines changed

docs/references/bib.bib

Lines changed: 16 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -2648,7 +2648,7 @@ @article{sedgewick:1986a
26482648
@article{williams:1964a,
26492649
abstract = {},
26502650
address = {New York, NY, USA},
2651-
author = {John William Joseph Williams},
2651+
author = {John William and Joseph Williams},
26522652
doi = {10.1145/512274.512284},
26532653
journal = {Communications of the ACM},
26542654
keywords = {algorithms, sorting, algorithm, heap, heapsort},
@@ -2705,3 +2705,18 @@ @techreport{rosenblatt:1957a
27052705
url = {},
27062706
year = {1957},
27072707
}
2708+
2709+
@article{shalevshwartz:2011a,
2710+
abstract = {We describe and analyze a simple and effective stochastic sub-gradient descent algorithm for solving the optimization problem cast by Support Vector Machines (SVM). We prove that the number of iterations required to obtain a solution of accuracy ${\epsilon}$ is ${\tilde{O}(1 / \epsilon)}$, where each iteration operates on a single training example. In contrast, previous analyses of stochastic gradient descent methods for SVMs require ${\Omega(1 / \epsilon^2)}$ iterations. As in previously devised SVM solvers, the number of iterations also scales linearly with $1/\lambda$, where $\lambda$ is the regularization parameter of SVM. For a linear kernel, the total run-time of our method is ${\tilde{O}(d/(\lambda \epsilon))}$, where $d$ is a bound on the number of non-zero features in each example. Since the run-time does not depend directly on the size of the training set, the resulting algorithm is especially suited for learning from large datasets. Our approach also extends to non-linear kernels while working solely on the primal objective function, though in this case the runtime does depend linearly on the training set size. Our algorithm is particularly well suited for large text classification problems, where we demonstrate an order-of-magnitude speedup over previous SVM learning methods.},
2711+
author = {Shai Shalev-Shwartz and Yoram Singer and Nathan Srebro and Andrew Cotter },
2712+
doi = {10.1007/s10107-010-0420-4},
2713+
journal = {Mathematical Programming},
2714+
keywords = {machine learning, svm, ml, classification, regression, loss functions},
2715+
month = {03},
2716+
number = {1},
2717+
pages = {3--30},
2718+
title = {{Pegasos: primal estimated sub-gradient solver for SVM}},
2719+
url = {https://doi.org/10.1007/s10107-010-0420-4},
2720+
volume = {127},
2721+
year = {2011},
2722+
}

0 commit comments

Comments
 (0)