Active Regression via Linear-Sample Sparsification
Xue Chen Eric Price
UT Austin
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 1 / 18
Active Regression via Linear-Sample Sparsification Xue Chen Eric - - PowerPoint PPT Presentation
Active Regression via Linear-Sample Sparsification Xue Chen Eric Price UT Austin Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 1 / 18 Agnostic learning Xue Chen, Eric Price (UT Austin) Active Regression
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 1 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 2 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd.
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
◮ f (x) = αTφ(x) for some φ : X → Rd. ◮ Example: univariate degree d − 1 polynomials.
◮ Will get
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 3 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
◮ Mean zero noise:
D ≤ ǫf ∗ − y2 D
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
◮ Mean zero noise:
D ≤ ǫf ∗ − y2 D
◮ Generic noise:
D ≤ (1 + ǫ)f − y2 D
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
◮ Mean zero noise:
D ≤ ǫf ∗ − y2 D
◮ Generic noise:
D ≤ (1 + ǫ)f − y2 D
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
◮ Mean zero noise:
D ≤ ǫf ∗ − y2 D
◮ Generic noise:
D ≤ (1 + ǫ)f − y2 D
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 4 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). ◮ Know D (which just defines f −
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). ◮ Know D (which just defines f −
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). ◮ Know D (which just defines f −
◮ Receive x1, . . . , xm ∼ D Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). ◮ Know D (which just defines f −
◮ Receive x1, . . . , xm ∼ D ◮ Pick S ⊂ [m] of size s Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). ◮ Know D (which just defines f −
◮ Receive x1, . . . , xm ∼ D ◮ Pick S ⊂ [m] of size s ◮ See yi for i ∈ S. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
◮ Can pick xi of our choice, see yi ∼ (Y |X = xi). ◮ Know D (which just defines f −
◮ Receive x1, . . . , xm ∼ D ◮ Pick S ⊂ [m] of size s ◮ See yi for i ∈ S.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 5 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
◮ This gives O(κ log d) sample complexity by Matrix Chernoff. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 6 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 7 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 7 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 7 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 7 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification
◮ Not with independent sampling (coupon collector). Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification
◮ Not with independent sampling (coupon collector). ◮ Analogous to Batson-Spielman-Srivastava linear size sparsification. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification
◮ Not with independent sampling (coupon collector). ◮ Analogous to Batson-Spielman-Srivastava linear size sparsification. ◮ Yes – using Lee-Sun sparsification. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification
◮ Not with independent sampling (coupon collector). ◮ Analogous to Batson-Spielman-Srivastava linear size sparsification. ◮ Yes – using Lee-Sun sparsification.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
◮ Essentially the same as leverage score sampling. ◮ Also analogous to Spielman-Srivastava graph sparsification
◮ Not with independent sampling (coupon collector). ◮ Analogous to Batson-Spielman-Srivastava linear size sparsification. ◮ Yes – using Lee-Sun sparsification.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 8 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
◮ Label every point =
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
◮ Label every point =
◮ Hence m = Θ(K log d + K
ǫ ) optimal.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
◮ Label every point =
◮ Hence m = Θ(K log d + K
ǫ ) optimal.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
◮ Label every point =
◮ Hence m = Θ(K log d + K
ǫ ) optimal.
◮ In this talk: mostly s = O(d log d) version. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
◮ Label every point =
◮ Hence m = Θ(K log d + K
ǫ ) optimal.
◮ In this talk: mostly s = O(d log d) version. ◮ Prior work: s = O((d log d)5/4) [Sabato-Munos ’14], Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
◮ Get x1, . . . , xm ∼ D. ◮ Pick S ⊆ [m] of size s. ◮ Learn yi for i ∈ S.
◮ m → ∞ =
◮ Hence s = Θ(d) optimal.
◮ Label every point =
◮ Hence m = Θ(K log d + K
ǫ ) optimal.
◮ In this talk: mostly s = O(d log d) version. ◮ Prior work: s = O((d log d)5/4) [Sabato-Munos ’14], s = O(d log d)
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 9 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 10 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 11 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 11 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 11 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 11 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 11 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 11 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class: Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class: ◮ Noise variance bounded for every sample: Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class:
x∼D f (x)2 ≈ s
◮ Noise variance bounded for every sample: Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class:
x∼D f (x)2 ≈ s
◮ Noise variance bounded for every sample:
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class:
x∼D f (x)2 ≈ s
◮ Noise variance bounded for every sample:
f ,x
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class:
x∼D f (x)2 ≈ s
◮ Noise variance bounded for every sample:
f ,x
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
◮ xi ∼ Di where Di depends on x1, . . . , xi−1. ◮ D1 = D′, D2 avoids points near x1, etc.
◮ Norms preserved for all functions in class:
x∼D f (x)2 ≈ s
◮ Noise variance bounded for every sample:
f ,x
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 12 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 13 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 13 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 13 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 13 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 13 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 14 / 18
◮ Unknown exactly what this is for Fourier-sparse signals. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 14 / 18
◮ Unknown exactly what this is for Fourier-sparse signals. ◮ d2 ≤ K d4 log3 d. [Chen-Kane-Price-Song ’16] Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 14 / 18
◮ Unknown exactly what this is for Fourier-sparse signals. ◮ d2 ≤ K d4 log3 d. [Chen-Kane-Price-Song ’16]
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 14 / 18
◮ Unknown exactly what this is for Fourier-sparse signals. ◮ d2 ≤ K d4 log3 d. [Chen-Kane-Price-Song ’16]
◮ d ≤ κ d log2 d. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 14 / 18
◮ Unknown exactly what this is for Fourier-sparse signals. ◮ d2 ≤ K d4 log3 d. [Chen-Kane-Price-Song ’16]
◮ d ≤ κ d log2 d.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 14 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 15 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 16 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 16 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 16 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 17 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 17 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 17 / 18
◮ Known net size is 2 ˜
O(d3).
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 17 / 18
◮ Known net size is 2 ˜
O(d3).
◮ Gives ˜
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 17 / 18
◮ Known net size is 2 ˜
O(d3).
◮ Gives ˜
◮ Gives ˜
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 17 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net? Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net?
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net?
◮ Logistic regression? Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net?
◮ Logistic regression?
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net?
◮ Logistic regression?
◮ Choose sample points sequentially. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net?
◮ Logistic regression?
◮ Choose sample points sequentially. ◮ Dynamically changing functions. Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
◮ O(K log d + K
ǫ ) unlabeled examples.
◮ O(d/ǫ) labeled examples
◮ Tight results via chaining and/or better net?
◮ Logistic regression?
◮ Choose sample points sequentially. ◮ Dynamically changing functions.
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 18 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 19 / 18
Xue Chen, Eric Price (UT Austin) Active Regression via Linear-Sample Sparsification 20 / 18