This item is only available for download by members of the University of Illinois community. Students, faculty, and staff at the U of I may log in with your NetID and password to view the item. If you are trying to access an Illinois-restricted dissertation or thesis, you can request a copy through your library's Inter-Library Loan office or purchase a copy directly from ProQuest.
Permalink
https://hdl.handle.net/2142/47015
Description
Title
Advances in Sparse Classification
Author(s)
Bharadwaj, Sujeeth
Contributor(s)
Hasegawa-Johnson, Mark
Issue Date
2009-12
Keyword(s)
compressed sensing
speech recognition
nonparametric speech recognition
sparse classification
sparse projection
Abstract
A recent result in compressed sensing (CS) allows us to perform
non-parametric speech recognition that is robust to noise, and that requires
few training examples. By taking fixed length representations of training
samples and stacking them in a matrix, we form a frame, or an over-complete
basis. Gemmeke and Cranen have shown that sparse projections onto this
frame recover the correct transcription with 91% accuracy at -5 dB SNR. We
propose that the goal of speech recognition is not sparse projection onto
training tokens, but onto training types. Sparse projection onto types can be
achieved by building a frame for each word in the dictionary, and stacking the
frames to form a rank 3 tensor. Speech recognition is performed by convex
linear projection onto the tensor, with sparsity enforced only in the index that
specifies type. We derive a mixed L1/L2 relaxation that can be globally
optimized using Newton descent.
Use this login method if you
don't
have an
@illinois.edu
email address.
(Oops, I do have one)
IDEALS migrated to a new platform on June 23, 2022. If you created
your account prior to this date, you will have to reset your password
using the forgot-password link below.