Sciweavers

Share
ICDM
2010
IEEE

Learning Preferences with Millions of Parameters by Enforcing Sparsity

8 years 3 months ago
Learning Preferences with Millions of Parameters by Enforcing Sparsity
We study the retrieval task that ranks a set of objects for a given query in the pairwise preference learning framework. Recently researchers found out that raw features (e.g. words for text retrieval) and their pairwise features which describe relationships between two raw features (e.g. word synonymy or polysemy) could greatly improve the retrieval precision. However, most existing methods can not scale up to problems with many raw features (e.g. English vocabulary), due to the prohibitive computational cost on learning and the memory requirement to store a quadratic number of parameters. In this paper, we propose to learn a sparse representation of the pairwise features under the preference learning framework using the L1 regularization. Based on stochastic gradient descent, an online algorithm is devised to enforce the sparsity using a mini-batch shrinkage strategy. On multiple benchmark datasets, we show that our method achieves better performance with fast convergence, and takes ...
Xi Chen, Bing Bai, Yanjun Qi, Qihang Lin, Jaime G.
Added 12 Feb 2011
Updated 12 Feb 2011
Type Journal
Year 2010
Where ICDM
Authors Xi Chen, Bing Bai, Yanjun Qi, Qihang Lin, Jaime G. Carbonell
Comments (0)
books