Sciweavers

23 search results - page 1 / 5
» Application specific loss minimization using gradient boosti...
Sort
View
ICASSP
2011
IEEE
12 years 8 months ago
Application specific loss minimization using gradient boosting
Gradient boosting is a flexible machine learning technique that produces accurate predictions by combining many weak learners. In this work, we investigate its use in two applica...
Bin Zhang, Abhinav Sethy, Tara N. Sainath, Bhuvana...
KDD
2008
ACM
120views Data Mining» more  KDD 2008»
14 years 5 months ago
Multi-class cost-sensitive boosting with p-norm loss functions
We propose a family of novel cost-sensitive boosting methods for multi-class classification by applying the theory of gradient boosting to p-norm based cost functionals. We establ...
Aurelie C. Lozano, Naoki Abe
NIPS
2007
13 years 6 months ago
A General Boosting Method and its Application to Learning Ranking Functions for Web Search
We present a general boosting method extending functional gradient boosting to optimize complex loss functions that are encountered in many machine learning problems. Our approach...
Zhaohui Zheng, Hongyuan Zha, Tong Zhang, Olivier C...
SDM
2012
SIAM
252views Data Mining» more  SDM 2012»
11 years 7 months ago
Learning from Heterogeneous Sources via Gradient Boosting Consensus
Multiple data sources containing different types of features may be available for a given task. For instance, users’ profiles can be used to build recommendation systems. In a...
Xiaoxiao Shi, Jean-François Paiement, David...
ICCV
2007
IEEE
14 years 6 months ago
Gradient Feature Selection for Online Boosting
Boosting has been widely applied in computer vision, especially after Viola and Jones's seminal work [23]. The marriage of rectangular features and integral-imageenabled fast...
Ting Yu, Xiaoming Liu 0002