Proactive Learning with Multiple Class-Sensitive Labelers
Recommend Documents
across multiple domains (assuming uniform reliability across classes) may not give ..... Specifically, when there exists
Oct 30, 2014 - Multiple Class-Sensitive Labelers. Seungwhan (Shane) Moon, Jaime Carbonell. School of Computer Science, C
and Skene [11] proposed using EM to improve the quality of simple majority ...... [5] V. S. Sheng, F. Provost, and P. G. Ipeirotis, âGet another label? Improving ...
Sep 9, 2010 - the round robin repeated labeling strategy (see Section 5). (There is no ...... following the discussion in Clemen and Winkler (1990). For our ...
d-dimensional synchronous CLA with n cells is a structure. A = (Zd, Φ,A,N, F), ..... include channel assignment algorithms for cellular mobile net- works and a ...
representations, and a learning environment with integrated representations. ...... Conference on Artificial Intelligence in Education, San Antonio, Texas, IOS ...
International Conference on Machine Learning. Two main reasons lead us to ... Multi-view learning is a natural, yet non-standard new problem setting; in brevity: learning from .... invited but we will also issue an open call for papers. Potential ...
Oct 5, 2011 - induction, Web page classification, advertisement removal, and discourse ..... a high accuracy) between work and personal emails based solely.
Jul 31, 2015 - and time-to-trigger (TTT) based on multiple criteria, viz HO ping pong (HOPP), HO failure (HOF) and UE speeds. ... network condition, network load, and UE speeds. Monitoring the system after manually changing the HO ...
Jul 31, 2015 - In this study, we propose the self-optimization of HO parameters based on .... Both strands of research show that FLC is a powerful tool for auto-tuning parameters .... where Tsearch is the target cell search delay when the target cell
Both works above exploit the same kernel directly for all classes, yet no ... benchmark data, we report some interesting findings and ex- planations concerning ...
Create an SDI application that contains more than one form. • Include Font and ....
on the menu bar and then click Add Windows Form to open the. Add New Item ...
labels are associated with sets or bags of patterns rather than individual .... Red and blue circles depict one positive bag and one negative bag on the left-hand ...
Nov 5, 2013 - online algorithms for learning f while taking into account the output ..... The idea of our second algorithm, MONORMA, is to combine the online ...
modelling anticancer drug activity [Barretina et al., 2012, Jang et al., 2014,. CCLE Consortium and GDSC Consortium, 2015]. Here, we applied EN re- gression ...
The pseudo code of the MT-DT algorithm is presented in Algorithm 1. The .... notated non-responsive emails constitute a binary classification task, Responsive.
Multiple Instance Learning, Support Vector Machine, Feature Selection, Alternating ...... in house making use of the LIBSVM C++ package for training the 2-norm ...
Amelia Zafra and Eva Gibaja and Sebastián Ventura. Department of Computer Science and Numerical Analysis. University of Córdoba. Email: [email protected], ...
TR-97{8. Multiple Attribute Learning with. Canonical Correlation Analysis and. EM algorithm. Shotaro Akaho. 3. Satoru Hayamizu y. Osamu Hasegawa.
large volumes of diagnostic data, and these data were analyzed by machine learning .... analysis using machine learning techniques, possibly resulting in new ...
Coverage and capacity of cellular network .... In my opinion, wireless mesh network in one lecture is not sufficient. ... AT&T, Verizon, Sprint, etc. was new to me.
Proactive Supply Chain Performance Management with ... › publication › fulltext › Proactive-... › publication › fulltext › Proactive-...by N Stefanovic · 2014 · Cited by 56 · Related articlesintegrates supply chain process model, online analytical
International Journal on Integrating Technology in Education (IJITE) Vol.4,No.3 ... Technology-enhanced learning, Adaptive teaching, student feedback based ...
ble-blind outcome study in patients with type 2 diabetes managed with diet ... Abbreviations: IDF, International Diabetes Federation; MI, myocardial .... Center when 50 and 75% of the target ... National Principal Investigators and the .... and by El
Proactive Learning with Multiple Class-Sensitive Labelers
Proactive Learning with Multiple Class-Sensitive Labelers
Seungwhan (Shane) Moon, Jaime Carbonell School of Computer Science, Carnegie Mellon University DSAA 2014 Conference 10/30/2014
Proactive Learning with Multiple Class-Sensitive Labelers
Seungwhan (Shane) Moon, Jaime Carbonell School of Computer Science, Carnegie Mellon University DSAA 2014 Conference 10/30/2014
Unlabeled Data is Abundant
3
Unlabeled Data is Abundant •
Imagine building a Vehicle classifier
Scarcity of labeled data 4
Active Learning
5
Active Learning
6
Query Strategies •
Uncertainty Sampling
•
Query by Committee
•
Entropy Based Sampling
•
Density Weighted Methods
•
and more …
7
Uncertainty Sampling
Label 1 Label 2 Unlabeled
Current Decision Boundary 8
Uncertainty Sampling
= most uncertain
Label 1 Label 2 Unlabeled
Current Decision Boundary 9
Assumptions in Traditional Active Learning
•
Annotator(s) always give perfect answers (oracle)
•
There is no difference in cost for querying different annotators
10
Proactive Learning [Carbonell et. al] •
Relaxes the following assumptions: •
Only a single annotator gives labels
•
Annotators always give perfect answers
•
Annotators are insensitive to costs —> utility optimization under budget constraint
11
Proactive Learning [Carbonell et. al] Multiple annotators They have different labeling accuracy (expertise) incur different cost
12
Proactive Learning [Carbonell et. al] Key Component: Estimating Labeler Accuracy Probability of getting a right answer for an unlabeled instance x, and an expert k
Limitation in previous literature on proactive learning
Labeler accuracy is independent of label in multi-class problems 13
Proactive Learning with Multiple Domain Experts: Anology Motivation Diagnosis of a patient with unknown disease (uncertainty in data)
14
Proactive Learning with Multiple Domain Experts: Anology Motivation Diagnosis of a patient with unknown disease (uncertainty in data) Given multiple physicians with different specialization (multiple class-sensitive experts) If we know the patient has seemingly cancer symptoms (posterior class probability) And that oncologist treats cancer issues (estimated labeler accuracy given a specific class) Better delegate a task to its respective expert 15
Proactive Learning with Multiple Domain Experts Problem Formulation (Objective)
: : : Greedy Approximation
16
Proactive Learning with Multiple Domain Experts Utility Criteria for Greedy Approximation
Jointly optimize for an instance and expert pair which - has high information value V(X) (instance) - has high probability of getting the right answer (both) - has low cost of annotation (expert) 17
Expert Estimation Estimating Expertise of Labeling Sources
class posterior probability of label for sample x being c
over set of categories
18
the estimated probability of expert k answering for label c
Expert Estimation Estimating Expertise of Labeling Sources
Per-class Reduced Estimation
19
Density Based Sampling for Multiclassification Tasks
Label 1 Label 2 Label 3 Unlabeled
Current Decision Boundary 20
Density Based Sampling for Multiclassification Tasks
Label 1 Label 2 Label 3 Unlabeled
Current Decision Boundary 21
Density Based Sampling for Multiclassification Tasks Def: Multi-class Information Density (MCID) (1) Density
(2) Unknownness
(3) Conflictivity Final Value Function 22
Density Based Sampling for Multiclassification Tasks Induce Density using a Gaussian Mixture Model
Each Mixture Sharing the Same Variance
Estimation via an EM Procedure
23
So far: New Proactive Learning Algorithm for Multiple Domain Experts
Multi-class Information Density (MCID) as a query strategy 24
Experiments Dataset
Simulated Noisy Labelers (except for Diabetes dataset) Narrow Experts: Classifier trained over partially noised dataset (expertise in only a subset of classes) Meta Expert: Classifier trained over the entire dataset 25
Baselines
Best Avg: learner always asks one of the narrow experts that has the highest average P(ans|x, k) Meta : learner always asks meta-oracle (expensive) BestAvg+Meta: joint optimization under uniform reliability assumption (Donmez et al., 2012) *Narrow: joint optimization using our algorithm *Narrow+Meta: with the presence of an meta oracle as well
Classification Performance Over Iterations
Cost Ratio of Narrow vs Meta: 1:6 27
Classification Performance for Different Cost Ratios
28
On other datasets
29
Classification Performance vs. Budget Allocated for Expertise Estimation
- Works for both when there are ground truth samples available & via majority votes - Is able to estimate expertise well enough with ~10% budget 30
Conclusions •
A new proactive learning algorithm with multiple class sensitive labellers accounts better than baselines
•
Efficient estimation of expert’s expertise via reduced per-class method
•
Multi-class Information Density (MCID) as a new active learning criteria for noised multi-class active learning 31
Future Work
•
Theoretical min-max bounds of the proposed algorithm, under different reliabilities and costs of the experts
•
Extend the framework to a crowdsourcing scenario with a larger pool of experts
32
Proactive Learning with Multiple Class-Sensitive Labelers
Seungwhan Moon, Jaime Carbonell Language Technology Institute School of Computer Science, Carnegie Mellon University DSAA 2014 Conference 10/30/2014 33
MCID Performance
34
Performance when expertise was estimated via Majority Vote