DA2PL 2016 Paderborn, Nov 7, 2016
Statistical inference for incomplete ranking data: A comparison of two likelihood-based estimators
Inés Couso Mohsen Ahmadi Eyke Hüllermeier
Statistical inference for incomplete Ins Couso ranking data: A - - PowerPoint PPT Presentation
DA2PL 2016 Paderborn, Nov 7, 2016 Statistical inference for incomplete Ins Couso ranking data: A comparison of two Mohsen Ahmadi likelihood-based estimators Eyke Hllermeier Statement of the problem Sample of N individuals. K
DA2PL 2016 Paderborn, Nov 7, 2016
Inés Couso Mohsen Ahmadi Eyke Hüllermeier
❖ Sample of N individuals. ❖ K alternatives. ❖ Every individual provides a pairwise comparisons. ❖ Goal: estimate the most popular complete ranking over
❖ Intermediate goal: estimate the probability distribution
N
i=1
N
i=1
π∈SK
N
i=1
N
i=1
π∈E(τi)
i=1 θπ−1(i) θπ−1(i)+θπ−1(i+1)+...+θπ−1(K) .
θi θi+θj .
i=1
j6=i
θi θi+θj
K
1,2
1,3
2,3
1,3
1,2
2,3
1,2
2,3
1,3
1,3
2,3
1,2
2,3
1,2
1,3
2,3
1,3
1,2
K(K−1) 2
❖ The MLM (solid), the FLM
(dashed) and the TLM (dotted) are compared.
❖ Three different coarsening
processes are considered.
❖ The same PL parameter is taken
in the three cases.
❖ Left column: Euclidean distance
true par. - par. estimate.
❖ Right column: Kendall distance
true mode - predicted ranking.
500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35
set λ1,2 = . . . = λ3,4 = 1/6. selected uniformly at random. In
500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4
xperiment, λ1,2 = 1 corresponds to the top-2
500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 500 1000 1500 2000 0.05 0.1 0.15 0.2 0.25 0.3 0.35
ranks: λi,j ∝ (8 − i − j). with a higher probability than
❖ MLM is theoretically the best one, but it involves to
❖ FML is simpler, but it ignores the coarsening process. It
❖ Biased estimations of the parameter do not imply non-
❖ Future directions: search for computational acceptable