Labeling Information Enhancement for Multi-label Learning with - - PowerPoint PPT Presentation

labeling information enhancement for multi label learning
SMART_READER_LITE
LIVE PREVIEW

Labeling Information Enhancement for Multi-label Learning with - - PowerPoint PPT Presentation

Labeling Information Enhancement for Multi-label Learning with Low-rank Subspace An Tao* , Ning Xu, and Xin Geng Southeast University, China Outline Introduction 1 The LIEML Algorithm 2 Experiments 3 Conclusion 4 Introduction 1


slide-1
SLIDE 1

An Tao*, Ning Xu, and Xin Geng Southeast University, China

Labeling Information Enhancement for Multi-label Learning with Low-rank Subspace

slide-2
SLIDE 2

Outline

1

Introduction

2

The LIEML Algorithm

3

Experiments

4

Conclusion

slide-3
SLIDE 3

1

Introduction

slide-4
SLIDE 4

Introduction

We call such label as logical label.

  • Each label is regarded to be either relevant or irrelevant to the instance.

Label set: Y = {sky, water, cloud, beach, plant, house}

Labeling information in multi-labellearning is categorical in essence.

Features: each pixels in the picture

In traditional multi-label learning:

slide-5
SLIDE 5

Introduction

However…

Q: The two pictures can’t be differed with only logical labels. A: To describe the pictures better, we extend the logical label to be numerical. Same logical label set Y = {sky, water, cloud, beach, plant, house}

This new label is called numerical label.

slide-6
SLIDE 6

Introduction

Specification of numerical label Specification of logical label

Use 𝒛" ∈ {−1,1}) to denote the logical label vector. Use to denote the numerical label vector.

  • Label element of 𝒛" = 1 : relevant to the instance.
  • Label element of 𝒛" = −1 : irrelevant to the instance.
  • Label element of

> 0 : relevant to the instance.

  • Label element of

< 0 : irrelevant to the instance.

  • Absolute value of label element of

: reflects the degree to which the label describes the instance.

slide-7
SLIDE 7

Introduction

Logical Label Numerical Label

Label Enhancement (LE)

Multi-label Model

Predictive Model Induction

Feature Overview of our LIEML algorithm for multi-label learning

+

Step 1: Step 2:

ü LE can be seen as a data preprocessing step which aims to facilitate in learning a better model.

slide-8
SLIDE 8

2

The LIEML Algorithm

slide-9
SLIDE 9

The LIEMLAlgorithm

Label Enhancement

Linear Model:

Symbol Definition:

  • : Dataset
  • : i-th instance vector
  • : i-th logical label vector
  • : i-th numerical label vector
  • is a weight matrix.
  • is a bias vector.
slide-10
SLIDE 10

The LIEMLAlgorithm

Label Enhancement

Linear Model:

  • For convenient describing, we set

.

The model becomes:

Label Enhancement

We then construct a stacked matrixZ:

Target Matrix

slide-11
SLIDE 11

The LIEMLAlgorithm

The optimization problem of LE becomes:

Label Enhancement

  • L(Z): logistic loss function

Not deviate too much

ü It prevents the label values in Z from deviating the original values too much.

slide-12
SLIDE 12

The LIEMLAlgorithm

Label Enhancement

Low-rank Assumption A: We assume that the stacked matrix Z belongs to an underlying low-rank subspace.

Not deviate too much Rank(Z)↓

  • L(Z): nuclear norm

and squared function Q: Why construct the stacked matrix Z? ü The stacked matrix Z is therefore an underlying low-rank matrix.

Labeling Information↑

slide-13
SLIDE 13

The LIEMLAlgorithm

Label Enhancement

Not deviate too much Rank(Z)↓

The target function T1 for optimization is yielded as:

slide-14
SLIDE 14

The LIEMLAlgorithm

Predictive Model Induction

The target function T2 we wish to minimize is: We build the learning model through an adapted regressor based on MSVR.

  • , ,

, , , , , , . .

  • is a nonlinear transformation of x to a higher dimensional feature space.
slide-15
SLIDE 15

The LIEMLAlgorithm

slide-16
SLIDE 16

3

Experiments

slide-17
SLIDE 17

Experiments

Experiment Configuration

Five evaluation metrics widely-used in multi-label learning:

  • BR, CLR, ECC, RAKEL, LP, and ML2

Six well-established multi-label learning algorithms:

  • Ranking-loss, One-error, Hamming-loss, Coverage, and Average precision

Ten benchmark multi-label data sets:

slide-18
SLIDE 18

Experiments

Experimental Results

slide-19
SLIDE 19

Experiments

Experimental Results

slide-20
SLIDE 20

Experiments

Experimental Results

LIEML ranks 1st in the most cases! The model of LE in LIEML is linear, but nonlinear in ML2, it is uneasy for LIEML to beat ML2 with the less efficient linear way. ü The results of the experiment validate the effectiveness of our LIEML algorithm for multi-label learning.

slide-21
SLIDE 21

4

Conclusion

slide-22
SLIDE 22

My personal website: https://antao.netlify.com/

Conclusion

This paper proposes a novel multi-label learning method named LIEML, which enhances the labeling information by extending logical labels into numerical labels. The labeling information is enhanced by leveraging the underlying low- rank structure in the stacked matrix.

Major Contribution More Information

Our lab website: http://palm.seu.edu.cn/

slide-23
SLIDE 23

Thank You!

slide-24
SLIDE 24

The LIEMLAlgorithm

Label Enhancement

To optimize the target function T1 :

Alternative Step 1: gradient decent Step 2: shrinkage ü To improve the speed of convergence, we begin with a large value µ1 for µ , and decay as .

slide-25
SLIDE 25

The LIEMLAlgorithm

Predictive Model Induction

To minimize T2 (Θ ;m ), we use an iterative quasi-Newton method called Iterative Re-Weighted Least Square (IRWLS). The solution for the next iteration ( Θ(k+1

) and m( k+1) ) of

is obtained via a line search algorithm along ( Θ and m ) .

  • ,

, , , , .

≈ The quadratic problem can be solved as:

slide-26
SLIDE 26

Experiments

Experiment Configuration

slide-27
SLIDE 27

Experiments

Experiment Configuration

Zhi-Hua Zhou, and Min-Ling Zhang. "Multi-label Learning." (2017): 875-881.

slide-28
SLIDE 28

Introduction

In traditional multi-label learning:

An instance

Some features Label 1 Label 2 Label d . . . Yes Yes No learning Model . . . No No Yes . . . predicting