Unveiling the World of Mass Excellence Cluster

undefined
 
 
 
 
Towards Extreme Multi-label Text Classification
Through Group-wise Label Ranking
Jie Xiong @ SEKE23
2023-07-08
 
 
 
 
1. Introduction
Extreme Multi-label text Classification 
(XMC) aims to
tag a document with 
relevant labels 
from an 
extremely
large label set
.
A
l
b
e
r
t
 
E
i
n
s
t
e
i
n
 
w
a
s
 
a
G
e
r
m
a
n
-
b
o
r
n
 
t
h
e
o
r
e
t
i
c
a
l
p
h
y
s
i
c
i
s
t
.
c
l
a
s
s
i
f
i
c
a
t
i
o
n
d
o
c
u
m
e
n
t
L
a
b
e
l
s
 
 
 
 
1. Introduction
 
 
 
 
2. Proposed Method
We proposed a novel Group-
wise label Ranking loss to
address the rank-missing and
label imbalance problems of
the existing XMC models.
We proposed a new XMC
model with GRank loss, named
X-Grank. We trained 
it 
end-to-
end to eliminate the cascaded
errors in existing methods.
 
 
 
 
2.1 GRank: Group-wise ranking loss
 1. The original intention: all positive labels have higher scores than the negative labels
 2. We made an equivalent transformation:
 3. Use log-sum-exp 
LSE
to approximate the max function and get the Group-wise
Ranking (GRank) loss:
 4.
 
Grank loss has a mutual constraint gradients and thus avoids label imbalance problem.
 
 
 
 
2.2 The application of Grank loss
T
o
t
a
l
 
l
o
s
s
 
i
n
 
M
o
d
e
l
 
T
r
a
i
n
i
n
g
T
h
e
 
l
o
s
s
 
o
f
 
t
h
e
 
r
e
t
r
i
e
v
i
n
g
 
s
t
a
g
e
T
h
e
 
l
o
s
s
 
o
f
 
t
h
e
 
r
a
n
k
i
n
g
 
s
t
a
g
e
 
 
 
 
3. Experiments: Datasets, Settings, Metrics
 
 
 
 
3.1 Recall results
 
 
 
 
3.2 Ranking results
 
 
 
 
3.3 Ablation analysis and Tail labels
 
 
 
 
3.4 Efficiency
undefined
 
 
 
 
Slide Note

Big Data Analysis & Intelligent Decision

Embed
Share

Cluster ENIGMASS aims to drive groundbreaking research in neutrino physics, foster collaborations, enhance scientific education, and support innovative projects. By leveraging manpower, education initiatives, and strategic investments, the cluster propels advancements in particle physics and instrumentation technology. With a focus on scientific communication and fostering partnerships, ENIGMASS paves the way for transformative discoveries in the realm of astroparticle physics.

  • Mass Excellence Cluster
  • Neutrino Physics
  • Scientific Collaboration
  • Particle Physics
  • Instrumentation Technology

Uploaded on Mar 03, 2025 | 0 Views


Download Presentation

Please find below an Image/Link to download the presentation.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.If you encounter any issues during the download, it is possible that the publisher has removed the file from their server.

You are allowed to download the files provided on this website for personal or commercial use, subject to the condition that they are used lawfully. All files are the property of their respective owners.

The content on the website is provided AS IS for your information and personal use only. It may not be sold, licensed, or shared on other websites without obtaining consent from the author.

E N D

Presentation Transcript


  1. Towards Extreme Multi-label Text Classification Through Group-wise Label Ranking Jie Xiong @ SEKE23 2023-07-08

  2. 1. Introduction Extreme Multi-label text Classification (XMC) aims to tag a document with relevant labels from an extremely large label set. Labels Positive label Group ?? document classification Albert Einstein was a German-born theoretical physicist. Negative label Group ?? 2

  3. 1. Introduction Problems in existing XMC Methods: The ranking learning between labels is missing, but we hope all scores in ??are greater than those in ??. The sparse positive labels lead to serious label imbalance in XMC, which may hinder the model convergence. 3

  4. 2. Proposed Method We proposed a novel Group- wise label Ranking address the rank-missing and label imbalance problems of the existing XMC models. We proposed model with GRank loss, named X-Grank. We trained it end-to- end to eliminate the cascaded errors in existing methods. loss to a new XMC 4

  5. 2.1 GRank: Group-wise ranking loss 1. The original intention: all positive labels have higher scores than the negative labels 2. We made an equivalent transformation: 3. Use log-sum-exp LSE to approximate the max function and get the Group-wise Ranking (GRank) loss: 4. Grank loss has a mutual constraint gradients and thus avoids label imbalance problem. 5

  6. 2.2 The application of Grank loss The loss of the retrieving stage The loss of the ranking stage Total loss in Model Training 6

  7. 3. Experiments: Datasets, Settings, Metrics 7

  8. 3.1 Recall results 8

  9. 3.2 Ranking results 9

  10. 3.3 Ablation analysis and Tail labels 10

  11. 3.4 Efficiency 11

  12. Thanks for Listening!

More Related Content

giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#giItT1WQy@!-/#