Large Language Models Do Multi-Label Classification Differently
Under review
We investigate how autoregressive LLMs perform multi-label classification, with a focus on subjective tasks, by analyzing the output distributions of the models in each generation step. We find that their predictive behavior reflects the multiple steps in the underlying language modeling required to generate all relevant labels as they tend to suppress all but one label at each step.
Recommended citation: Marcus Ma*, Georgios Chochlakis*, Niyantha Maruthu Pandiyan, Jesse Thomason, and Shrikanth Narayanan. "Large Language Models Do Multi-Label Classification Differently." arXiv preprint arXiv:2505.17510 https://arxiv.org/abs/2505.17510