Text Classification
Text classification aims to automatically categorize text into predefined categories, driven by the need for efficient and accurate information processing across diverse domains. Current research focuses on leveraging large language models (LLMs) like BERT and Llama 2, often enhanced with techniques such as fine-tuning, data augmentation, and active learning, alongside traditional machine learning methods like SVMs and XGBoost. These advancements are improving the accuracy and efficiency of text classification, with significant implications for applications ranging from medical diagnosis and financial analysis to social media monitoring and legal research. A key challenge remains ensuring model robustness, interpretability, and fairness, particularly when dealing with imbalanced datasets or noisy labels.
Papers
ChatGraph: Interpretable Text Classification by Converting ChatGPT Knowledge to Graphs
Yucheng Shi, Hehuan Ma, Wenliang Zhong, Qiaoyu Tan, Gengchen Mai, Xiang Li, Tianming Liu, Junzhou Huang
The Benefits of Label-Description Training for Zero-Shot Text Classification
Lingyu Gao, Debanjan Ghosh, Kevin Gimpel
Differentiate ChatGPT-generated and Human-written Medical Texts
Wenxiong Liao, Zhengliang Liu, Haixing Dai, Shaochen Xu, Zihao Wu, Yiyang Zhang, Xiaoke Huang, Dajiang Zhu, Hongmin Cai, Tianming Liu, Xiang Li
Graph Neural Networks for Text Classification: A Survey
Kunze Wang, Yihao Ding, Soyeon Caren Han
MEGClass: Extremely Weakly Supervised Text Classification via Mutually-Enhancing Text Granularities
Priyanka Kargupta, Tanay Komarlu, Susik Yoon, Xuan Wang, Jiawei Han
Multidimensional Perceptron for Efficient and Explainable Long Text Classification
Yexiang Wang, Yating Zhang, Xiaozhong Liu, Changlong Sun