Study Feature
Research on "Study Feature" broadly investigates the performance and limitations of various machine learning models across diverse tasks, focusing on areas like data compression, emotion recognition, remaining useful life prediction, and medical image generation. Current studies heavily utilize large language models (LLMs) and deep convolutional neural networks (CNNs), often exploring techniques like transfer learning, prompt engineering, and ensemble methods to improve model accuracy and robustness. This research is significant for advancing both fundamental understanding of model capabilities and for developing practical applications in fields ranging from healthcare and industrial maintenance to natural language processing and security.
Papers
Large Language Models as Financial Data Annotators: A Study on Effectiveness and Efficiency
Toyin Aguda, Suchetha Siddagangappa, Elena Kochkina, Simerjot Kaur, Dongsheng Wang, Charese Smiley, Sameena Shah
A Study on the Use of Simulation in Synthesizing Path-Following Control Policies for Autonomous Ground Robots
Harry Zhang, Stefan Caldararu, Aaron Young, Alexis Ruiz, Huzaifa Unjhawala, Ishaan Mahajan, Sriram Ashokkumar, Nevindu Batagoda, Zhenhao Zhou, Luning Bakke, Dan Negrut
Building a Language-Learning Game for Brazilian Indigenous Languages: A Case of Study
Gustavo Polleti
How to be fair? A study of label and selection bias
Marco Favier, Toon Calders, Sam Pinxteren, Jonathan Meyer
Can ChatGPT Detect DeepFakes? A Study of Using Multimodal Large Language Models for Media Forensics
Shan Jia, Reilin Lyu, Kangran Zhao, Yize Chen, Zhiyuan Yan, Yan Ju, Chuanbo Hu, Xin Li, Baoyuan Wu, Siwei Lyu