Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- ArticleOctober 2024
SpecFormer: Guarding Vision Transformer Robustness via Maximum Singular Value Penalization
Computer Vision – ECCV 2024Pages 345–362https://doi.org/10.1007/978-3-031-72949-2_20AbstractVision Transformers (ViTs) are increasingly used in computer vision due to their high performance, but their vulnerability to adversarial attacks is a concern. Existing methods lack a solid theoretical basis, focusing mainly on empirical training ...
- research-articleAugust 2024
NegativePrompt: leveraging psychology for large language models enhancement via negative emotional stimuli
IJCAI '24: Proceedings of the Thirty-Third International Joint Conference on Artificial IntelligenceArticle No.: 719, Pages 6504–6512https://doi.org/10.24963/ijcai.2024/719Large Language Models (LLMs) have become integral to a wide spectrum of applications, ranging from traditional computing tasks to advanced artificial intelligence (AI) applications. This widespread adoption has spurred extensive research into LLMs across ...
- research-articleJuly 2024
Dynamic evaluation of large language models by meta probing agents
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 2591, Pages 62599–62617Evaluation of large language models (LLMs) has raised great concerns in the community due to the issue of data contamination. Existing work designed evaluation protocols using well-defined algorithms for specific tasks, which cannot be easily extended to ...
- research-articleJuly 2024
CompeteAI: understanding the competition dynamics of large language model-based agents
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 2526, Pages 61092–61107Large language models (LLMs) have been widely used as agents to complete different tasks, such as personal assistance or event planning. While most of the work has focused on cooperation and collaboration between agents, little work explores competition, ...
- research-articleJuly 2024
Open-vocabulary calibration for fine-tuned CLIP
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 2121, Pages 51734–51754Vision-language models (VLMs) have emerged as formidable tools, showing their strong capability in handling various open-vocabulary tasks in image recognition, text-driven visual content generation, and visual chatbots, to name a few. In recent years, ...
-
- research-articleJuly 2024
Selective mixup helps with distribution shifts, but not (only) because of mixup
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 1958, Pages 47948–47964Context. Mixup is a highly successful technique to improve generalization by augmenting training data with combinations of random pairs. Selective mixup is a family of methods that apply mixup to specific pairs e.g. combining examples across classes or ...
- research-articleJuly 2024
The good, the bad, and why: unveiling emotions in generative AI
- Cheng Li,
- Jindong Wang,
- Yixuan Zhang,
- Kaijie Zhu,
- Xinyi Wang,
- Wenxin Hou,
- Jianxun Lian,
- Fang Luo,
- Qiang Yang,
- Xing Xie
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 1161, Pages 28905–28934Emotion significantly affects our daily behaviors and interactions. Although recent generative AI models, such as large language models, have shown impressive performance in various tasks, it remains unclear whether they truly comprehend emotions and ...
- research-articleJuly 2024
Position: what can large language models tell us about time series analysis
- Ming Jin,
- Yifan Zhang,
- Wei Chen,
- Kexin Zhang,
- Yuxuan Liang,
- Bin Yang,
- Jindong Wang,
- Shirui Pan,
- Qingsong Wen
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 895, Pages 22260–22276Time series analysis is essential for comprehending the complexities inherent in various realworld systems and applications. Although large language models (LLMs) have recently made significant strides, the development of artificial general intelligence (...
- research-articleJuly 2024
Position: TRUSTLLM: trustworthiness in large language models
- Yue Huang,
- Lichao Sun,
- Haoran Wang,
- Siyuan Wu,
- Qihui Zhang,
- Yuan Li,
- Chujie Gao,
- Yixin Huang,
- Wenhan Lyu,
- Yixuan Zhang,
- Xiner Li,
- Hanchi Sun,
- Zhengliang Liu,
- Yixin Liu,
- Yijue Wang,
- Zhikun Zhang,
- Bertie Vidgen,
- Bhavya Kailkhura,
- Caiming Xiong,
- Chaowei Xiao,
- Chunyuan Li,
- Eric Xing,
- Furong Huang,
- Hao Liu,
- Heng Ji,
- Hongyi Wang,
- Huan Zhang,
- Huaxiu Yao,
- Manolis Kellis,
- Marinka Zitnik,
- Meng Jiang,
- Mohit Bansal,
- James Zou,
- Jian Pei,
- Jian Liu,
- Jianfeng Gao,
- Jiawei Han,
- Jieyu Zhao,
- Jiliang Tang,
- Jindong Wang,
- Joaquin Vanschoren,
- John C Mitchell,
- Kai Shu,
- Kaidi Xu,
- Kai-Wei Chang,
- Lifang He,
- Lifu Huang,
- Michael Backes,
- Neil Zhenqiang Gong,
- Philip S. Yu,
- Pin-Yu Chen,
- Quanquan Gu,
- Ran Xu,
- Rex Ying,
- Shuiwang Ji,
- Suman Jana,
- Tianlong Chen,
- Tianming Liu,
- Tianyi Zhou,
- William Wang,
- Xiang Li,
- Xiangliang Zhang,
- Xiao Wang,
- Xing Xie,
- Xun Chen,
- Xuyu Wang,
- Yan Liu,
- Yanfang Ye,
- Yinzhi Cao,
- Yong Chen,
- Yue Zhao
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 813, Pages 20166–20270Large language models (LLMs) have gained considerable attention for their excellent natural language processing capabilities. Nonetheless, these LLMs present many challenges, particularly in the realm of trustworthiness. This paper introduces TRUSTLLM, a ...
- research-articleJuly 2024
A general framework for learning from weak supervision
- Hao Chen,
- Jindong Wang,
- Lei Feng,
- Xiang Li,
- Yidong Wang,
- Xing Xie,
- Masashi Sugiyama,
- Rita Singh,
- Bhiksha Raj
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 290, Pages 7462–7485Weakly supervised learning generally faces challenges in applicability to various scenarios with diverse weak supervision and in scalability due to the complexity of existing algorithms, thereby hindering the practical deployment. This paper introduces a ...
- research-articleJune 2024
Diversify: A General Framework for Time Series Out-of-Distribution Detection and Generalization
IEEE Transactions on Pattern Analysis and Machine Intelligence (ITPM), Volume 46, Issue 6Pages 4534–4550https://doi.org/10.1109/TPAMI.2024.3355212Time series remains one of the most challenging modalities in machine learning research. Out-of-distribution (OOD) detection and generalization on time series often face difficulties due to their non-stationary nature, wherein the distribution changes ...
- surveyMarch 2024
A Survey on Evaluation of Large Language Models
- Yupeng Chang,
- Xu Wang,
- Jindong Wang,
- Yuan Wu,
- Linyi Yang,
- Kaijie Zhu,
- Hao Chen,
- Xiaoyuan Yi,
- Cunxiang Wang,
- Yidong Wang,
- Wei Ye,
- Yue Zhang,
- Yi Chang,
- Philip S. Yu,
- Qiang Yang,
- Xing Xie
ACM Transactions on Intelligent Systems and Technology (TIST), Volume 15, Issue 3Article No.: 39, Pages 1–45https://doi.org/10.1145/3641289Large language models (LLMs) are gaining increasing popularity in both academia and industry, owing to their unprecedented performance in various applications. As LLMs continue to play a vital role in both research and daily use, their evaluation becomes ...
- research-articleJanuary 2024
Optimization-Free Test-Time Adaptation for Cross-Person Activity Recognition
Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies (IMWUT), Volume 7, Issue 4Article No.: 183, Pages 1–27https://doi.org/10.1145/3631450Human Activity Recognition (HAR) models often suffer from performance degradation in real-world applications due to distribution shifts in activity patterns across individuals. Test-Time Adaptation (TTA) is an emerging learning paradigm that aims to ...
- research-articleDecember 2023
Distilling out-of-distribution robustness from vision-language foundation models
NIPS '23: Proceedings of the 37th International Conference on Neural Information Processing SystemsArticle No.: 1428, Pages 32938–32957We propose a conceptually simple and lightweight framework for improving the robustness of vision models through the combination of knowledge distillation and data augmentation. We address the conjecture that larger models do not make for better teachers ...
- research-articleDecember 2023
Adaptive Memory Networks With Self-Supervised Learning for Unsupervised Anomaly Detection
IEEE Transactions on Knowledge and Data Engineering (IEEECS_TKDE), Volume 35, Issue 12Pages 12068–12080https://doi.org/10.1109/TKDE.2021.3139916Unsupervised anomaly detection aims to build models to effectively detect unseen anomalies by only training on the normal data. Although previous reconstruction-based methods have made fruitful progress, their generalization ability is limited due to two ...
- research-articleNovember 2024
PromptRobust: Towards Evaluating the Robustness of Large Language Models on Adversarial Prompts
- Kaijie Zhu,
- Jindong Wang,
- Jiaheng Zhou,
- Zichen Wang,
- Hao Chen,
- Yidong Wang,
- Linyi Yang,
- Wei Ye,
- Yue Zhang,
- Neil Gong,
- Xing Xie
LAMPS '24: Proceedings of the 1st ACM Workshop on Large AI Systems and Models with Privacy and Safety AnalysisPages 57–68https://doi.org/10.1145/3689217.3690621The increasing reliance on Large Language Models (LLMs) across academia and industry necessitates a comprehensive understanding of their robustness to prompts. In response to this vital need, we introduce PromptRobust, a robustness benchmark designed to ...
- research-articleOctober 2023
Non-IID always Bad? Semi-Supervised Heterogeneous Federated Learning with Local Knowledge Enhancement
- Chao Zhang,
- Fangzhao Wu,
- Jingwei Yi,
- Derong Xu,
- Yang Yu,
- Jindong Wang,
- Yidong Wang,
- Tong Xu,
- Xing Xie,
- Enhong Chen
CIKM '23: Proceedings of the 32nd ACM International Conference on Information and Knowledge ManagementPages 3257–3267https://doi.org/10.1145/3583780.3614991Federated learning (FL) is important for privacy-preserving services by training models without collecting raw user data. Most FL algorithms assume all data is annotated, which is impractical due to the high cost of labeling data in real applications. To ...
- research-articleSeptember 2023
Power control algorithm for wireless sensor nodes based on energy prediction
Wireless Networks (WIRE), Volume 30, Issue 1Pages 517–532https://doi.org/10.1007/s11276-023-03504-4AbstractConventional wireless sensors have difficulty solving the problem of energy limitation, especially in sensor networks in hard-to-reach extreme areas. In order to solve the problem that it is difficult to charge wireless sensors in the field using ...
- research-articleAugust 2023
Exploring Vision-Language Models for Imbalanced Learning
International Journal of Computer Vision (IJCV), Volume 132, Issue 1Pages 224–237https://doi.org/10.1007/s11263-023-01868-wAbstractVision-language models (VLMs) that use contrastive language-image pre-training have shown promising zero-shot classification performance. However, their performance on imbalanced dataset is relatively poor, where the distribution of classes in the ...