TY - JOUR
T1 - Trustworthy Machine Learning in the Era of Foundation Models
AU - Han, Bo
N1 - Publisher Copyright:
© 2025 IEEE. All rights reserved, including rights for text and data mining, and training of artificial intelligence and similar technologies.
PY - 2025/11
Y1 - 2025/11
N2 - This position article examines trustworthy machine learning with foundation models by investigating the four essential aspects. In learning, we describe how pretraining, fine-tuning, and reinforcement learning enable models to acquire generalizable knowledge and emphasize the importance of high-quality, unbiased training data, as well as robust training methods. In reasoning, we summarize the fundamental methodology of training-free, post-training, and test-time scaling methods that enhance logical deduction, reasoning transparency, and systematic safety. In planning, we incorporate neurosymbolic methods that combine adaptable neural capabilities with formally verifiable symbolic reasoning, ensuring safe and accountable decision making. In multimodality, we investigate the need for multimodal integration, where aligning information from different sensory input sources is important to mitigate biases and errors. The article presents an interdisciplinary vision for incorporating capability, robustness, safety, and explainability to establish trustworthy foundation models, paving the way for their reliable deployment in real-world applications, such as financial and clinical decision making.
AB - This position article examines trustworthy machine learning with foundation models by investigating the four essential aspects. In learning, we describe how pretraining, fine-tuning, and reinforcement learning enable models to acquire generalizable knowledge and emphasize the importance of high-quality, unbiased training data, as well as robust training methods. In reasoning, we summarize the fundamental methodology of training-free, post-training, and test-time scaling methods that enhance logical deduction, reasoning transparency, and systematic safety. In planning, we incorporate neurosymbolic methods that combine adaptable neural capabilities with formally verifiable symbolic reasoning, ensuring safe and accountable decision making. In multimodality, we investigate the need for multimodal integration, where aligning information from different sensory input sources is important to mitigate biases and errors. The article presents an interdisciplinary vision for incorporating capability, robustness, safety, and explainability to establish trustworthy foundation models, paving the way for their reliable deployment in real-world applications, such as financial and clinical decision making.
UR - https://www.scopus.com/pages/publications/105024074580
U2 - 10.1109/MIS.2025.3621938
DO - 10.1109/MIS.2025.3621938
M3 - Journal article
AN - SCOPUS:105024074580
SN - 1541-1672
VL - 40
SP - 73
EP - 79
JO - IEEE Intelligent Systems
JF - IEEE Intelligent Systems
IS - 6
ER -