Failed Experiments Boost AI's Chemical Reaction Predictions by 10%
June 14, 2025
A recent study published in Science Advances highlights the significance of negative experimental results in enhancing AI model training, particularly for predicting chemical reactions.
Researchers trained a transformer-based language model using data from both successful experiments and at least 40 times as many unsuccessful experiments, leading to improved prediction accuracy.
The IBM team built upon their previous work with transformer models, utilizing chemical reactions sourced from United States Patent and Trademark Office patents for their training.
The findings suggest that incorporating data from both successful and unsuccessful experiments can significantly enhance the accuracy of AI models in predicting chemical reactions.
The fine-tuned model demonstrated over a 10% improvement in predicting successful reactions compared to a model trained solely on successful data.
Mara Graziani, the principal investigator, emphasizes the need for a shift in academic publishing incentives to promote the reporting of failed experiments, as the current system often prioritizes significant individual contributions.
To achieve this, the researchers optimized a latent space representation, allowing for better differentiation between successful and unsuccessful reactions.
A significant challenge in this field is the lack of venues for publishing negative experimental results, which limits the data available for training AI models.
The research involved creating reward functions for reinforcement learning from human feedback, a method commonly used in other fields but less prevalent in chemistry due to the sparse nature of positive data.
The analysis included two types of unsuccessful experiments: those yielding unexpected but relevant products and those that provided no significant insights, both of which offer valuable data.
Graziani compares the learning process in chemistry to language learning, where understanding errors is crucial for mastering complex rules.
Summary based on 1 source
Get a daily email with more AI stories
Source

IBM • Jun 13, 2025
How to make AI models more accurate: Embrace failure