At QCon SF 2024, Wenjie Zi of Grammarly presented on the challenges inherent in machine learning projects. She began by citing daunting statistics: historical studies show failure rates as high as 85%, with recent research indicating little improvement. This high failure rate highlights a significant issue in the field: despite advances in AI technology, applying these technologies effectively in business contexts remains a substantial challenge.
The presenter highlighted five common pitfalls in machine learning projects. The first is tackling the wrong problem, where efforts are focused on issues that don’t align with the real business need. The second is challenges arising from data, such as poor quality, limited quantity, or biases that compromise the model. The third is the struggle to turn a successful model into a product, often due to integration and deployment challenges. The fourth is offline success but online failure, where models perform well in controlled settings but fail in real-world use. Lastly, unseen non-technical obstacles, such as resistance from stakeholders or organizational misalignment, can block progress.
There’s a famous saying in the machine learning world: garbage in, garbage out. Machine learning projects rely entirely on recognizing patterns in data. Therefore, if the data is flawed, it is highly likely that the conclusions drawn from the study will not be trustworthy. – Wenjie Zi
A central theme of Wenjie’s talk was the lifecycle of a machine learning project, which typically includes stages such as defining business goals, collecting and processing data, training models, deploying them, and monitoring their performance. She pointed out that failures often occur at various stages due to the lifecycle’s complexity. She emphasized the importance of having clear project objectives from the outset.
Another major challenge discussed was data management, encapsulated by the phrase “garbage in, garbage out.” The quality of data directly influences the success of machine learning projects. Issues such as data leakage, inadequate sample sizes, and biased data sets can lead to flawed conclusions and model failures. Zi noted that even sophisticated models from big tech companies and leading universities are not immune to these fundamental errors.
The transition from model development to production, known as MLOps, was another critical area. This requires an integrated approach involving multiple teams and systems, which increases the risk of failure. She highlighted the need for robust infrastructure and operations to support machine learning applications, noting that the actual machine learning code often constitutes only a small part of the overall system.
Later, Zi advocated for the “failing fast” approach as beneficial in machine learning projects. By quickly identifying inviable projects, teams can avoid further waste of resources and pivot towards more promising initiatives. This approach is part of a broader cultural shift towards embracing failures as learning opportunities.
Towards the end of her talk, Zi shared strategies to overcome these challenges, advocating for clear business goal definitions, rigorous data management practices, and a strong focus on the end-to-end integration of machine learning projects. Zi concluded her presentation with a quote from Charlie Munger, emphasizing the importance of learning from one’s own experiences and minimizing reliance on second-hand knowledge—a perspective that resonates deeply within the machine learning community.
Developers interested in learning more about Zi’s presentation may watch the InfoQ website where a video of her presentation will be available in the coming weeks.