For the majority of model developers, a combination of the two — human and automation — is where you’ll see the best balance between quality and accuracy versus lower costs and efficiency. We’ll explore why humans still need to be in the loop today.
With the recent advancements in LLMs, it’s increasingly common to hear discussions about automating annotation for machine learning (ML) models. It understandably raises a lot of questions among model developers—especially when coupled with the perception that human annotation is slow, expensive, and error-prone, whereas automation is cheaper and faster.
So which is the right choice for your model, based on where the technology stands today?
For the majority of model developers, a combination of the two — human and automation — is where you’ll see the best balance between quality and accuracy versus lower costs and efficiency.
In this post, we’ll explore why humans still need to be in the loop today. In Part 2, we’ll discuss where automation works best today in annotation workflows.
Most ML models today are “closed systems” that learn patterns, structures, and relationships from a fixed dataset during training and fine-tuning. These datasets are often enormous, pulled from publicly available information on the internet. Due to the inherent variability and inaccuracies in the data, it unsurprisingly results in a certain level of noise, which impacts model performance.
Improving these models requires new knowledge — embedding context, making corrections, etc. — that can only be added by human labelers and annotators today.
Why is this the case? If humans are replaced by models in the labeling or annotation workflow, the information within the model may be reorganized more efficiently but the amount of knowledge or capabilities within that closed system would be mostly unchanged.
Example: If a teacher asked students to correct each other’s quizzes, without providing the correct answers, the students’ collective knowledge might become better distributed since some understood parts of the material better than others. However, no new knowledge is created here and concepts that were missed are still unlikely to be fully grasped. It’s not until the teacher steps in to correct the remaining errors and teach new concepts that real learning occurs. Think of the teacher as the human annotators and the students as the model.
It’s true that not all models function as closed systems — there is ongoing research focused on developing AI agents that can autonomously pull in external information to enrich their knowledge base or add onto existing capabilities using external tools (such as accessing a calculator).
Currently, these types of ML models are at a developmental stage where they’re still being explored. They also don’t eliminate the need for human input, rather, they shift where human knowledge is required. Models will still need carefully curated examples to learn how to use these tools or data effectively and they’ll still need downstream human validation and editing to ensure outputs are high quality, accurate, and generating the right outcomes.
When improvements are made to a model, it directly correlates with how effectively human knowledge is transferred into the training data. As a result, it’s crucial that the information humans inject into the model is both accurate and relevant and addresses the weak spots where the model will benefit the most.
However, this comes with a couple of challenges.
Given all of these challenges and limitations, how should models and humans work together to produce the best results? Read Part 2 for more.
Image credit: Yutong Liu & Kingston School of Art / Better Images of AI / Talking to AI 2.0 / Liscenced by CC-BY 4.0
Jerome Pasquero holds a Ph.D. in electrical engineering from McGill University and has gone on to build leading-edge technologies that range from pure software applications to electromechanical devices. He has been a key contributor to the design of innovative and successful consumer products that have shipped to millions of users. Jerome is listed as an inventor on more than 120 US patents and has published over 10 peer-reviewed journal and conference articles. For the past 5 years, Jerome has been leading a number of AI product initiatives.