Gartner Predicts by 2027, Organizations Will Use Small, Task-Specific AI Models Three Times More ..
Gartner, Inc. predicts
that by 2027, organizations will implement small, task-specific AI models, with usage
volume at least three times more than those of general-purpose large
language models (LLMs).
While
general-purpose LLMs provide robust language capabilities, their response
accuracy declines for tasks requiring specific business domain context.
“The
variety of tasks in business workflows and the need for greater accuracy are
driving the shift towards specialized models fine-tuned on specific functions
or domain data,” said Sumit Agarwal, VP Analyst at Gartner.
“These smaller, task-specific models provide quicker responses and use less
computational power, reducing operational and maintenance costs.”
Enterprises can
customize LLMs for specific tasks by employing retrieval-augmented generation
(RAG) or fine-tuning techniques to create specialized models. In this process,
enterprise data becomes
a key differentiator, necessitating data preparation, quality checks,
versioning and overall management to ensure relevant data is structured to meet
the fine-tuning requirements.
“As enterprises increasingly
recognize the value of their private data and insights derived from their
specialized processes, they are likely to begin monetizing their models and
offering access to these resources to a broader audience, including their
customers and even competitors,” said Agarwal. “This marks a shift from a
protective approach to a more open and collaborative use of data and
knowledge.”
By commercializing their
proprietary models, enterprises can create new revenue streams while
simultaneously fostering a more interconnected ecosystem.
Implementing Small Task-Specific
AI models
Enterprises looking to implement
small task-specific AI models must consider the following recommendations:
· Pilot Contextualized Models: Implement small, contextualized models in
areas where business context is crucial or where LLMs have not met response
quality or speed expectations.
· Adopt Composite Approaches: Identify use cases where single model
orchestration falls short, and instead, employ a composite approach involving
multiple models and workflow steps.
· Strengthen Data and Skills: Prioritize data
preparation efforts to collect, curate and organize the
data necessary for fine-tuning language models. Simultaneously, invest in
upskilling personnel across technical and functional groups such as AI and data
architects, data scientists, AI and data engineers, risk and compliance
teams, procurement teams and business subject matter experts, to effectively
drive these initiatives.
Leave A Comment