By using this site, you agree to the Privacy Policy and Terms of Use.
Accept
World of SoftwareWorld of SoftwareWorld of Software
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Search
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
Reading: New IT roles are putting AI to the test
Share
Sign In
Notification Show More
Font ResizerAa
World of SoftwareWorld of Software
Font ResizerAa
  • Software
  • Mobile
  • Computing
  • Gadget
  • Gaming
  • Videos
Search
  • News
  • Software
  • Mobile
  • Computing
  • Gaming
  • Videos
  • More
    • Gadget
    • Web Stories
    • Trending
    • Press Release
Have an existing account? Sign In
Follow US
  • Privacy
  • Terms
  • Advertise
  • Contact
Copyright © All Rights Reserved. World of Software.
World of Software > News > New IT roles are putting AI to the test
News

New IT roles are putting AI to the test

News Room
Last updated: 2026/05/17 at 12:55 PM
News Room Published 17 May 2026
Share
New IT roles are putting AI to the test
SHARE
New IT roles in companies are critically examining the output of AI tools in order to improve results in the future.

BalanceFormCreative / Shutterstock

Currently, many companies are moving toward putting together teams to evaluate AI. The newly created positions, functions or roles will become an indispensable safety net for organizations that introduce AI tools. As more and more AI pilot projects move into widespread production, the new teams should better evaluate AI results.

Subscribe to our CIO newsletter for more exciting insights, outlooks and background information for the IT community.

The rapid rise of AI agents has led to AI assessment teams beginning to take shape in recent months, reports Yasmeen Ahmad, managing director of product management, data and AI cloud at Google Cloud. Companies that now observe the behavior of AI agents in practice would realize that evaluation is not a one-time step but must be an ongoing practice.

KI-Evaluation – mehr als nice to have

At Google, AI assessment teams are embedded in the agent development groups, where both functions occur simultaneously. “As the agent developers work, the evaluation takes place in parallel, creating a shorter iteration cycle,” says Ahmad.

“Other companies have begun to set up AI evaluation workgroups within their larger AI and IT departments,” adds Maksim Hodar, CIO of software company Innowise.

In some cases, companies would combine data architects, security officers, and compliance officers into the evaluation team rather than hiring new employees from scratch. The groups would adopt a hybrid position between programming and ethical business practices. “It’s safe to say that AI evaluation teams are currently moving from a nice-to-have to a necessity.”

Automate responsibility?

Hodar has also observed that more and more companies are moving away from isolated AI implementation and are placing a greater focus on the “safety net”. Although a number of new tools for observability and governance, for example, focus on preventing AI errors, technology alone is not a complete solution. According to Hodar, people will be needed to decide whether the AI ​​tool is in line with company values ​​and regulations such as the GDPR. “Technology provides information, but the evaluation team still ultimately gives the green light because accountability cannot be automated.”

Human assessment teams need data from observability tools, but the technology itself cannot provide the necessary context for AI models and agents to correct incorrect results, says Google expert Ahmad. AI agents have become very good at passing output checks in test environments, but evaluation teams are needed to track their results in real-world situations. “Agentic applications may pass the initial unit test for a specific scenario, but agentic systems are non-deterministic decision makers and therefore behave unpredictably,” says Ahmad. You can’t test all the potential behaviors they might exhibit in the real world.

Understand the context of AI errors

While an observability tool provides data on token and tool usage, as well as tool failures and reasoning errors, human “evaluators” are required. They could fix many problems and provide context for common inference errors made by agents.

“When our internal assessment teams spend time with our AI agents, a large part of what they do is explore why the reasoning logic failed in some places,” explains Ahmad. The solution is usually to provide the right context at the right levels in the agent so that it can draw better conclusions.

Testing in a complex environment

“A good evaluation team also addresses several other aspects, including governance, cultural readiness, alignment with company workflows, and measurable business impact of AI tools,” adds Noe Ramos, vice president of AI operations at Agiloft, a contract lifecycle management provider. Technology alone cannot solve all these problems.

“The biggest hurdle isn’t technical – it’s human,” Ramos says. You can buy powerful tools and still have problems if people don’t trust them, don’t understand them, or don’t see how they fit into their work.

Like Hodar and Ahmad, Ramos sees growing demand for AI evaluation teams, although these roles would emerge as a skill set rather than formalized job titles.

Sometimes less is more

“AI evaluation is ultimately not just about security, but about ensuring that AI provides clarity and certainty of action rather than more unrest,” argues Ramos. Your company puts it internally like this: “We use AI to promote clarity and action – not to overwhelm teams with more dashboards.” Her team includes a head of AI operations, an AI agent engineer, and a head of GPT and AI systems. The aim is to integrate the evaluation into Agiloft’s AI operating model.

As the maturity level of organizations increases with increasing AI use, the leap towards a disciplined use of tools requires a structured evaluation function. “In my experience, one of the biggest risks is that AI initiatives are driven by the loudest voices rather than real operational priorities,” says Ramos. Rather, AI development should focus on amplifying the most grounded ideas to maximize the impact of AI in the enterprise.

According to Ramos, in most organizations the evaluation role or function must sit at the interface between IT, security, data management and operational stakeholders. Those responsible for AI evaluation also needed a deep understanding of how the company works. “One of the reasons why AI assessment fails is that companies don’t really understand their own workflows,” said Ramos. AI can only be evaluated intelligently if workflows are mapped, bottlenecks are identified and priorities are coordinated. (ajf/jd)

Sign Up For Daily Newsletter

Be keep up! Get the latest breaking news delivered straight to your inbox.
By signing up, you agree to our Terms of Use and acknowledge the data practices in our Privacy Policy. You may unsubscribe at any time.
Share This Article
Facebook Twitter Email Print
Share
What do you think?
Love0
Sad0
Happy0
Sleepy0
Angry0
Dead0
Wink0
Previous Article the return of the alliance on processors thanks to the Intel 18A-P process the return of the alliance on processors thanks to the Intel 18A-P process
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Stay Connected

248.1k Like
69.1k Follow
134k Pin
54.3k Follow

Latest News

the return of the alliance on processors thanks to the Intel 18A-P process
the return of the alliance on processors thanks to the Intel 18A-P process
Computing
the battery of hybrids wears out faster than that of electric ones, here’s why
the battery of hybrids wears out faster than that of electric ones, here’s why
Mobile
AI boom is not enough: California is planning a digital tax for cloud software
AI boom is not enough: California is planning a digital tax for cloud software
Software
Best Lease Origination Software for Faster Quote-to-Contract Workflows
Best Lease Origination Software for Faster Quote-to-Contract Workflows
Trending

You Might also Like

8 tips on how companies can become diverse and inclusive
News

8 tips on how companies can become diverse and inclusive

6 Min Read
Body Language: What you need to know about body language
News

Body Language: What you need to know about body language

7 Min Read
The best smart rings in the test
News

The best smart rings in the test

14 Min Read
Amazfit T-Rex Ultra 2 in the test: Large, strong and robust outdoor smartwatch
News

Amazfit T-Rex Ultra 2 in the test: Large, strong and robust outdoor smartwatch

12 Min Read
//

World of Software is your one-stop website for the latest tech news and updates, follow us now to get the news that matters to you.

Quick Link

  • Privacy Policy
  • Terms of use
  • Advertise
  • Contact

Topics

  • Computing
  • Software
  • Press Release
  • Trending

Sign Up for Our Newsletter

Subscribe to our newsletter to get our newest articles instantly!

World of SoftwareWorld of Software
Follow US
Copyright © All Rights Reserved. World of Software.
Welcome Back!

Sign in to your account

Lost your password?