AI specialist Naren Gangavarapu has warned against organisations treating the costs of consuming AI services the same way they manage traditional software licenses, comparing the potential for AI tokens to burn money to that of a poker machine. 

Naren, who previously led a large tech modernisation program at Northern Beaches Council in Sydney, tells ADAPT that organisations need clear visibility into their use of AI tokens, the basic units of data that AI models use to process and generate text, images or audio content.

Users are charged based on the number of tokens processed by large language models (LLMs).

To keep AI costs manageable as tools are deployed at scale, Naren says organisations need to recognise that token consumption occurs with every interaction, effectively every ‘keystroke’. 

Enterprises that don’t make this distinction could end up spending huge amounts of money that they can’t claw back, he says. 

“You need to know how many tokens you’re buying and how many tokens are being consumed. This is very important. 

Basically, your tokens have to be budgeted per use case, and you have to make sure that you ‘take stock’ on a daily or weekly basis”, he says.  

Costs can also be controlled through organisational policies that govern AI model selection, ensuring prompt workflows use models that are aligned and behave as expected. 

Policies should define when more complex, and more expensive models should be used, versus when simpler direct prompts and cheaper models are sufficient, he says.

“You need to have business ownership of these budgets so that [the] technology [group] doesn’t overspend because they get excited about building models and might end up giving a huge bill to the business.”

Agents need a ‘control plane’ 

The bulk of Australian CIOs surveyed by ADAPT say they are investing in AI agents, but many remain in the initial awareness and active exploration stages.

As of February 2026, 22% said they had targeted agentic implementations live in selected teams (up from 12% last August), while only 8% are scaling agents across their operations.

Naren says many agents can be turned on and deployed quickly across organisations, often faster than the process of hiring staff and navigating governance requirements.

He says agents without governance are experiments but those with a control plane become a platform. He compares this to how an air traffic control centre, which provides visibility for aircraft leaving and landing at an airport.

“[With agents], you need to know what you can control. What can the agents do and what can’t they do? Then you need to build observability, which at any point tells you what’s happening. 

“There are incidents where agents have been trained to do something and they have gone and deleted databases. Say that’s happened and you have to explain…what’s happened [to the organisation]. There has to be auditability; this means you have to be observing it [the agent] in real time.

“And then the governance part is put on top…the policies on which models are used…and what data an agent has been approved to access”, he says.

Naren highlights that while most enterprises have ‘systems of record’ or knowledge in their core applications, intellectual property (IP) concerns are often preventing AI capabilities from being integrated into these platforms.

Instead, data is staged in platforms like Databricks and Snowflake, with strict access controls applied.

“Then you give selective control to agents in data stages so that your core is protected. That way, your security and IP access is manageable”, he says.

Finally, an intervention layer – which monitors, intercepts and regulates the actions of autonomous agents – provides a ‘kill switch’ that allows humans to override agentic processes.

“If we don’t have that, it’s quite unpredictable as to what some of these agents can do because they’re learning so quickly. If you don’t architect [agentic infrastructure in this way], the agents can thrive along with your risk…your cost gets out of control and you have no idea [which agent] has done what”, he says.

Naren adds that many organisations continue to struggle with maintaining accurate inventories of APIs and microservices, leaving them potentially exposed to cyber attackers. These risks, he says, are amplified when agents are allowed to proliferate across organisations without the right controls.

“We need to make sure that we bring in those governance controls and be extra careful to ensure that governance and observability are built in from the get-go to manage the scale and risk aspect of [AI agents].”

Intelligence by subscription, AI consciousness

Naren is sceptical about the CEO narrative framing AI intelligence as a subscription service or suggesting that AI models may become conscious.

He argues the tech market has created hype and a narrative that puts unnecessary pressure on organisations to buy these solutions.

“This is an adoption tactic that [vendors] use. I’ve seen this happen with CRMs, with cloud, with every new product that comes out in the market. Organisations have already invested huge amounts of money in their core systems; they’re not going to move away from them. They’ve invested hundreds of millions of dollars. 

“So how do you shake that up? By creating this consistent narrative. There are enough people with enough intelligence around the world with public representation as well as within governments, that are closely watching this.”

He says governments should play an active role in ensuring organisations, countries and their citizens are “not impacted by such narratives” or pressured into hasty decisions.

“AI achieving consciousness? I don’t see it in my career…it’s human beings that train it [AI].”

Contributors
Byron Connolly Head of Programs & Value Engagement at ADAPT
Byron Connolly is a highly experienced technology and business journalist, editor, corporate writer, and event producer, and ADAPT’s Head of Programs and... More

Byron Connolly is a highly experienced technology and business journalist, editor, corporate writer, and event producer, and ADAPT’s Head of Programs and Value Engagement.

Prior to joining Adapt, he was the editor-in-chief at CIO Australia and associate editor at CSO Australia. He also created and led the well-known CIO50 awards program in Australia and The CIO Show podcast.

As the Head of Programs, Byron creates valuable insights for ADAPT’s community of senior technology and business professionals, helping them reach their organisational and professional goals. With over 25 years of experience, he has a passion for uncovering stories about the careers and personal philosophies of Australia’s top technology and digital executives.

When he is not working, Byron enjoys hot yoga, swimming, running, and spending time with his family.

Less
Naren Gangavarapu Transformation and AI Specialist
Naren is an award-winning executive with over 25 years of experience implementing technology strategies that drive measurable business outcomes. He has a... More
Naren is an award-winning executive with over 25 years of experience implementing technology strategies that drive measurable business outcomes. He has a proven track record of managing and executing large-scale digital transformation initiatives across complex service industries, including transportation (rail, roads, maritime), planning and environment, and all levels of public and private sectors.
He specialises in developing and implementing robust digital strategies that not only enhance customer experiences but also build resilient digital infrastructures and optimise investments. His passion lies in turnaround projects, where I thrive on transforming challenges into growth opportunities, ensuring that every technological advancement drives tangible, sustainable results for the business.
Less
budgets management compliance