Logo

AI-driven data: preparing for a zero-trust future

Gartner report finds that, by 2028, as AI data proliferates, organisations will shift to a zero-trust model to safeguard business outcomes.

  • 2 weeks ago Posted in

Gartner forecasts that by 2028, a significant portion of organisations will adopt a zero-trust approach to data governance. This shift arises from the increasing presence of AI-generated data, which complicates verification processes.

"Organisations can no longer implicitly trust data or assume it was human generated," stated Wan Fui Chan, Managing VP at Gartner. As AI-generated content becomes more prevalent and indistinguishable from human-created data, implementing authentication and verification measures is crucial to protecting business and financial outcomes.

Large language models (LLMs), often trained on diverse sources including books and research papers, are at risk of repetitive AI content. Gartner's 2026 CIO and Technology Executive Survey found that 84% of respondents plan to boost GenAI funding.

This surge in both AI adoption and investment will lead to models being trained more on former AI outputs. The consequence could be model collapse where AI results might fail to mirror reality.

"As AI-generated content becomes more prevalent, regulatory requirements for verifying ‘AI-free’ data are expected to intensify in certain regions, cited Chan, emphasising the variances in global regulatory standards. Identifying and tagging AI-generated data will be critical.

Success in this regulatory landscape depends on tool availability and workforce expertise in information management and metadata solutions. This will support data cataloguing, differentiating proactive organisations.

Proactive management practices, like active metadata management, provide advantages. Such practices allow organisations to swiftly analyse and automate decisions across their datasets.

  1. Appoint an AI Governance Leader: Assign a role specifically for AI governance to ensure robust zero-trust policies. This leader should coordinate closely with data and analytics teams.
  2. Foster Cross-Functional Collaboration: Form teams across cybersecurity and data sectors to assess AI data risks and strengthen policies accordingly.
  3. Leverage Existing Governance: Enhance current governance frameworks, tuning security, and metadata management to incorporate AI-related policies.
  4. Adopt Active Metadata Practices: Utilise real-time alerts to identify and correct data inaccuracies or biases, safeguarding critical systems.

Such actions will be a significant in combatting the risks of unchecked AI-generated data, preserving organisational integrity in a rapidly evolving digital landscape.

Snowflake partners with OpenAI in a $200M collaboration, integrating AI models for enhanced data...
SUSE introduces a framework to evaluate digital sovereignty gaps as per the 2025 EU guidelines.
Wipro presents an AI-enabled approach to enterprise processes, combining consulting and automation...
NFON AG strengthens its position in AI communications with a new modular licence model and partner...
Major expansion in UK's AI programme offers free training to equip 10 million individuals with...
Hack The Box and LinkedIn Learning have joined forces, aiming to address the cybersecurity skills...
1Password launches an enhanced Partner Program to support the growing demand for identity security...
A recent report by Zapier highlights the pivotal role AI will play in shaping enterprise operations...