Enterprise Artificial Intelligence: Building Trusted AI in the Sovereign Cloud
The decade of responsible intelligence has begun — are you ready?
Enterprise AI is hitting a wall: Public models aren’t trained on your business data, but you can’t hand over your organization's proprietary information to a public system. The definitive roadmap for this new reality is Enterprise Artificial Intelligence: Building Trusted AI in the Sovereign Cloud, a new book written by OpenText leaders. Listen now to learn why this book is a must for organizations looking to move from isolated AI experiments to enterprise-grade deployments.
Learn more here: https://www.opentext.com/resources/enterprise-artificial-intelligence-building-trusted-ai
Enterprise Artificial Intelligence: Building Trusted AI in the Sovereign Cloud
Chapter 1: The evolution of enterprise data
Use Left/Right to seek, Home/End to jump to start or end. Hold shift to jump forward or backward.
AI doesn’t exist apart from data — it is data. Explore enterprise data — what it is, how it’s used, and how it can be optimized with an AI engine and governed by an Enterprise Information Management (EIM) platform in the cloud.
Chapter 1. The Evolution of Enterprise Data Not long ago, Enterprise Data was treated like an attic box, filled with old records, reports, and compliance paperwork. It was something you stored, not something you lived with. You climbed up to the attic only when you needed to check a number, prove a point, or satisfy an auditor. But today, data lives on the main floor. It's awake, wired, and running the business in real time. It informs every decision, powers every transaction, and if placed out on the street or made public, it can reveal more than you ever intended. But keep this data private and protected and it will drive your competitive edge. Artificial intelligence doesn't exist apart from data. It is data. Remembered, organized, and activated in motion. The enterprise is the same old house, but with the attic emptied and the brain downstairs. That's why the privacy and security laws that govern data must now be used to govern AI. As we move into the cognitive era and AI develops at faster rates, organizations and agencies will be required to treat information as a managed asset across its life cycle, not a passive archive. This chapter breaks down enterprise data, what it is, how it's used, and how it can be optimized with an AI engine and governed by an Enterprise Information Management platform in the cloud. We'll explore AI as the intelligence layer of your enterprise's information fabric, embedded into content services and analytics and stitched together by your operational business processes. Around 90% of the world's information expected to live inside organizations. Emails, documents, records, workflows, transactions, communications. Public LLMs are trained on open internet. Not enough for enterprise grade regulated insights. Hyperscalers don't always solve for data sovereignty, privacy, or industry specificity. Enterprises want to own and govern their content, compliance, security, competitive advantage. Agentic AI only delivers value when powered by trusted proprietary enterprise data. The hidden web. The real landscape, ten parts private, one part public. Most of the world's useful enterprise information lives behind the firewall. IDC reports that an organization's unstructured content, emails, reports, documents, images, recordings, makes up almost 90% of all data. This private data vastly outweighs the public web content that fuels today's generative AI. Yet much of it remains unmanaged, fragmented, or trapped across silos. That imbalance matters. The ratio of private to public data is roughly 10 to 1, which means the vast majority of the world's intelligence potential is hidden from public models. The real competitive advantage sits inside the enterprise. In contracts, design files, invoices, maintenance logs, clinical notes, and correspondence, provided it's governed, connected, and trusted. Enterprise information management was designed for exactly this challenge. EIM unifies, secures, and operationalizes enterprise data so it can be used responsibly and strategically. Managing information well means knowing where it lives, who owns it, who can see it, and when it changes. But not all data is created equal. Structured data, numbers in a database, can be sorted, queried, and reported with ease. Unstructured data, words, images, video, voice resists that order. It requires indexing, context, and classification. That's why the technologies that manage numbers and the technologies that manage words must differ. To understand the value of unstructured data, consider the scale of what sits beneath every business record. One employee might appear as a row in an HR database, but is also linked to thousands of documents, resumes, contracts, pay slips, correspondence, and performance reviews. An asset, whether an aircraft engine or a power turbine, might exist as a single record in an enterprise resource planning ERP system. But it's surrounded by a dense web of manuals, quality reports, inspections, and maintenance logs. Together, structured and unstructured information form the deep or hidden web of the enterprise, the data that's invisible to public search but vital to daily operations. Every digital artifact contributes to this hidden layer. Every email, report, draft, image, and chat thread. As mobile and collaboration technologies have moved inside the enterprise, the variety and velocity of content has exploded. EIM evolved to capture, classify, and govern this complexity, to make sure that what's created is also understood, retrievable, and compliant. But while EIM brought order to enterprise information, AI is now exposing its next frontier. Generative models, trained mostly on public data, can write, summarize, and predict. But they can't act within a business. They lack the governed, permissioned internal data that powers real decisions. Without it, AI can't perform agentic tasks like approving invoices, scheduling maintenance, or interpreting engineering drawings. To move beyond this plateau, AI needs what information management has offered for decades, a framework of secure, compliant, permissioned enterprise data. Only then can an AI operate responsibly inside the firewall, not just imitate intelligence from the outside. Managing this data means knowing where it lives, who owns it, who can see it, and when it changes. These fundamentals of EIM, permissions, metadata, and lifecycle control, are what make enterprise information trustworthy. These functions must now be applied to AI. Where to start? Let's take a look at where enterprise information lives. Where the data lives. Types of enterprise information. Every organization produces information for a handful of reasons, to record operations, enable communication, preserve knowledge, comply with regulations, and deliver value to customers. What began as structured record keeping, transactions, invoices, and inventory ledgers on mainframes expanded into a web of communication and collaboration, emails, shared documents, and digital workspaces. Today, these information streams can be understood as three broad classes of enterprise data that matter differently for AI. Human-generated content, machine generated data, and transactional or business network data. Each has its own structure, governance requirements, and role in AI model training. Together, they form the foundation for agentic intelligence inside the enterprise. Human generated content, the language of the enterprise. Human generated content includes documents, emails, scans, multimedia, case notes, and other forms of communication. It is rich in meaning and nuance but inherently unstructured. This is the content management domain, where information carries personal, contextual, and often sensitive data that demands careful classification, metadata tagging, and life cycle management. These materials contain the policy, precedent, and language that define how a business operates. Training AI on such content requires de-identification and strict governance, but the payoff is significant. This is where intent, business rules, and institutional knowledge live. When properly managed, unstructured content becomes the foundation for retrieval augmented generation, RAG, prompt libraries, and natural language reasoning, capabilities that allow agentic AI to act with understanding, not just automation. Machine generated data the Enterprise Nervous System. Machine generated data comes from the systems that power the organization, logs, telemetry, performance metrics, and monitoring streams. It's high in volume, velocity, and structure and tells the story of what the enterprise is doing in real time. This is the domain of observability and operational awareness, where every event and anomaly leaves a trace. Machine data provides the causal signals AI needs to act intelligently on infrastructure, detecting patterns, predicting failures, or recommending corrective actions before users notice. Its challenges lie in scale, retention cost, and the need to map raw signals to business context. When combined with policy and human content, this data enables an AI agent to respond autonomously while maintaining audibility and compliance. Transactional and business network data, the source of truth. Transactional data, purchase orders, invoices, shipping notices, and other structured messages represents the legal and economic truth of business operations. These records define the obligations between companies and are essential for compliance, taxation, and audit. Because they follow well-defined schemas and carry high semantic precision, they provide a reliable foundation for reasoning without ambiguity. For agentic AI, transactional data is what anchors decisions in fact. It enables agents to reconcile financial records, forecast cash flow, and identify exceptions in complex supply chains without fabricating results. Integrating this structured information with the unstructured and operational layers creates a full view of the enterprise, what's happening, why it's happening, and what to do next. In the following feature, MOBIS, a car manufacturer, created a parts production system designed to ensure quality and savings throughout the supply chain, informed by analytics and business intelligence. Case study MOBIS Headquartered in Seoul, South Korea, with subsidiaries in approximately 40 countries worldwide, MOBIS manages the supply chain for automobile industry heavyweights. The company created a parts production system designed to ensure quality and savings throughout the supply chain, from purchasing and inventory to sales and logistics, helping its clients stand out in the competitive automobile industry. MOBIS Parts Australia PTY Limited is the automotive supplier's Australian subsidiary. The automotive industry has ongoing competitive challenges, with other brands coming up with new products, new sales strategies, and new pricing methodologies. As a result, MPAU has to ensure that they have adequate systems and technologies in place to keep products competitive and operations agile. To react responsively to changing demands or competitors' offerings, the company needed a new business intelligence system that would support real-time inventory and dealer network reporting, monitor sales performance and pricing offers from vendors, and offer analytical capabilities to predict future sales and inventory requirements. From a logistics side, we are able to get clear visibility on business operations by integrating information from the back end to the front end of the BI system, allowing us to analyze the information coming from the back end system. IT Manager, MPAU. After testing several systems, MPAU ultimately chose an analytics suite based on its robust functionality, as well as its ease of use. The latter was a key factor to ensure end users would take to the system naturally to deepen distributor engagement by integrating seamlessly and intuitively into the day-to-day operations of approximately 140 to 160 users and dealers. The solution was able to integrate with data sources throughout the MPAU operations and dealer network with dashboards that offered each department from inventory and warehousing through sales and logistics a snapshot into their daily activities. Analytics capabilities give the company competitive advantage with the ability to not only view historical sales and inventory, but to forecast future needs as well. Now, instead of relying on a cumbersome reporting process and inaccurate predictions, users can compare historical data to current sales information in real time and project future sales. The result is a more efficient business environment with informed decision making that allows users to access and interact with data more reliably and the company to operate with greater agility in a competitive market. When information types work together. As illustrated in the above case study, the real power of enterprise data emerges when different types of information converge. When human knowledge, machine telemetry, and business transactions are governed, connected, and contextualized, they form the living architecture of an intelligent enterprise. Consider a finance shared services group using an AI assistant to reconcile mismatched purchase orders and invoices. The assistant analyzes invoice images in OCR, optical character recognition, output, in simpler terms, human generated content, then checks them against ERP transaction records, transactional data, and reviews system logs showing when invoices were received or approved, machine data. With the unified layer of metadata, capturing document lineage, access rights, and timestamps, the AI assistant can generate a defensible recommendation that reduces cycle time and preserves audit integrity. In another example, a legal department agent preparing litigation hold letters draws from policy documents and prior communications, human generated content, uses filing timelines and case metadata, transactional data, and verifies server access logs to confirm custodianship, machine data. The result is a draft that's both accurate and compliant, produced in minutes instead of days. The same principle plays out at scale in real organizations, as illustrated in the following case study. Case study, an independent energy company. An independent energy company, which is also a state-run public service, serves more than 150,000 customers. The company needed a way to manage growing volumes of information while maintaining compliance across multiple regulatory regimes. As a high performance company, we need to ensure that the right information is in the right place at the right time so we can make the right decision, explains the corporate records manager at the company. To do this, we have to incorporate compliance mandates, overcome information organization challenges, and constantly improve business processes. By extending enterprise information management with AI-driven search and automation, the company can now locate and act on information faster than ever. Intelligent retrieval services both structured and unstructured content, emails, spreadsheets, reports, and PDFs within one secure environment. AI summarization helps employees interpret large sets of engineering data and regulatory records, while machine learning models flag retention issues and automate compliance checks. The company now manages compliance and performance in tandem, using AI to make governance more proactive and less manual. EIM and Enterprise AI together allow the energy company to balance regulatory obligations with operational agility, turning information governance into an engine for insight rather than a constraint. These examples, from automated finance reconciliation to enterprise-wide energy management, illustrate the same truth. AI and automation deliver value not from any single data set, but from the relationships between them. When data is unified, trusted, and contextualized, it becomes more than information. It becomes intelligence. Every organization's digital story begins with its formats. The files, records, and containers we use to store information reflect the technologies and priorities of their time, from punched cards and print spool files to APIs built in JavaScript object notation, JSON, and AI ready datasets. Looking back across 50 years of enterprise information reveals a simple truth. Every leap in computing created a corresponding leap in content. The pre-web mainframe era 1970s to 1980s. The first generation of enterprise data lived inside the mainframe. It was structured, rigid, and optimized for machine efficiency. Fixed with text records, COBOL, common business oriented language, data files, and indexed ISAM, index sequential access method, formats replaced handwritten ledgers and paper journals. Storage was scarce, so every byte counted. Batch processing dominated, and the focus was on throughput rather than interaction. Systems were designed to process overnight, producing printed reports and spool files the next morning. Encoding schemes like EBCDIC, extended binary, coded decimal interchange code, kept everything proprietary and tightly coupled to the hardware that produced it. These early systems established the foundation for structured data discipline. They introduced schema control, versioned record formats, and the beginnings of what would become metadata, the idea that each field had meaning. The mainframe's rigidity forced organizations to think of data as an asset long before the term data governance existed. The PC and desktop publishing era, 1980s to 1990s. The arrival of the personal computer brought liberation along with fragmentation. Suddenly, employees could create content independently of the mainframe. Word perfect and early Microsoft Word documents, spreadsheet files, and desktop publishing outputs proliferated across offices and floppy disks. Information creation shifted from corporate centers to individual desktops. Reports, memos, and presentations multiplied in new digital formats like Doc, XLS, and PostScript. For the first time, documents were visual, editable, and printable at scale. This democratization of content fueled productivity but fractured control. Data that once lived in centralized systems was now scattered across hard drives and file shares. This era marked the birth of the content explosion that would eventually drive demand for enterprise-wide content management systems. Client, server, and web 1.0, 1990s. As organizations networked their systems, information began to move online. The rise of the web introduced HTML pages, GIF, and JPEG images, and early XML for structured exchange. Internal internets mirrored public websites, and the challenge shifted from creation to discovery. Client server architectures allowed employees to share databases and applications while browsers made it possible to publish information universally. The need for indexing and search sparked the first metadata models and document management systems. This brought about the search revolution, which was characterized by information's need for lifecycle thinking, creation, storage, retrieval, and disposal. Web 1.0 turned corporate knowledge into something dynamic, connected, and increasingly complex. It also laid the groundwork for governance. Once you find data, you need to decide who else should access it. Web 2.0 and the Collaboration Era, 2000s. The early 2000s introduced a more social and participatory web. Rich media, PDFA archival standards, XML and JSON formats, and multimedia encodings like MP3 and MP4 became common currency. User-generated content and collaboration tools entered the enterprise. Email archives, portals, and wikis joined traditional records systems. Content management evolved to handle persistent shareable documents that crossed departmental and even organizational lines. This was also the era when regulatory pressure met digital skill. The Sarbanes-Akley Act, the Health Insurance Portability and Accountability Act, HIPAA, and other compliance regimes forced companies to prove not only what they knew, but when and how they knew it. This convergence of regulation and collaboration cemented the needs for records management, version control, and policy-driven archiving, core pillars of modern EIM. The Big Data, Multimedia, and Mobile Era. By the 2010s, information had outgrown documents, streams, telemetry, and time series data flowed from mobile devices, sensors, and apps. New analytic file formats, Avro, Parquet, Orc, Optimize Rown, optimized for scale and speed became standard in data lakes and cloud warehouses. Video, voice, and large image repositories expanded exponentially as smartphones and digital experiences became the norm. Object storage and content delivery networks redefined the concept of file, turning everything into an addressable blob with metadata wrappers. This proliferation of data gave rise to the hidden web, the vast universe of unstructured content behind the firewall. It was both an opportunity and a liability for the enterprise, an opportunity to mine with analytics and apply AI training, but risky in terms of cost and compliance. The cognitive era late 2010s to the present. Today's content landscape is fluid, interconnected, and multimodal. Data moves not just between people and systems, but between machines. APIs, especially REST, representational state transfer, and GraphQL interconnect microservices. Application specific bundles, containers, notebooks, structured logs, represent new hybrid document types. This era is defined by interoperability and automation. Information is both consumed and produced by AI, machine learning, and digital agents. Tokenized text corpora, embedded metadata, and semantic tagging enables retrieval augmented generation and contextual reasoning. Where earlier eras optimize for format efficiency, today's focus is meaning. The modern enterprise must unify structured and unstructured data across every modality, voice, image, text, and transaction into governed ecosystems that support both analytics and intelligent action. In the cognitive era, formats aren't static. They're interfaces between human intent and machine understanding. The same life cycle that once applied to documents, capture, manage, process, search, archive, now extends to knowledge itself. Across each era, the theme remains consistent. Technology changes, but the need for trust and context endures. From COBOL reports to cloud APIs, every new format redefines not just how data is stored, but how it is governed, shared, and understood. The lesson is simple. Information management evolves with the medium. What began as control over files has become control over intelligence. The formats of the past were about legibility. The formats of the future are about learning. Why governance comes first? Enterprise information management has always been about that trust. It organizes the information estate across capture, manage, process, search, and archive, tying content lifecycle directly to the business processes that create it. Done well, governance improves insight, reduces risk, and lowers compliance costs. It's not an add-on to AI strategy. It's the precondition. Discover how UBS has centralized its information in an EIM platform to govern its information and comply with regulations in the case study below. Case study, UBS. In response to the compliance requirements posted by Sections 302 and 906 of the Sarbanes Ackley Act, UBS, one of the world's leading financial institutions, implemented an internal certification process for financial reports, in which senior executives formally certify their financial figures and processes using a sub-confirmation process. During the internal certification process, appropriate persons are notified via email when their input is required and are then granted personalized access to the relevant documents on the UBS intranet. All relevant processes are archived and tracked in a log file. The CEO and group controller, generally the CFO, issue a final certification for the Security Exchange Commission only when all internal processes have been completed. The UBS Corporate Governance Portal enables the company's business managers worldwide to collaborate in developing internal and external business reports. Relevant departments have access to complete overview and status of the certification processes at all times. All related processes have been automated and simplified, expediting the certification process. Mapping EIM to AI, Internet, Intranet, and Extranet. EIM provides a useful mental model for understanding AI maturity. Just as information moves from public to private domains, AI evolves from broad generalization to contextual intelligence. Internet equals generative AI. The outermost layer represents public knowledge, open data, and generalized language models that are excellent for ideation, first drafts, and exploration. Generative AI works much like the Internet itself, vast, connected, and creative, but limited by a lack of organizational context or precision. Intranet equals Agentic AI. The middle layer mirrors an organization's internal network. Here, data is private, permissioned, and workflow aware. Agentic AI can reason over internal systems, act on approved workflows, and make bounded decisions under governance controls. This is where AI stops merely describing and starts doing, automating tasks, augmenting staff, and enforcing policy through action. Extranet equals Artificial General Intelligence, AGI. The innermost layer represents the future frontier, where AI collaborates safely across organizations and systems. Like an extranet connecting trusted partners, AGI would reason fluidly across boundaries, sharing insights while maintaining trust and compliance between entities. As AI moves inward, from public to private to shared domains, its context, accuracy, and value increase. But so too does the need for governance. The deeper the intelligence operates within your organization's core data, the greater the responsibility to secure, audit, and align it with human and regulatory boundaries. By the numbers, the energy cost of training AI. Training a foundational model like GPT-3 consumed approximately 1,287 megawatt hours of electricity, emitting around 502 metric tons of CO2, roughly equivalent to the annual emissions of 112 gasoline-powered cars. In 2024, a study found that up to 30% of the power used during large language model training runs is wasted through inefficient scheduling and hardware usage, meaning the same outcome could be achieved with significantly less energy. Projections indicate that AI training demand could consume 8 TW in 2024 and reach 652 TW by 2030, representing an over 80-fold increase in electricity use in just six years. Data quality and the physics of learning. Garbage in, garbage out has never been more relevant. The accuracy of any AI model depends on the quality of the data it consumes. In statistical terms, more data improves probability. If that data is relevant, consistent, and clean. Modern machine learning doesn't simply add volume, it learns to assign weight to the signals that matter most through repeated training and feedback. Over time, the system develops a sense of what's meaningful and what's noise, much like a human learning through experience. When those inputs are incomplete, inconsistent, or poorly governed, the model fills the gaps on its own. That's when hallucinations happen. Confident, convincing answers that are entirely wrong. As model complexity increases, so does the risk of these errors. The antidote is curation, grounding AI in governed, high-integrity data. There's also a cost to quality ratio at play. The better the data, the less energy and time a model wastes in training and inference. As AI models grow larger and compute demands rise, the physics of learning becomes a matter of efficiency as much as accuracy. Well-curated data reduces redundancy, minimizes reprocessing, and reduces the deployment footprint of AI operations. The new figure of merit lies in the balance between data quality, training time, and energy consumption. A reminder that better governance isn't just safer, it's smarter and more sustainable. Why AI Must Follow Data's rules? Traditional software processed data and moved on. AI doesn't, it remembers. Every piece of information it encounters becomes part of its internal landscape, shaping how it reasons, responds, and behaves in the future. That memory makes AI different, and it makes governance essential. If data has always required rules, AI now extends those rules to a new frontier. The same life cycle that governs enterprise information, capture, manage, process, search, and archive or dispose, must now apply to intelligent systems. We have to decide what a model is allowed to learn, what it should retain, what it must forget, and how its knowledge can be verified or audited over time. Without those boundaries, memory becomes liability. Uncontrolled accumulation turns information into risk. Disciplined lifecycle management turns it into insight and value. Governance isn't just about protecting data anymore. It's about teaching intelligence how to remember responsibly. In the following case study, we explore how MAN Diesel and Turbo is using EIM as its foundation for governance and to achieve compliance. Case study MAN Diesel and Turbo. MAN Diesel and Turbo, headquartered in Augsburg, Germany, is the world's leading manufacturer of large board diesel engines and turbo machinery. The company employs around 14,900 staff at more than 100 international sites. Primarily in Germany, Denmark, France, Switzerland, the Czech Republic, India, and China. Diesel engines in container freighters or luxury liners are some of the largest products in the world and among those with the longest lifespan. They have to function for decades and be regularly maintained. One of the world's leading manufacturers in this field, MAN Diesel and Turbo, needs to keep important technical documents for a minimum of 30 years and sometimes indefinitely. The company was looking for an information management solution to ensure high quality maintenance and successfully refute any claims for liability arising from alleged construction faults. To help achieve compliance, MAN Diesel and Turbo turned to records management capabilities contained within an extended EIM solution, along with Application Governance and Archiving, AGA. The combined solutions bring together diverse applications to preserve information and context. Approximately 1,000 service staff in Germany and Denmark use it to archive over 4,000 process-related transaction files every day. In many service processes, paper-based transactions are now a thing of the past as existing paper files are being digitized. MAN Diesel and Turbo is saving valuable time spent on searches and maintaining paper archives along with the large number of their digital archives. The integrated solution is also reducing maintenance requirements through the replacement of legacy systems, enabling the company to modernize its infrastructure, digitally transform key processes, and comply with regulations. AI's Plateau on Data and How to Move Beyond It. By 2026, more than 80% of enterprises will be using generative AI models or APIs in production. That scale of adoption raises the stakes on governance. Spending on AI governance tools is expected to more than quadruple by 2030 as organizations work to manage risk, permissions, lineage, and model oversight while moving from experimentation to full integration. Enterprise use of generative AI is now mainstream, but much of that activity still sits on top of public data and generic models. Great for content, limited for action. In 2024, 65% of organizations regularly used Gen AI, a share that has continued to rise into 2025, yet many remain stuck at good demos rather than operational impact. The gap isn't enthusiasm, it's data. To work for your business, AI needs governed access to private, permissioned information so it can reason in context and execute workflows safely. Experience shows that scale without strong foundations rarely delivers results. Companies with mature data and AI capabilities consistently outperform their peers. Those that break away do so because they treat data strategy and governance as first principles, not afterthoughts. In practical terms, that means bringing private content under control, enforcing access permissions, and applying policy-rich metadata so models can retrieve relevant information. Act within guardrails and demonstrate accountability. When AI operates inside the firewall, connected to your governed data estate, it stops guessing and starts working. Instead of offering general answers, it can take informed accountable action. It can resolve an invoice exception by referencing the purchase order, vendor terms, and approval history. It can draft and route a policy update that automatically respects permissions, retention schedules, and regulatory requirements. It can answer a support case in natural language using approved knowledge and record that interaction in the system of record. Each of these capabilities depends on the same EIM backbone that reduces fragmentation, governs access, and connects unstructured information to the business processes that rely on it. To oversimplify, an EIM platform provides order. It governs information across systems, silos, and geographies. AI provides context. It learns from that governed data to deliver insights, automation, and decision support. In the cognitive era, AI will unlock the next generation of information management. The implication is straightforward. Without private permission data and the governments to use it responsibly, generative AI hits a ceiling. It can summarize the internet, but it can't approve an invoice, schedule a repair, or resolve a customer exception inside your systems. The way forward lies in enterprise data that is cataloged, classified, and access controlled, supported by auditable pipelines that allows AI agents to pull facts, take defined actions, and leave a verifiable trail. That's how organizations turn widespread adoption into lasting business value. The Fast Five Download. One, AI maturity starts with data maturity. AI is only as strong as the data it learns from. Generative models built on public data hits a plateau. Agentic AI requires governed, private, and permissioned information. Invest in data foundations before scaling AI capability. Use EIM to identify high value private data sets and apply AI where process value is clear. Two, governance is the new infrastructure. The principles that keep enterprise data compliant, metadata, permissions, lifecycle control, and auditability are now prerequisites for AI. Governance defines how intelligence learns, remembers, and acts safely within your organization. Three, move inward for value, Internet, Intranet, Extranet. Public data powers generative AI, content. Internal data powers agentic AI action. And connected ecosystems will one day power AGI collaboration. Each step inward increases accuracy, accountability, and value, and demands stronger controls. Four, data quality defines AI performance and its footprint. Curated, clean data reduces hallucinations, improves reliability, and lowers compute waste. The new figure of merit balances quality, training time, and energy consumption. Better data governance now means faster, greener, and more accurate AI later. 5. AI must follow data's rules. Unlike traditional software, AI remembers what it sees. That makes its memory part of your governance landscape. Treat AI learning as a life cycle. Decide what models can learn, what they should retain, what they must forget, and how you'll audit them.