- datapro.news
- Posts
- How the last 6 months of AI rewrote the rules of data engineering
How the last 6 months of AI rewrote the rules of data engineering
THIS WEEK: Part 2 of The Twelve Days of AI Christmas

Dear Reader…
Welcome back to our festive journey through 2025's defining AI moments. If the first half of the year laid the groundwork for transformation, the second half accelerated the revolution. What began as efficiency gains and productivity tools evolved into something far more profound: a fundamental reimagining of what data engineering means in an age of autonomous text to action AI.
As the seasons changed, the industry confronted hard truths about cost, governance, and the widening gap between those who could harness AI and those left behind. Here are the final six days of our AI Christmas.

Day Seven: The Context Window Singularity
In July, OpenAI's release of ChatGPT-5 delivered on the promise of the "context window singularity." With a massive context window of 10 million tokens (roughly 750,000 words), this model fundamentally changed the scope of what an AI could "keep in mind" simultaneously.
For data engineers, this was a true game changer. The most immediate application was in modernising legacy systems. Engineers could feed the entire documentation, database schema, stored procedure code, and application logic of a 20-year-old mainframe system into the model in a single prompt. ChatGPT-5 could map thousands of tables and complex business logic instantly, erasing the massive cognitive overhead previously required to understand tangled, undocumented architectures.
The model's agentic skills and advanced reasoning allowed it to build, test, and optimise ETL pipelines in hours rather than weeks. One test case showed the model building a complete ETL pipeline, a task estimated at three weeks for a human team, in just six hours.
However, this capability came with a stinging infrastructure reality check. While the API abstracted complexity for casual users, organisations running private instances or fine-tuned models hit a wall. Processing a 10-million token context required massive memory allocations that exceeded standard enterprise hardware configurations. The cost of inference for these massive contexts proved prohibitive for many real-time applications. Organisations discovered that "near-infinite memory" came with "near-infinite costs" if not managed correctly.
This trend created a paradox: AI could now understand everything, but affordably processing that understanding became the new bottleneck. It forced a discipline of "context curation," deciding what not to send to the model, and validated the need for the "lean stack" approach identified at the year's beginning.
Day Eight: The Great Divergence
By late summer, a critical divergence emerged in how AI was being adopted across the economy. Comprehensive research from Anthropic and usage data from OpenAI revealed "dual trajectories" that were reshaping the digital landscape.
On the consumer side, AI was acting as a powerful equaliser. OpenAI's analysis of 2.5 billion daily interactions across 700 million weekly users showed that AI was closing gender gaps (usage shifted from 80% masculine to 48% masculine) and reducing income-based disparities.
Conversely, enterprise usage was concentrating power. Anthropic's Economic Index showed that 77% of enterprise API usage followed automation patterns, compared to roughly 50% on consumer platforms. Furthermore, AI usage intensity correlated directly with national GDP (0.7% increase per 1% GDP growth).
This divergence had profound implications for data strategy and the competitive landscape. Enterprises weren't just using AI as a helpful assistant; they were deploying it as "replacement infrastructure." Automated workflows were replacing human processes entirely. "Lights-out" processes, where operations run without human intervention, became the goal for data-mature organisations.
This created a competitive wedge. Organisations with sophisticated, "AI-ready" data infrastructure could leverage these automation patterns to pull ahead exponentially. The statistic that 77% of enterprise usage was automation-focused became a blueprint for how businesses were fundamentally rewiring their operational DNA.
Conversely, organisations with poor data quality, siloed systems, or legacy debt were locked out of the "automation dividend." They couldn't automate because their data wasn't trusted. This widened the gap between data-mature and data-poor firms, threatening the survival of the latter.
Data engineers found themselves at the centre of this storm. They were no longer just supporting business analysts; they were building the very rails upon which the automated enterprise ran.
Day Nine: The Governance Reckoning
In August, the AI industry was rocked by a bombshell report from MIT's Networked Agents and Decentralised AI initiative. The study, titled "The GenAI Divide: State of AI in Business 2025," claimed that 95% of generative AI pilot projects at companies were failing to reach production with measurable value.
The report detailed a brutal "funnel of failure": while 80% of organisations explored AI tools and 20% launched pilot projects, only 5% reached production with a marked impact on profit and loss. The study identified the primary culprits not as model limitations, but as engineering and organisational failures.
One of the most significant concepts introduced was the "verification tax." Because AI models (even advanced ones) tended to be "confidently wrong," employees were forced to spend excessive time double-checking every output. This verification time eroded the promised productivity gains, effectively neutralising the return on investment.
The report also highlighted that companies were attempting to layer generative AI on top of "already broken, messy workflows." Without addressing underlying process issues and data fragmentation, AI simply accelerated the production of bad outcomes.
This failure rate served as a massive vindication for foundational data engineering. The "boring" work of data quality and governance was suddenly recognised as the only path to the 5% success bracket.
The successful 5% of companies prioritised making data "AI-ready," cleaning, organising, and governing data before applying models. They established robust frameworks for data lineage, privacy, and compliance. The lack of governance was identified as a critical failure point. Successful teams established clear AI adoption strategies and governance frameworks early, rather than treating them as an afterthought.
The report highlighted that the highest returns weren't in flashy consumer-facing bots, but in unglamorous "back-office automation" (data summarisation, pipeline monitoring, anomaly detection). This reoriented data teams towards operational efficiency and away from "hype" projects.
August 2025 was the month the hype died, and the real work of data governance was recognised as the primary determinant of AI success.
Master ChatGPT for Work Success
ChatGPT is revolutionizing how we work, but most people barely scratch the surface. Subscribe to Mindstream for free and unlock 5 essential resources including templates, workflows, and expert strategies for 2025. Whether you're writing emails, analyzing data, or streamlining tasks, this bundle shows you exactly how to save hours every week.
Day Ten: The Death of Batch
In September, the frontier of AI moved beyond language into the physical world with the rise of Large Behaviour Models (LBMs). Unlike large language models, which operated in a "world of words," LBMs were focused on behaviour, embodiment, and action. They learned from observing physical demonstrations and analysing sensor streams to control robots and autonomous systems.
For data engineers, the rise of LBMs represented the "death of batch." Traditional ETL pipelines, built to process static snapshots of text data overnight (batch processing), were fundamentally incompatible with LBMs. LBMs required continuous, asynchronous streams from the messy, unpredictable physical world.
Engineers faced a "multi-rate, multi-modal, asynchronous mess" of data. This included robotic sensors producing thousands of measurements per second (force, torque, acceleration), camera feeds, lidar scans, and human teleoperation inputs.
This trend forced a complete re-architecture of the enterprise data stack. The LBM era mandated a hybrid compute model. Gigantic base model training happened on cloud clusters, but inference (the robot's actual decision-making) had to occur on-device (at the edge) within milliseconds. Data engineers had to design synced data pipelines that straddled the edge and cloud, ensuring logs and retraining data flowed back to the cloud without choking bandwidth. This brought chips like NVIDIA's Jetson Thor into the data engineer's domain.
The data engineer's role shifted from data "janitor" to "quality assessor of behaviour." Engineers had to develop programmatic scoring techniques to classify human demonstration data by "learnability" or efficiency. Was the teleoperator's movement efficient, or was it sloppy? This subjective quality assessment became a core engineering task.
The goal became building "data circulatory systems," living, real-time architectures that were always pumping and never still, continuously feeding the "real-time data flywheels" of the LBMs.
Day Eleven: The Silent Revolution
By October, a "silent revolution" was underway in the core discipline of data modelling. The traditional three-tier approach (conceptual, logical, physical) that had defined data architecture for decades was being dismantled and rebuilt by AI.
The "text-to-schema paradigm" allowed engineers to convert complex business requirements directly into foundational entity-relationship diagrams in hours rather than weeks. This utilised generative AI to parse documents and user stories, eliminating the "blank canvas" problem and dramatically accelerating stakeholder validation.
The specific demands of AI workloads forced a "logical model bifurcation." Traditional logical models split into two distinct, parallel blueprints that data architects had to govern: a business-facing semantic layer enforcing consistency for business intelligence and standardising business definitions, and a specialised logical feature model optimising for machine learning operations, managing granular, high-quality derived values in feature stores to ensure model reproducibility.
Perhaps most radically, the philosophy of governance was inverted. Traditional governance involved auditing data after it was deployed. The new AI-driven systems enabled proactive, real-time governance that prevented issues before they occurred.
Automated, real-time, column-level data lineage tracking became mandatory for debugging and regulatory compliance, extending beyond static tables to cover complex AI workflows and notebooks. In consumption-based cloud platforms, inefficient data models directly translated to budget overruns. Cost governance was elevated to a primary architectural consideration, managed by AI systems that automatically handled indexing and resource allocation.
The role of the data architect evolved from a manual "drafter" of diagrams to a "refiner and governor" of AI-suggested structures, requiring a blend of traditional modelling skills and new machine learning operations competencies.
Day Twelve: The Action Imperative
As the year drew to a close in December, the cumulative impact of agentic AI, LBMs, and deep research crystallised into the final, defining trend: the governance of autonomous action.
The industry realised that the "text-to-action" capability, where AI doesn't just suggest a plan but executes it in the real world, was the single greatest opportunity and risk for 2026. The coming year would be defined not by the intelligence of models, but by the governance of their autonomy.
This realisation set the stage for the "data engineering mandate for 2026." Future architectures would need to move beyond governing data quality to governing action itself.
Data engineers began conceptualising new architectural layers. Transactional safety layers would provide protocols to revert unauthorised AI actions or "roll back" real-world changes where possible. Action lineage systems would trace not just data transformations, but the decision pathways taken by autonomous agents. "Why did the agent order 10,000 units of stock?" required a verifiable audit trail of the agent's reasoning chain.
Constraint modelling involved hard-coding boundaries and ethical constraints directly into the data model and API layers that no AI agent, regardless of its intelligence, could cross.
As we close the book on 2025, the transformation is complete. The data engineer of 2024, a builder of pipelines and a cleaner of tables, is obsolete. In their place stands a new professional profile: The AI systems Architect.
They manage behaviour, not just bytes. They curate context, ensuring the 10-million token window is filled with signal, not noise. They govern action, building the safety rails for autonomous agents. They optimise efficiency, leveraging the "lean AI" stack to avoid cloud bankruptcy.
The seismic shifts of 2025 have settled into a new landscape. The hype has evaporated, replaced by the hard, cold reality of engineering physics: latency, memory, cost, and risk. For those who can master these elements, the opportunities are boundless. For those who cannot, the 95% failure rate awaits.
As we look towards 2026, one thing is clear: data engineering is no longer a support function. It is the central nervous system of the AI-driven enterprise. And just like the twelve days of Christmas, each gift of 2025 has accumulated into something far greater than the sum of its parts.

