Sage Analyst Summit: Keeping the Winning Playbook While Evaluating Emerging Changes to the Game

Connect, grow, deliver

TBR spent two days in Atlanta, listening to and speaking with Sage’s management team as part of the company’s annual Analyst Summit, and we walked away impressed. This is a company that knows itself and its strengths. It knows where it needs to improve. It knows where the pain points and constraints are, and has always done a good job of navigating between the two.

 

Most importantly, the company knows its customers, which should come as no surprise considering how long Sage has been serving its SMB install base. Sage has leveraged these strengths and established a large, sticky install base from which to pursue opportunities adjacent to its core business.

 

Sage is focused on three interlocking areas — connect, grow, deliver — which President Dan Miller described during the event:

  • Connect through trusted partner networks
  • Grow by winning new logos through a verticalized suite motion
  • Deliver real, measurable productivity using AI

Each pillar represents a separate part of the company’s go-to-market strategy, but Grow stands out as the most vital to the company’s growth trajectory. Landing and expanding with new logos is the company’s greatest source of revenue growth, with vertical-specific and business operations solutions offering some of the greatest upsell potential. Aligned with this strategy, the company is a disciplined but active acquirer, onboarding new IP to enhance these sales motions.

 

Long-term, AI presents opportunities for the company to upsell into its finance and accounting (F&A) core. As Sage leans into its strengths while building for the future, its ability to scale AI and industry depth across a known and trusted customer base may prove to be the company’s most valuable asset.

Landing with F&A, then expanding with payroll, HR and operations management

Sage’s land-and-expand strategy starts with a stronghold in finance and builds outward through operational adjacencies. Most customers enter through core accounting —typically via Intacct — and expand into areas like payroll, HR, and inventory or distribution management as their needs mature. Vertical-specific modules are critical to this motion, especially in midmarket industries where Sage can tailor functionality to operational nuances.

 

The company reinforces expansion by packaging these capabilities into suites, streamlining procurement and positioning itself as more than just a financial system. Sales teams are trained to identify expansion triggers early; signs like API adoption, workflow customization or manual process bottlenecks often indicate opportunities. Although the company’s product maturity varies across the portfolio, Sage has seen success in service- and product-centric verticals, enabling the company to upsell and cross-sell. This approach, combined with a focus on ease of integration and strong partner involvement, is helping Sage grow account value without overpromising in its product road map.

AI at Sage: Workflow-first, ROI-driven

Sage management spent much time discussing its ambitions in AI. From TBR’s perspective, the tone was very grounded. Although the company will never be at the cutting edge of AI innovation, management did a great job of articulating the current opportunities to upsell AI capabilities. Finance and accounting workflows offer many sales opportunities for Sage to pursue, and the company is investing in R&D to capitalize on them. Similar to many of its application peers, Sage intends to approach agentic AI and generative AI development on a use-case-by-use-case basis. In Sage’s case, this is even more prudent as SMB customers face greater budgetary restrictions and require ROI to be realized in the first year.

 

Sage management highlighted AP automation, time-saving prompts and variance analysis as key areas where the company is achieving success with AI-powered automation. Like several peers, the company’s Copilot solution serves as the unified user interface (UI) for engaging embedded AI tools. Long-term, management expects to see this UI become more adaptive, guiding the user through an automated workflow. Guided prompting was another area of focus, and the company is building a library of prompts for end users to leverage as they perform specific tasks. Under the hood, the company intends to run its AI tools on internally trained models built on top of a third-party. CTO Aaron Harris discussed two of these tools: Sage Accounting LLM and APDoc2Vec.

 

As a reminder, Sage partnered with Amazon Web Services (AWS) over a year ago to collaborate on F&A models, and management highlighted the continued effort to build a new multitenant, dependency-based stack.

 

Long-term, TBR expects this work to be pivotal in reducing the cost of running AI workloads, while internally developed models with lower parameter counts than big-name large language models (LLMs) will further enhance cost efficiency at inference. Meanwhile, Sage is still figuring out how to monetize AI, but the industry default is to implement a tiered system. Some high-compute copilots may eventually carry usage-based fees, especially in forecasting, but for now, the priority is to show clear value and price accordingly.

 

In 2025 no conversation is complete without recognizing the platform implications of agentic workflows. Behind the scenes, Sage is preparing for an agent-first architecture by integrating emerging frameworks, such as Model Context Protocol (MCP) or Agent 2 Agent (A2A), directly into its platforms. The long-term goal is to coordinate these through super agents and plug into the broader agent ecosystem (Salesforce, Microsoft, Google), but this is still only part of the long-term road map.

 

That said, the company is building for the future, with an emphasis on data model consistency, dependency-based deployment, and orchestration layers capable of managing multi-agent chains. This is all being done with AWS in the background, keeping the platform anchored at the infrastructure layer.

Sage deepens its partner relationships

Sage’s partner and go-to-market strategy is built for focus and leverage. The company cannot cover every vertical or service need on its own, so partners are central to how it sells, delivers and scales. The revamped Sage Partner Network is tighter, with clear roles across sell, build and serve motions, and expectations tied to growth, not just activity. Multiyear vertical plans, coinvestment and execution discipline are now baseline requirements.

 

Internally, the GTM engine runs through SIGMA, which ties product planning to what the direct and partner channels sell. Sales teams are trained to package suites, identify expansion triggers, and position the platform by vertical need, rather than a feature checklist. To prepare for the platform’s evolution, Sage is already laying the groundwork for a more extensible ecosystem, including plans for an agent marketplace that would give partners a direct path into the next wave of product delivery.

Staying the course and preparing for what lies ahead

Sage’s story at its annual Analyst Summit was not necessarily one of reinvention. Land and expand has been the company’s strategy for years, and it has worked well so far. By anchoring in finance, expanding through vertical suites and operation management, and keeping partners close to the motion, Sage is executing with clarity around who it serves and how it wins. Meanwhile, the platform is evolving, AI is taking shape, and the architecture is catching up to the ambition. None of the company’s claims felt like overpromising.

 

In a market filled with transformation stories, Sage is running a disciplined play. The question is whether it can maintain that discipline as it scales and converts its product investments, especially in AI and agentic workflows, into tangible value for the customers it already knows best.

SAP Sapphire 2025: Legacy Application Leader Moving Confidently Into a Data and AI Future

Staking a claim in a best-of-suite future

At SAP Sapphire 2025, one thing became immediately clear: SAP is no longer chasing the cloud market — it is positioning itself to define it. While best-of-breed has long been the enterprise default, a growing segment of the market is leaning toward consolidation: fewer vendors, tighter integration, faster outcomes. SAP sees an opening. With its dominance as a system of record and a broad portfolio spanning platforms and line-of-business (LOB) suites, the company believes it is uniquely equipped to serve these best-of-suite buyers and made a compelling case at Sapphire that it is actively working to turn this vision into reality.

 

SAP’s messaging has focused heavily on customers already operating in the cloud, shifting attention away from the sizable portion of its base still tethered to ECC (ERP Central Component). The forward-looking emphasis may be warranted. Although cloud migrations remain a strategic priority, they have been at the center of SAP’s story for the better part of a decade. While the customer mix still skews toward legacy deployments, TBR estimates that cloud revenue accounts for more than 60% of SAP’s total corporate revenue, presenting a solid base from which to expand contract sizes.

 

In addition to migration efforts, the company has built out a suite of integration, robotic process automation and data assets — many with high attach rates — that are driving much of its commercial cloud momentum. While TBR believes SAP will continue steadily transitioning legacy customers to the cloud, its land-and-expand strategy among new logos (born-in-the-cloud, midmarket) and existing customers leaning into modernization will provide ample growth opportunities to build on top of migration-related gains. For this reason, TBR believes SAP was justified in prioritizing its platform-centric cloud strategy at Sapphire 2025. The company has built a compelling cloud business, and that road map deserves to be in the spotlight.

Building an agentic flywheel

SAP’s central metaphor this year — the “flywheel” — describes a loop in which enterprise applications feed business data into a semantic layer, which powers AI agents that act on the data and push outcomes back into the apps. Put simply: if you own the context, you control the outcome. SAP believes its depth of structured business data gives it a defensible advantage in the race toward agentic AI. Fragmented stacks, the company argues, are the Achilles’ heel of enterprise automation. SAP promises to reduce the cost and complexity of AI adoption by delivering deeply integrated, outcome-oriented capabilities across its entire suite of products.

The state of Joule as a UI for the AI era

SAP wants 2025 to be the year agents move from prototype to production. Joule remains the user interface (UI), but it is now positioned as an orchestration layer, not just a chatbot. The company showcased use cases ranging from accounts receivable prioritization to automated financial close and proactive risk flagging. These scenarios emphasized traceability. For example, each step is visible to the end user, and each recommendation is auditable. That transparency, enabled by LeanIX, signals SAP’s commitment to building enterprise-grade controls around automation.

 

Today, most of these agents are operating in relatively structured environments. Financial workflows, inventory management and procurement tasks offer well-bounded problems. The leap to agents that navigate fuzzier terrain — customer onboarding, scenario planning or partner collaboration — has not happened yet. Agentic systems will continue to be built on a use-case-by-use-case basis, which takes time. SAP is developing more tasks and, at the event, showed a demo of Joule working through a tariff shock scenario. It featured each member of a fictional C-Suite reacting in real time using embedded AI: the CFO reallocating capital, the chief revenue officer rerouting demand, the COO managing supply constraints, and the chief human resources officer rebalancing skills.

 

In TBR’s opinion, the demo felt like an oversimplification of a complex issue, but we were still impressed by the information an agent could gather and the actions it was able to execute. Obviously, agentic AI stands to be highly disruptive to SaaS workflows, and TBR believes SAP is playing the game well. Long-term, the breadth of the company’s ERP and LOB portfolios offers a massive amount of whitespace for innovation, enabling the company to continue attacking the opportunity on a use-case-by-use-case basis as it rides the wave.

Prioritizing semantic cohesion over data consolidation

SAP has spent years refining its data strategy. While Datasphere offered value in real-time processing, it was never intended to serve as a central data platform — especially with Snowflake, Databricks and Google Cloud leading in that space. The launch of Business Data Cloud (BDC) acknowledges this external reliance, advancing the same ambition Datasphere once aimed for: a harmonized, semantically enriched, agent-ready data layer.

 

BDC’s zero-copy architecture and native integrations with platforms like Databricks reflect this evolution. SAP is betting on semantic fabrics, not data lakes. Knowledge graphs across HR, finance and procurement add structure, while embedded governance ensures auditability. This plays to SAP’s strengths. The offering feels tailored to existing customers and midmarket newcomers, especially those with aggressive AI ambitions.

 

That said, harmonized data remains one of the hardest problems in enterprise IT. BDC assumes a level of data maturity that many SAP customers have not yet achieved. A large portion of the install base remains on premises, but for those already in the cloud — or willing to invest — the value proposition is becoming clearer. And SAP’s traction among net-new logos suggests the offering resonates with digital-native buyers looking to operationalize AI quickly.

Turning channel partners into strategic collaborators

The biggest partner takeaway from Sapphire was that SAP is no longer content with resell-and-implement motions. It wants deeper collaboration. The flywheel — applications, data, AI — only spins fast enough when partners are embedded into engineering, orchestration and execution. That shift has forced SAP to rearchitect how it manages partner access, tools and territory, with trust becoming a central pillar of its partner strategy.

 

SAP is also handing over the sales motions for its innovation stack. Partners now have access to the same internal tools used to build and deploy agents: Joule Studio, Prompt Optimizer, LeanIX, SAP Build and WalkMe. This is not only enablement but also an invitation to build within the stack. But access comes with expectation. These tools require fluency, not just familiarity. SAP wants to work with a deeper class of partner that can move from implementation to cocreation.

 

Equally important: territory. SAP is expanding partner-led coverage, particularly in North America and Europe. The new SAP Referral Program, scheduled to launch in 3Q25, formalizes this shift. Strategic partners will now own more of the Customer Value Journey — sales, delivery and post-sales engagement — especially in midmarket and vertical contexts.

 

Perhaps the most strategic move, though, is cultural. SAP is not just training partners; it is also increasingly transferring responsibility. KPMG is delivering structured Joule certifications. Accenture is codeveloping production agents. Capgemini is integrating Databricks into SAP’s data stack. Meanwhile, PartnerEdge, SAP’s overall partner program, is evolving to reward cloud performance, AI capability and vertical differentiation. Success in these areas will see the greatest investment and visibility from SAP.

SAP moves ahead with strategic clarity

All told, Sapphire 2025 marked a turning point, not because SAP introduced a radically new vision but because the company finally appears ready to execute on the one it has been quietly building for years. The narrative has matured, the tools are in place, and the platform is coherent. And the partners, customers and product ecosystem are starting to move together. Some heavy lifting remains, such as around migrations, data harmonization and partner fluency, but if SAP can stay focused on delivering scalable value through agentic AI, integrated data platforms and partner-enabled execution, the next chapter of the company’s growth story will look a lot less like catching up to the cloud and a lot more like leading in it.

Hitachi Digital Services Brings IT + OT + AI to Mission-critical Systems

From May 20 to 21, 2026, in Dallas, Hitachi Digital Services hosted close to 100 analysts and advisers for a two-day summit featuring Hitachi leaders, clients and technology partners. The following reflects TBR’s observations from the event, informal conversations with Hitachi leaders and clients, and TBR’s ongoing analysis of the overall IT services and professional services space.

Hitachi Digital Services perfectly positioned to take advantage of IT-OT convergence and AI adoption

Two tectonic technology shifts happening now and opening massive revenue opportunities play straight into Hitachi’s strengths, positioning Hitachi Digital Services for significant growth over the next five years. With IT and OT convergence accelerating and enterprises increasingly adopting AI at scale, Hitachi Digital Services can rely on Hitachi’s engineering DNA, deep expertise in specific industries, and firsthand experience in manufacturing as the foundation for a build-integrate-operate value proposition. The analysts and advisers event highlighted how well Hitachi Digital Services brings the entire Hitachi company to its clients, remains rooted in Hitachi’s engineering approach to everything, and maintains its focus on strategic priorities, even as the market shifts all around them.

 

Although smaller than the typical IT and professional services firms covered by TBR, Hitachi Digital Services warrants close attention in the coming years. Its distinctive business model, strategic direction and performance — particularly its combination of IT, OT, AI and domain expertise — position Hitachi Digital Services as a potential outlier in the IT services landscape.

Hitachi’s broader capabilities and assets help Hitachi Digital Services deliver unique combination of IT, OT and AI

Though only a two-day event, Hitachi included an impressive breadth of client speakers and panel participants, all from Hitachi Digital Services’ core industries, while also bringing technology partners into the discussion, occasionally simultaneously. One major telecom player described their relationship with Hitachi Digital Services as a client and with Hitachi overall as a go-to-market technology partner. An automotive client praised Hitachi Digital Services’ ability to “make our problems their problems to solve” and called out Hitachi’s “asset knowledge,” which blended IT capabilities with OT experience.

 

During onstage presentations and informal discussions with TBR, clients noted the importance of Hitachi Digital Services’ ability to seamlessly tap into the larger Hitachi organization and bring deeply technical engineering expertise to bear. Not surprisingly, clients attending the event also noted their well-established relationships with Hitachi, reinforcing Hitachi Digital Services’ messages around sustained focus on clients’ evolving issues. TBR’s assessment of Hitachi Digital Services’ client service value proposition and technology partner alliance strategy could be best summed up by a comment from a cloud vendor on stage with Hitachi Digital Services: “showing up together with Hitachi, delivering results, day after day.”

 

On Day 1 of the summit, a senior Hitachi Digital Services executive said the company maintained an “R&D and engineering first, consulting second mindset,” a sentiment echoed throughout the event. Another executive noted that innovation “starts with Hitachi engineering R&D.” Discussions around IT-OT convergence and Industry 5.0 included elements of consulting but remained rooted in engineering. For example, when discussing IT and OT convergence, Hitachi leaders described the company’s state-of-the-art railcar manufacturing plant in Maryland, a real, physical, operational test bed for digital innovations and Hitachi Digital Services’ IT + OT + AI value proposition. For Industry 5.0, Hitachi executives discussed Hitachi Digital Services’ overall AI strategy and noted the confluence of human and AI collaboration aided by AI agents.

 

In TBR’s view, the consistent undercurrent of engineering and R&D separates Hitachi Digital Services from peers in the IT services and consulting space, in part because of the credibility behind Hitachi’s “engineering first” assertion. Any company can say it has engineering capabilities; Hitachi builds railcars.

Domain expertise and organizational intelligence needed for broader AI adoption and agentic AI

Regarding AI, leaders at Hitachi Digital Services emphasized that effective AI begins with deep knowledge. Hitachi brings a strong foundation in business, operations, design, manufacturing and maintenance — an extensive base of expertise that it leverages to power its AI initiatives. In short, Hitachi Digital Services professionals argued they have more robust intelligence to feed into AI models and algorithms than most peers. Hitachi leaders also highlighted the company’s focus on organizational intelligence and its long-standing capabilities around “capturing the intelligence of complex organizations.”

 

In TBR’s view, Hitachi’s only complex organization likely provides a useful starting point for understanding the applicability of AI in organizational intelligence. Regarding agentic AI, Hitachi Digital Services leaders said AI agents must be specific to a domain, such as industrial AI, engineering AI or cybersecurity AI — an assertion that plays to Hitachi’s strengths. They also expect agentic AI adoption will accelerate in 2026, particularly around process and application modernization. One Hitachi Digital Services leader anticipated a shift from “minimum viable product” to “minimum viable agent,” a neat turn of phrase that might indicate both a shift toward a more consulting-focused mindset and internal pressures to deploy more AI-enabled assets. TBR notes that every one of Hitachi Digital Services’ competitors is working through those same shifts.

Mission-critical IT + OT + AI

In TBR’s view, Hitachi Digital Services’ ability to lean on the broader Hitachi company’s engineering expertise and real-world physical assets provides Hitachi Digital Services with clear separation from peers in the IT services space. If Hitachi Digital Services can deliver results in mission-critical — and constantly moving — settings, such as railway operations, clients in less dynamic environments can be assured of its capabilities.

Hitachi leaders noted shifts in the broader IT and OT spaces, such as the greater attention being paid to cybersecurity risks in OT and CFOs being generally more aware of the costs of funding fully separate IT and OT systems, which TBR believes may accelerate IT-OT convergence. Complemented by AI and infused with domain expertise, Hitachi Digital Services, in TBR’s view, should maintain a clearly differentiated position within the digital transformation and IT services space.

 

As a potential partner for large ISVs and cloud hyperscalers, Hitachi Digital Services’ combination of IT, OT and AI capabilities provides easily understandable differentiation. For large enterprises in Hitachi Digital Services’ target industries, Hitachi’s track record in mission-critical environments and the company’s sustained R&D and innovation create a compelling value proposition.

 

TBR will be more closely tracking Hitachi Digital Services’ evolution, particularly as IT-OT convergence accelerates and Hitachi Digital Services’ role in the broader ecosystem grows.

 

Graph: Degree of Consideration for Expanding Alliance Relationships (Source: TBR 1H25)

Salesforce Fills Its Data Governance Gap, Assembling an End-to-end Platform to Power Agentic Workflows

Informatica acquisition is big step in already robust platform strategy

As agentic AI begins to blur the lines between applications and platforms, Salesforce continues to demonstrate its intent to lead in embracing change. Ever since the company’s 2018 acquisition of MuleSoft, purchasing platform vendors has, on average, been an every-other-year occurrence — and Salesforce continued this pattern in May when it finally entered into a definitive agreement to acquire Informatica. This marked the culmination of a yearlong pursuit, as the two companies were unable to reach an agreement at a rumored $11 billion price point in April 2024.

 

Waiting paid off for Salesforce, which is now expected to close at a price point of $8 billion — $3 billion less than the previously rumored figure. While the acquisition cost decreased due to a decline in Informatica’s publicly traded stock price, the company’s IP only became more relevant to Salesforce’s data and AI strategy over the year. As Salesforce CEO Marc Benioff put it, the merger will result in “the most complete agent-ready data platform in the industry.” It is a bold statement, but there is merit to the claim — Salesforce’s cohesive platform portfolio aligns well with the company’s Agentforce ambitions.

 

Informatica is an important addition to this cohesive portfolio. Data governance and management was a lingering capability gap — one of growing importance due to the lack of standardized data formats across acquired IP and the third-party applications offered via AppExchange.

 

Connecting customers’ entire front-office estate to Agentforce will require these data formats to be standardized across these systems, compelling Salesforce to offer these capabilities natively. This will bring Salesforce a big step closer to delivering on its promise to offer customers a unified, end-to-end portfolio capable of supporting agentic workflows. Selling the story and delivering on customers’ ROI expectations will be the next step, requiring Salesforce to be an innovator in contextualized model development as the company looks to unlock new use cases.

Informatica and its place within Data Cloud

With Informatica, Salesforce will gain a data management and governance leader. TBR covers Informatica’s product development strategy in greater depth in the recent special report Data Quality & Governance Pillars, and Ecosystem-led Approach Mark Informatica’s Entry Into Agentic AI. To quickly reiterate some of the big takeaways:

  • Informatica has become more ambitious over the past year in its agentic AI pursuits. The company argues that agents that can reason and act across workflows require far more than large language models. They need clean data, reliable orchestration and deep context. Informatica views its metadata-rich foundation — anchored in CLAIRE (Cloud-scale AI-powered Real-time Intelligence Engine) — as a critical advantage in this emerging stack. With this foundation, the company will target the AI opportunity from two angles. As TBR describes in the special report linked above, the two approaches include “Informatica for GenAI, in which customers use IDMC’s [Intelligent Data Management Cloud] data management capabilities to enable enterprises’ GenAI use cases; and GenAI from Informatica, where customers leverage Informatica’s GenAI offerings.” CLAIRE Copilot entered general availability at the event, while CLAIRE GPT, which is already used by over 550 customers, continues to automate cataloging, data access and metadata queries. In addition, Informatica made its formal entry into the agentic AI space. Eight new CLAIRE Agents were announced — focused on core data management tasks like quality, ingestion, lineage and modernization — and are scheduled to enter preview in the second half of 2025.
  • Ecosystem development is a major priority for Informatica as well. The firm highlighted its expanding relationship with both Microsoft and Salesforce. Microsoft and Informatica continue to deepen their ties, most notably with the integration of Informatica Data Quality as a native application within Microsoft Fabric, while CLAIRE Copilot, built using Azure OpenAI, reinforces the alignment. Still, this development is occurring outside the Salesforce umbrella. Under Salesforce ownership, integrations between Informatica and Microsoft will likely be deprioritized. However, the impact of the acquisition on Informatica’s broader relationships with third-party SaaS vendors will depend on Salesforce’s commitment to maintaining its open ecosystem approach.

Meanwhile, the company’s standard platform assets are likely of greater relevance to Salesforce. Informatica’s IDMC will bring ingestion pipelines and ETL (Extract, Transform, Load) tooling purpose-built for hybrid environments — something Salesforce would otherwise need to build or buy elsewhere. And with Informatica’s governance and metadata engine (CLAIRE), Salesforce will gain explainability and policy controls that are critical for deploying AI in regulated environments.

 

In short, Informatica will equip Salesforce with the ingestion and governance layers Data Cloud is missing — layers that are now prerequisites for scaling AI credibly in the enterprise. Crucially, Informatica’s prebuilt integrations and policy automation tools lower the barriers to entry for enterprise teams experimenting with agents, accelerating time to value for Agentforce deployments.

Salesforce is looking for a way to accelerate top-line performance with an end-to-end portfolio capable of powering automated workflows

Informatica joins MuleSoft, Slack, Own and Tableau in the ranks of large IP purchases Salesforce has made in pursuit of its platform-meets-application layer vision. While AI is still a relatively new part of this conversation, building a platform backbone and centralizing workflow management around Slack is a concept that has been around since the business messaging application was acquired in 2020.

 

The emergence of agentic AI brings a powerful new technology to drive deeper automation, which aligns perfectly with this existing goal but also elevates the need for a robust, well-rounded data layer. Agentic outcomes are contingent upon access to quality, standardized data and metadata, and the ability to natively offer modern governance capabilities via Data Cloud will help Salesforce deliver the positive ROI it is chasing with Agentforce.

 

This moves Salesforce closer to realizing the CRM-plus-data-plus-AI vision that the company’s leadership has been championing. That said, execution remains to be seen. MuleSoft and Tableau are healthy growth drivers for Salesforce, but Data & AI makes up just $1 billion of the company’s $37.2 billion in annual recurring revenue (ARR) as of CY1Q25. Agentforce contributes only $100 million in ARR.

 

Despite Agentforce’s triple-digit year-to-year growth, this scale means the contribution is not enough to move the needle in terms of total top-line performance. TBR continues to believe this will come with time as the still-nascent platforms mature in the market, though this will require the company to sustain the platform’s triple-digit growth in the short term, followed by high-double-digit growth in the medium term.

 

Internally, Salesforce is hiring into its sales teams and setting a goal for each of its account executives to land an Agentforce deal within the year. Ecosystem development will complement this strategy as Salesforce works with IT services partners to sell Agentforce and accelerate time to value for early adopters. TBR’s conversations with ecosystem channel partners have led us to suspect that the platform’s growth will accelerate in the future, but Salesforce will need to execute in the days and months ahead.

Informatica purchase will be margin accretive in CY2026

To briefly touch on the economics of the transaction, Salesforce anticipates the acquisition will be accretive to operating margins and cash flows by the end of the second year post-integration, so TBR does not expect it to derail the progress the company has made in establishing a new margin floor.

 

Near-term impacts are expected but should roll off quickly, allowing Salesforce to continue enjoying the ample free cash flow it has earned through its disciplined approach to cost-cutting. This will provide the company with meaningful liquidity, positioning it to remain a strategic acquirer long-term, as evidenced by the less splashy Convergence acquisition, which was announced in May.

Long-term acquisition strategy

TBR expects Salesforce to focus on integrating Informatica before committing to other large purchases. That said, smaller IP- and talent-driven acquisitions in the AI space will likely ramp up, similar to the recently announced acquisition of Convergence, an AI startup focused on agentic AI research. These smaller acquisitions will play a role in Salesforce’s portfolio development strategy.

 

Moreover, the company’s operating efficiency plan more than doubled free cash flow between 2023 and 2025, providing Salesforce with the capital to fund an accelerated acquisition strategy. AI will be a primary target of these acquisitions, though vertical-specific front-office solutions could also be major considerations as the company broadens its portfolio to create new native applications to connect to Agentforce.

Salesforce’s acquisition of Informatica is not only additive but also connective

By bringing governance, metadata and ingestion capabilities in-house, Salesforce will fill critical platform gaps that limit the company’s ability to scale Agentforce and deliver on its broader AI vision. The pieces now fit: a unified stack spanning CRM, data and AI. Execution, of course, is the next hurdle.

 

Agentforce is still in the early days of development, and the company’s ability to drive meaningful ARR from the platform will depend on sustained growth, smart ecosystem plays and clear ROI for customers. But once Informatica is in place, Salesforce will be better equipped to turn its AI ambitions into enterprise outcomes — and to do so in a unified way.

PwC Japan: Trust, Unity and Focus

On April 15 and 16, PwC Japan hosted over 20 analysts, a partner and PwC executives for a day and a half summit at the company’s Technology Laboratory in Tokyo. Chief Strategy Officer and Chief Innovation Officer Kenji Katsura set the tone when opening the meeting by explaining that over the course of the event attendees would be hearing from leaders across PwC’s businesses — including audit, tax, deals and consulting — highlighting the importance of PwC’s strategy to deliver the full range of the firm’s expertise to clients. While ensuring that PwC’s services remain highly relevant to clients in Japan, the firm’s GTM strategy is closely aligned with its global network. This alignment allows PwC Japan to leverage the best practices and innovations from across the network while also contributing homegrown insights and advancements that can benefit clients worldwide.

The event included demos, presentations and one-on-one breakouts that allowed analysts to gain firsthand knowledge of PwC Japan’s evolving strategy. Demonstrating culture and an understanding of local business dynamics still provides an important nuance that allows PwC to elevate the value of its services, especially in the current uncertain geopolitical environment, as having highly country-specific capabilities and dedicated staff may become a greater asset than more explicitly globalized organizations. The following write-up summarizes TBR’s takeaways from the event and provides a glimpse into what we believe will set the next chapter of PwC’s business.

 

As a node in PwC’s ever-expanding ecosystem, Technology Laboratory brings art and science together through physical assets

Starting with a hands-on demo led by Technology Laboratory lead Shinichiro Sanji, PwC’s Technology Laboratory team demonstrated the pivot the firm has begun making in its interactions with clients, where the outcome of workshops has evolved beyond the art of the possible and into tangible solutions backed by the use of physical assets. Responsible for PwC’s tech-driven go-to-market strategy, which builds a business pipeline through cross-industry collaboration, Technology Laboratory in Tokyo is a first of its kind and enables the firm to demonstrate how it can interlink buyers beyond the traditional IT and/or finance functions into operational technology.

 

The cocreation of assets with commercial and public clients as well as academia will also help PwC demonstrate understanding and connectivity of physical AI and revamp industries where industrialized robotics are heavily in use. The Experience Center also remains an integral part of the Technology Laboratory’s success as creating future scenarios based on market and industry trends necessitates market insights delivered through multidisciplinary skills teams. With the Technology Laboratory and Experience Center housed adjacent to each other, it streamlines collaboration and interaction between teams when needed, while maintaining enough separation to enhance the unique value each team contributes during client workshops.

 

Meanwhile, as PwC strengthens its local relationships, it also continues to build nodes across its member firms that demonstrate an appetite for innovation and support client needs. For example, the firm recently announced the opening of the OT Cyber Lab in Dubai, United Arab Emirates (UAE) and the expansion of PwC’s Reinvention Lab in Australia, where it added hands-on capabilities for clients to test out advanced technologies.

We see PwC’s Business Experience Technology (BXT) and Business Model Reinvention (BMR) frameworks as the connecting glue between the various labs across PwC member firms, especially as the BXT raised the bar high for the firm’s consultants to drive outcomes. Now BMR provides a structured approach for the labs to have guided conversations with clients around their functional technology needs.

Maintaining continuity in executing against strategic priorities brings PwC closer, one member firm at a time

Following the Technology Laboratory demo, Masataka Kubota, Chair and Territory Senior Partner, PwC Japan Group CEO, kicked off the event by setting a high-level agenda centered on five megatrends: Climate Change, Tech Disruption, Demographic Shift, Fracturing World and Social Instability. These themes guided PwC’s presentations across its Assurance, Tax, Advisory and Consulting practices, with each area exploring or delving deeper into these trends.

 

PwC executives’ emphasis on reinforcing the alignment of PwC Japan and PwC Global strategies around BMR, trust, sustainability, and AI and data closely resonated with what TBR heard in a pair of similar events in the fall of 2024 in EMEA and the U.S. Kubota further amplified PwC member firms’ efforts around unity. PwC Japan’s regional customization and diversity remain vital to the firm’s success, especially as local clients lean on the company’s expertise and guidance to navigate choppy international market conditions.

 

Masaki Yasui, PwC Japan Group’s Consulting leader, continued the discussion, providing an update and insights into the firm’s line of business. Doubling in size in terms of people since late 2018, PwC Japan’s Consulting practice now houses 5,130 of the firm’s more than 12,700 employees in Japan and has grown at a double-digit rate for around 10 years — a pace that is expected to be sustained through 2030, according to Yasui. This will be an impressive achievement for PwC Japan’s consulting business, given the macroeconomic headwinds and impact on discretionary spending and consequently on consulting opportunities.

 

According to TBR’s management consulting research, consulting revenue experienced a deceleration in 2024 of 3.1% year-to-year, down from 8.1% in 2023, a trend we expect to continue throughout 2025.

Deploying tech-enabled arbitrage model will test PwC’s readiness to transform its professional services model, with consulting being most ripe for it

Leaning on the firm’s ongoing success rooted in its client-centric approach, focus on priority accounts, portfolio expansion and investment in growth areas provided a strong foundation for PwC Japan’s consulting growth. While these efforts are not unique to PwC, we believe what has helped the firm thus far is its collaborative culture across lines of business, making it appealing for partners to be part of. Growth will likely come from PwC Japan investing in new services to meet demand with BMR, strengthening its digital core, and front-office transformation serving as the lead in parts of the clients’ discussions.

 

Meanwhile, focusing on multinational clients while tapping into the power of PwC Japan’s Business Network will help the firm bridge opportunities with global clients in and outside Japan, further amplifying the need for better member firm collaboration. PwC Japan recognizes the evolving professional service market dynamics and has built out a new three-layer model for sustainable growth, with the key separation between the first and second layers being less human-dependent. Executing on the second and third layers, which are largely focused on accounting for changes in staffing and commercial models enabled through data intelligence and data monetization, will once again test PwC Japan’s collaborative culture, as often such initiatives are more challenging to be sold and managed internally than to bring externally with clients.

 

Many professional services companies grapple with similar challenges largely because of the time and materials (T&M) commercial model that consulting companies typically employ for such services. Developing data monetization and a fee-for-service type of model essentially requires companies like PwC to depart from the T&M model. We believe PwC Japan is one step ahead of many of its competitors in that regard as its consulting business has largely been driven by fixed-price engagements, making it easier to bridge into the fee-for-service setup it is looking to pursue as part of its layers two and three strategies.

 

The challenge for PwC Japan will be to educate other member firms on how to approach the opportunities, as elsewhere T&M remains the predominant commercial model. Further, PwC Japan, just like other member firms, sees managed services as part of its ability to continue to grow its business. While it is still a small portion of the current revenue composition (about 4% to 6% at the global level, according to TBR estimates), the firm continues to explore avenues to augment that gap, with leaning on its tax and cyber practice capabilities, delivery partners, and acquisitions among its top choices.

PwC’s Industrial Structural Transformation framework: A road map to the firm’s ability to execute through integrated scale and packaged services

Here lies the opportunity for PwC: developing a framework that demonstrates unity and focus while relying on its core success around trust. During the event, PwC unveiled an Industrial Structural Transformation (IST) framework, which we believe provides the necessary road map to execute against the firm’s aspiration of what the next chapter of the firm might look like. Bringing together all parts of the business — Assurance, Tax, Advisory and Consulting — and replicating them across industries while using data intelligence and a fee-for-service commercial model can help PwC build a foundation that resonates across all member firms.

 

Starting with use cases in Japan across segments like mobility provides a glimpse into how the framework can be applied as the evolution of AI extends into the physical world, and how the architecture of the entire industry can be transformed beyond the use of the software to include the business and social rules. Developing the backbone of going to market through integrated scale is the first step. Executing against it, especially bringing use cases outside of Japan to other member firms, which often deal with their local client issues, can prove to be a rewarding challenge.

Three key areas which PwC Japan discussed at length during the event — sustainability, risk management and tax — brought to light the closer collaboration and applicability of the framework at scale between the various parts of PwC’s portfolio. Sustainability and risk management bring together assurance and consulting while tax provides a conduit for closer collaboration with consulting.

 

Developing solutions that are function-specific provides the necessary horizontal connection across these areas while relying on the firm’s industry knowledge to demonstrate value and deliver outcomes. We believe a large portion of PwC’s success will come from including its technology partners at every step of development, deployment and management of its IST framework as partner feedback and knowledge management have become as critical as ever to understand how professional services companies go to market.

Application of technology solutions beyond the marketing hype elevates PwC Japan’s soberness and readiness to support the global network in handling macro disruptors

Virtually no market discussions today exclude data and AI. Takuya Fujikawa, PwC Japan Group Chief AI Officer, Data and AI Leader, took the opportunity to provide an update on the firm’s data and AI practice at the onset of the second day of the event. One important nuance struck TBR about PwC Japan’s portfolio: its applicability of AI across multiple domain areas, rather than simply drilling down on generative AI or agentic AI, which have been the predominant focus in similar settings. Adhering to open data principles while relying on its industry and functional expertise provides a strong foundation for the firm’s data and AI portfolio with examples around threat intelligence, intelligent business analytics or power market analytics highlighting various use cases.

 

With PwC US recently launching the Agentic Operating System, we expect PwC Japan to lean on these experiences and capabilities and follow suit with agentic AI solutions that meet local client needs for driving efficient operations. We expect the rollout to be bidirectional and other member firms to look to collaborate with PwC Japan on emulating its portfolio offerings. Scaling the use of such solutions will fit nicely and support PwC Japan consulting’s three-layer growth model, especially around the data intelligence and data monetization opportunities and the tech-enabled arbitrage model.

Additionally, discussions throughout the second day provided deeper insights into other strategic domains of PwC Japan’s portfolio, such as sustainability and trust, including cybersecurity. Common themes across value creation enabled through technology amplified the need for better alignment among practice areas within PwC Japan, especially as buyers’ focus in each area has shifted toward translating respective market challenges to business implications.

 

Within sustainability, PwC remains focused on creating customer value by emphasizing cross-business themes, including decarbonization and biodiversity, with most of the opportunities centered on government reporting mandates, further demonstrating the need for collaboration between audit and consulting services.

 

Withing trust, leadership highlighted cybersecurity and cyber intelligence services – both important elements of PwC Japan’s strategy – including its collaboration with delivery partners, such as with TIS Inc. for use of remote monitoring, alert responses and emergency interventions, among other security services.

Enabled by a humble and gracious culture, PwC Japan sets the bar high for what’s next in the firm’s strategy evolution

Just as technologies arm PwC consultants with tools to solve complex business challenges, the Technology Laboratory provides the enabling environment for all presentations, as regardless if executives spoke about audit and assurance, tax, advisory or consulting, they all leaned on a piece of technology that helped them connect the dots between art of the possible and the answer through the tangible.

 

As macroeconomic uncertainty persists, PwC has realized that instead of trying to fix issues outside its control, it is better to focus on its own transformation to prepare for addressing client challenges. We recognize there is still an opportunity for the firm to strengthen relationships with both internal and external stakeholders, including its alliance partners. Building a strong, common foundation for member firms is a crucial first step for integrated scaling. This involves connecting the Technology Laboratory and Experience Centers to deploy BXT and BMR at scale. Leveraging their experience with fixed-price contracts, they can then pivot toward fee-for-service models, providing the necessary frameworks for collaboration.

 

PwC Japan Group and its leadership set the bar high for its member firm counterparts in what it would take for the firms to work together more closely. Starting with better service line collaboration and portfolio offerings rollouts to establishing common KPIs and shared services are some of the prerequisites, and PwC Japan seems to have the ingredients to make it work all together. And while clients are less worried about how a firm operates and are more focused on solving their business and technology problems, working with a vendor that has its own house in order certainly makes it an easier partnership.

EY Reimagines Global Mobility: Human-centric, Tech-enabled and Business-critical

EY Global Mobility Reimagined 2025, Barcelona: Over two days in Barcelona, EY hosted more than 150 clients and a few industry analysts for its first in-person EY Global Mobility Reimagined conference since 2019. During the event, TBR spoke with EY leaders, EY technologists and EY clients from a diverse set of countries and industries. Nearly all the client attendees serve within their enterprise’s talent, mobility or human resources organizations, and a common vibe throughout the event was the changing challenges facing HR professionals. The following reflects both TBR’s observations and interactions at the event and our ongoing research and analysis of EY. Three themes emerged over the two days of the conference, both from the presentations and in discussions with EY leaders and EY clients. First, rapidly changing technology, particularly AI, permeates every aspect of mobility, but the EY leaders and conference attendees returned repeatedly to the need to keep humans at the core. Second, EY did not emphasize or sell what EY can do but rather kept the focus on clients’ problems. A Tech Connect showcase featured cool new EY software and solutions, but the conference plenaries and breakout sessions never veered into a sales pitch for EY’s solutions. Third, in discussions with EY leaders, TBR heard a clear strategy for continued rapid growth and evolving technology alliances, underpinned by a commitment to managed services.

Humans remain central to mobility, although technology can help

Mobility — moving talent around the world on short- and long-term assignments — is inherently stressful for the employee and risk-inducing for the employer, so while technologies can improve the processes and mitigate risks, the experience remains a human one. Even with generative AI (GenAI) and agentic AI, everyone strives to keep humans fully at the center. Three moments during the conference highlighted this theme.

 

During a breakout session focused on emerging technologies, EY noted that nearly all current mobility-focused technologies and platforms have been designed around corporate requirements and policies. In the near future, technologies will be designed around the employee experience. EY’s new Microsoft Teams app for mobility, described below, provides an example of that shift. Second, during a panel discussion about the ethical concerns around AI adoption, one EY leader noted that even if agentic AI and other tools replace many of mobility professionals’ day-to-day tasks, nothing can replace the human touch, especially during a stressful time like an international relocation. Once again, the technology must enhance the employee experience. Lastly, EY professionals noted during a breakout panel on immigration that employers have had a mindset shift with respect to permanent residency.

 

Previously, employees tried to shift their residency status in a foreign country without assistance from their employer, reflecting employers’ concerns that once established in a country, those employees would be inclined to stay, perhaps necessitating a split with the employer. In some countries — Saudi Arabia was cited as a prime example — annual visa and work permit renewals are both expensive and stressful. Over a long-term assignment, paying for an employee to gain permanent residency could be cost effective and, by demonstrating support and bringing corporate resources to bear, could increase employee retention. Happy employees who stay longer and build better relationships with clients lead to better returns for the company on its investment in talent and mobility.

The event offered a forum for clients to discuss challenges and how they are coping

EY sold without selling. Every session included EY partners describing the firm’s views of the challenges facing mobility professionals and HR teams overall, but, with the exception of the Tech Connect showcase, EY’s capabilities were not front and center. In the majority of sessions, EY’s capabilities and the firm’s ability to help address those challenges simply did not come up. The EY partners focused on setting the overall parameters of the discussion, providing context around the challenges clients face, and then allowing those clients to engage with EY and with other clients about how they are coping.

 

In TBR’s view, this approach separates EY from peers while also reflecting the ethos TBR has observed at other EY events, such as the Strategic Growth Forum. EY provides clients a comfortable space to talk about issues and commiserate, without a hard EY sell. One attendee told TBR the conference made him feel better because his problems were not nearly as dire as the other professionals he spoke with during the event.

 

Navigating GenAI: Insights, Strategies and Opportunities for 2025 — Watch “2025 Predictions for GenAI” on demand now

Mobility practice serves as the glue across EY globally

In sidebar discussions with TBR, EY leaders’ comments reinforced two trends about Mobility — and the firm’s People Advisory Services – overall. First, People Advisory Services is growing ahead of the overall firm. The practice has been investing in managed services capabilities and scale, with an appreciation that noncommoditized managed services will be a significant component of People Advisory Services revenues. Second, Mobility remains an essential part of the glue that keeps EY operating as a global firm serving global clients.

 

In addition, during the event TBR noted a clear emphasis on Microsoft as a strategic partner, but EY noted expectations that other technology alliance partners, including ServiceNow, will become increasingly strategic to Mobility. Regarding the third point, Mobility services have challenged other Big Four firms, in part because — as practiced by EY — the burdens include forced cross-border coordination by and shared resources from separate member firms, managing a plethora of niche providers and technology partners, and deploying a software business model.

 

Countering those burdens, Mobility can serve as a centripetal force, helping align EY’s Tax, Audit and Advisory practices and giving additional weight to the firm’s global capabilities (and leadership, notably). As a complementary service to consulting or tax, Mobility advances EY’s client retention strategies, particularly with its largest clients. Internal benefits and external rewards. Win- win.

Emerging technologies begin to permeate HR

AI could not be ignored, in part because the notion of agentic AI-related disruption was an underlying current throughout the event. While not diminishing the challenges of adopting emerging technologies, EY professionals repeatedly stressed the need to adopt soon, smartly and with a long-term plan in place. In a breakout session, EY professionals used a now/next framework to describe a few trends in emerging technologies (including the corporate requirements and employee experience described above).

 

Currently, HR professionals and employees must wrangle with multiple technologies and platforms to execute on mobility challenges; in the near term, everyone will enjoy streamlined technologies with cohesive data-sharing strategies. Today, HR professionals rely on dashboards to provide analytics on mobility and other People Advisory Services issues, but soon predictive AI-driven analytics will provide insights and quicker decision making, with fewer (or perhaps no) dashboards.

 

Notably, when surveyed during the breakout session, the majority of HR professionals in attendance opted for “streamlined technologies” as their top priority. In a separate session, client attendees said their greatest expectation from AI would be data analytics and enhanced reporting.

A few other technology-centric comments and observations from the event:

  • EY partners said Mobility professionals did not need to wait for the next GenAI update or release — the technology needed is here and can be applied now.
  • Previous notions about data complexity may be outdated as the technology exists now to handle that complexity — HR professionals should focus on what they want to accomplish, not whether their data is perfect. (Side note: in the same discussion, EY partners observed that the biggest roadblock to adoption remains the availability of quality data.)
  • The ethics around GenAI remain … murky. EY partners noted the environmental impact of energy-hungry data centers and suggested a gap exists between innovation and accountability, eventually cautioning for a go-slower approach to AI adoption.

Overall, technology played across every aspect of Mobility with the common theme around enhancing the employee experience and measuring how EY’s Mobility practice can benefit a company’s strategy and employee retention, and even improve relationships with clients. In short: use technology wisely, with help from EY.

EY integrates mobility management into Microsoft Teams

The Tech Connect showcase included nine solutions, most notably EY Mobility Pathway, a corporate mobility management tool; EY Mobility Carbon Tracker, a customizable tool for scenario planning and carbon footprint measuring; and new Microsoft Teams app for mobility, a seat-based SaaS offering deployed as an application on Microsoft Teams. The last one stole the show. Employees do not need to log on to another platform, remember another password or navigate an unfamiliar app, but rather add the new Microsoft Teams app for mobility app to their Teams experience. The software can be configured to clients’ specific mobility needs, such as shipping dates, travel, housing, tax and other elements of the international relocation journey.

 

The new Microsoft Teams app for mobility looks and feels like a Teams app, has all the employer data, and seamlessly — as the employee experiences it — pulls in data and information from the third-party providers the employer uses, such as shipping companies or short-term housing agents. EY partners explained that new Microsoft Teams app for mobility is currently live with a few clients, and will go live soon across more of EY.

 

In TBR’s view, EY made a significant strategic decision in embedding new Microsoft Teams app for mobility into Microsoft Teams and not creating a separate employee-centric dashboard. This keeps employees in an environment they are already comfortable working in and avoids additional stress during a difficult time. EY’s commercial model for new Microsoft Teams app for mobility requires the firm to invest in software support and maintenance capabilities, but feeds into the firm’s overall managed services play.

Immigration rises to C-level topic

An immigration session resonated with TBR, in part because the TBR principal analyst in attendance once stamped visas at a U.S. embassy, but also because of the political issues that were openly discussed. EY partners noted that 2024 was a “super year” for elections globally, and immigration issues featured prominently in election politics in many countries. Extrapolating to global enterprises, EY partners made a convincing case that immigration has become a boardroom issue. EY’s Batia Stein and Chris Gordon noted that chief human resources officers (CHROs) and other executives surveyed by EY said the top option for solving talent gaps is moving talent where it is needed, no matter where on the globe — so, mobility.

 

As described above, assisting with permanent residency can alleviate some employee stress and enhance client and employee retention. In addition, using technology to enhance the employee mobility experience is not simply the right thing to do for employees; EY also believes the Mobility practice can be a business driver. And at a time when compliance issues have become more frequent and fraught, exacerbated by immigration raids and joint immigration and tax audits, Mobility can be a business driver for EY, too.

People Advisory Services global infinity loop reflects EY’s approach to clients’ issues

EY has a visual of its People Advisory Services Tax practice that features an infinity loop with People Advisory Services on one side and People Managed Services on the other, with all the related offerings creating an endless cycle of services, surrounded by EY’s other practices and offerings, such as Strategy & Transactions and Sustainability. The infinity loop helps understand EY’s positioning of its services and, perhaps more importantly, reflects EY’s understanding of its clients’ needs and challenges.

 

Companies keep recruiting, hiring, paying, rewarding, moving, repatriating, retiring and hiring in an endless loop, and EY has capabilities — including consulting, tax and software — that can accelerate movement around that endless loop. EY did not need to say that at the Global Mobility Reimagined conference, as clients understood it already. EY also has a stated ambition to grow People Advisory Services to more than $3 billion by 2031. Absent the worst possible global political and economic scenarios, including a drastic curtailment of global mobility, TBR believes that ambition is perhaps a bit too modest.

Data Quality & Governance Pillars, and Ecosystem-led Approach Mark Informatica’s Entry Into Agentic AI

Building on last year’s theme of modernization, where Informatica highlighted innovations to fast-track migrations from legacy PowerCenter and Master Data Management (MDM) to Informatica Data Management Cloud (IDMC) in the cloud — accounting for over half of Informatica’s business — Informatica World 2025 in Las Vegas was all about agentic AI, and the crucial, yet still sometimes overlooked, role of data. With its ability to apply reasoning to handle more complex, multipart workflows, agentic AI has rapidly emerged as AI’s next frontier. Although agentic AI promises increased productivity, AI agents require a few key elements, including orchestration, a vast knowledge base and governance. With the wealth of metadata in CLAIRE, the AI engine powering IDMC, it was only a matter of time until Informatica used some of its core differentiators to push into the agentic AI space to not only craft a future for more autonomous data management but also to give customers the tools needed to build and orchestrate their own agents.

 

Informatica enters agentic AI race

Informatica employs two strategies relative to generative AI (GenAI): Informatica for GenAI, in which customers use IDMC’s data management capabilities to enable enterprises’ GenAI use cases; and GenAI from Informatica, where customers leverage Informatica’s GenAI offerings. Those products include CLAIRE Copilot, which entered general availability at the event, and CLAIRE GPT, which is used by over 550 clients to streamline tasks within IDMC such as pulling datasets and interacting with catalogs. On the annual conference’s 25th anniversary, Informatica formally entered the agentic AI space with a similar approach, giving customers the ability to consume AI agents within IDMC for more autonomous data management and a tool for letting customers build, manage and orchestrate their own AI agents.

  • CLAIRE Agents: As part of the GenAI from Informatica strategy, Informatica introduced eight new agents to support tasks across the data life cycle. The agents, which are expected to begin preview in 2H25, are Data Quality, Data Discovery, Data Lineage, Data Ingestion, ETL [Extract, Transform Load], Modernization, Product Experience and Data Exploration. Architecturally speaking, these agents will round out the IDMC platform, sitting above the metadata system of intelligence, with CLAIRE Copilot and GPT acting as user experience (UX) overlays, where customers can interact with these agents.

Nearly every facet of IT has emerged as a prominent use case for GenAI, including data management, and customers are looking for ways to streamline more system-level tasks. Provided customers are prepared to move from manual — or even predictive and conversational AI engineering — to agentic AI, these new data management agents can help absorb a lot of back-end data management tasks, including developing the data pipelines and reducing some of the burden on the user, whose primary focus now becomes managing the agents.

 

  • AI Agent Engineering: In agentic AI, the hyperscalers and SaaS vendors are racing to position as the AI control tower. As more vendors push into the PaaS space and the resulting AI agents convolute the applications layer, the question becomes, “Which set of vendors are positioned to abstract that complexity in a governed way?” At the event, Informatica entered the space with the introduction of AI Agent Engineering, a new tool to help users build their own agents using the popular low-code/no-code drag-and-drop experience. Additionally, AI Agent Hub, which acts as a marketplace within AI Agent Engineering, helps users find, manage and connect these agents, including not just Informatica’s CLAIRE agents but also the tools customers are already using to build agents, such as Amazon Bedrock, Azure OpenAI, Google Vertex and Salesforce’s Agentforce. When it comes to building and orchestrating AI agents, customers have a range of options, but one of the compelling things about Informatica entering this space is its ability to provide the federated governance and access controls around these agents without disrupting existing workflows. Governance remains one of the leading barriers to GenAI adoption, and while some overlap will always exist between the hyperscalers and data ISVs, the hyperscalers recognize Informatica’s reputation for helping customers build trust in their data and ability to apply that trust in a vendor-agnostic way.

 

Unlock the potential of generative AI (GenAI) in your enterprise by understanding the critical role of unstructured data management – Watch The Emerging Data Ecosystem on demand now

Ecosystem developments

As we often discuss, Informatica maintains a high degree of neutrality and can effectively work across a range of technology partners without introducing significant overlap. Maintaining its commitment to working within the technology ecosystem, Informatica announced new product integrations across its technology partners, including the following highlights.

  • Microsoft: Microsoft’s play at the PaaS layer (e.g., Synapse, Power Platform) and ability to extend the Dynamics 365 data model to enable Customer 360 analytics make it an invaluable, somewhat unique partner to Informatica. Reaffirming Informatica’s commitment to Microsoft, CLAIRE Copilot was built using the Azure OpenAI Service. The launch of Microsoft Fabric last year seemed to mark a turning point in the alliance, as Informatica was granted status as an early design partner for Fabric, which has amassed 21,000 paid customers in the span of 18 months. Essentially, this status allows Informatica to make its Data Quality tool available as a native service, so customers can profile and assess data using Informatica as it gets ingested into Microsoft Fabric via the OneLake repository in real time. At the event, Informatica made Data Quality available (in public preview) as its own Fabric application. In addition, as part of its commitment to staying relevant within the Microsoft Fabric ecosystem, Informatica will start supporting Apache Iceberg in Microsoft Fabric, which is important as Microsoft looks to cement its commitment to open standards. These developments come as part of a new strategic agreement between Informatica and Microsoft, which implies not just a focus on R&D but also investment in the joint go-to-market approach. Having Informatica exist as a first-class citizen within the Microsoft stack could make the case for customers to explore other components of IDMC, creating deeper synergies with services partners like KPMG that use Informatica and Microsoft Fabric, both internally and externally for data modernization and transformation.
  • Salesforce*: Though Informatica and Salesforce technically had a preexisting alliance, the partnership was formalized at Informatica World 2025 with the announcement that IDMC will be integrated with Salesforce’s Agentforce. Specifically, Informatica plans to deliver MDM SaaS with Agentforce, effectively putting the 360-degree wrapper around agents that customers build in Salesforce’s platform. As previously mentioned, customers will also be able consume Agentforce via Informatica AI Agent Engineering upon availability later this year. Salesforce sees Informatica as a key player in the market and is looking to strengthen its play in data management and governance in accordance with Agentforce, so this partnership is a win for Salesforce. In turn, Informatica seems to recognize the role Agentforce will play in the AI ecosystem for sales and service use cases, and it will be interesting to see how this partnership progresses and if Salesforce ends up joining Informatica’s seven other, more established technology partners.

On the services side, Informatica continues to cement its value across nine core global systems integrator (GSI) partners, which collectively staff 30,000 Informatica professionals. In 2024, Informatica earned 15,000 certifications, up over 20% year-to-year. Vendor sentiment and our own conversations with enterprise IT decision makers suggest that for AI to effectively scale, data needs to be in the cloud. As such, modernization will continue to be a big focus for Informatica and its partners through 2025. This includes AI-powered modernization and potentially using the new CLAIRE Agents, specifically Modernization, to help migrate on-premises data to IDMC. When it comes to agentic AI, Informatica’s new innovations should open new doors for services partners to not only modernize data management tasks ahead of GenAI deployments but also help clients create new custom agents (using AI Agent Engineering), including those tailored to certain industries, and make them relevant within existing workflows.

 

Between the technology partners and GSIs, Informatica works with a robust ecosystem of partners in a triparty approach, where resources from a hyperscaler, GSI and Informatica are brought together to help customers modernize their data faster and, by default, hasten AI’s time to value. When we survey and speak to alliance decision makers at IT services firms, data management comes up as one of the top areas for partner-led growth, signaling to the ecosystem that they will continue to invest in resources to guide conversations with customers with the technology maturity to address the data foundations ahead of GenAI.

Conclusion

Agentic AI has a lot of promise but also some challenges. The proliferation of AI agents will create more best-of-breed complexity — which we know customers are trying to move away from — and heighten concerns around data privacy and governance. Informatica’s move into the agentic AI space with both CLAIRE Agents and AI Agent Engineering is certainly in step with the market; we all know AI agents do not exist in silos, so Informatica’s ability to work within its ecosystem of tech partners and connect agents in a vendor-neutral way is particularly compelling. Meanwhile, Informatica’s robust engineering relationships with the hyperscalers, as evidenced by Informatica’s Data Quality integration with Microsoft Fabric, will continue to elevate its standing with the big GSIs and foster a compelling triparty alliance approach focused on helping customers get their data ready for AI.

*After Informatica World, on May 27, Informatica entered into an agreement to be acquired by Salesforce. Informatica will continue to operate as a stand-alone entity until the acquisition closes, likely in Salesforce’s FY27. Please see TBR’s Salesforce coverage for further insights.

Oracle Redefines Data Intelligence in Full-stack Approach

Oracle pivots around data intelligence, owing to its full-stack approach

Oracle has long offered a modern analytics stack tailored to multiple personas and workloads, such as through the Oracle Analytics Cloud (OAC) and Autonomous Data Warehouse (ADW), coupled with the operational data — where the true value exists — in Oracle’s Fusion, NetSuite and Industry Applications. But the 2023 launch of Fusion Data Intelligence (FDI) marked a major shift in Oracle’s analytics strategy and early vision for data intelligence, where data is used not only for static reporting of one-and-done use cases but also for continual predictive insights made possible by AI.
 
As a reminder, Oracle delivers FDI as a single-SKU application, an approach not all peers take, so Fusion customers are connected to their data through the quarterly Fusion updates, potentially causing minimal disruption to the workflow, which is an important enabler of data intelligence. From a technical perspective, FDI also comes with prebuilt AI and machine learning (ML) models, data science capabilities, and even its own separate set of intelligent applications (e.g., Supply Chain Command Center) that are persona-specific, allowing customers to act on a particular use case without leaving FDI, which is now the fastest-growing application across the entire Oracle corporation.
 

Oracle Data Intelligence (Source: TBR)


 
Importantly though, effective data intelligence is about not only the application but also the underlying architecture and whether it can effectively support structured and unstructured data for complex analytics use cases. We all know Oracle Cloud Infrastructure (OCI) has become a critical component of Oracle’s business, and because of the infrastructure layer, Oracle has a top-down advantage that many other players cannot provide.
 
The 2024 launch of Intelligent Data Lake reaffirmed how Oracle wants to further bridge the gap between applications and infrastructure, with an architecture that integrates with ADW and OAC. Essentially, Intelligent Data Lake is a reworking of existing OCI capabilities, such as cataloging and integration, to create a single abstraction layer that, in true data lake fashion, allows customers to query data on object storage, with support for popular data format frameworks including Apache Iceberg.
 
Many peers have been moving more squarely into the data lake space to make it easier for customers to build AI applications on top of a single copy of data. But in the case of Oracle, Intelligent Data Lake serves as the glue between the infrastructure and applications. With Intelligent Data Lake, Oracle has essentially redelivered its analytics tools as part of the Data Intelligence Platform, offering another key layer that could make the case for best-of-breed customers to consolidate more of their data and business intelligence estates on Oracle.
 
Regarding those application components, customers can leverage FDI as a single product, but this extends to NetSuite, Oracle Health (Cerner) and Industry Applications. For instance, last year Oracle launched Energy & Water Data Intelligence, leveraging insights from Industry Applications like Oracle Utilities Customer Cloud Service. More notably, as Oracle pivots around data intelligence, the company is taking steps to help customers access non-Oracle data sources.
 
For instance, last year Oracle launched a native Salesforce integration with FDI so customers can combine their CRM and Fusion data within the lakehouse architecture. This means Oracle customers can access Salesforce data with FDI the same way they can with Fusion. It will be interesting to see if Oracle more aggressively expands the data ecosystem in the future, particularly within the back office, to deliver FDI’s value to those outside the Oracle SaaS base.

FDI aligns with partners’ digital transformation ambitions

One of the compelling things about Oracle’s full-stack approach to analytics, from infrastructure up to applications, is that it prevents Oracle from getting caught up in a traditional BI RFP and instead enables the company to sell Oracle Data Intelligence as part of a broader enterprise transformation, which aligns with global systems integrators’ (GSIs) business models. Today, most of GSIs’ Oracle business comes from the applications side, and doing a Fusion SaaS implementation (e.g., ERP, HCM [human capital management]) and then introducing FDI to break down integration barriers and ultimately make those Fusion apps more “intelligent” appears to be a common motion.
 
In some cases, FDI is also displacing different components of a build-your-own data strategy. For example, we recently heard a compelling example from Infosys, which modernized a customer’s analytics stack by migrating from Snowflake and Informatica to FDI, which was then integrated with external systems, including NetSuite. In a scenario like this, it is clear that having a lot of data in the Oracle ecosystem can influence a customer’s decision to consolidate on FDI, but it also speaks to the role Oracle plays on the infrastructure side, as FDI address not only the analytics pieces but also the underlying data tasks, including the data pipelines, and absorb system-level tasks like ETL (Extract, Transform, Load).
 
Oracle’s full-stack approach to analytics makes a compelling case for consolidation, helping partners create value by eliminating disparate integrations and unlocking ROI. This is particularly true for partners that are perhaps willing to abandon the typical tech-agnostic approach and recommend Oracle as the primary choice from a data and analytics perspective. If Oracle engages a broader external data ecosystem in the future, as discussed above, partners will need to make sure they look beyond the applications layer and leverage Oracle’s broad PaaS and IaaS capabilities for custom development use cases.

Fujitsu Expands AI Strategy in Europe, Emphasizing Collaboration, Compliance and Customization

‘AI can be a knowledge management accelerator, but only if well-fed by an enterprise’s own data’

In February TBR met with two AI leaders in Fujitsu’s European Platform Business to better understand the company’s approach to the AI market, its evolving AI capabilities and offerings, and what we can expect as 2025 unfolds. Maria Levina, AI Business analyst, and Karl Hausdorf, head of AI Business, gave a detailed presentation focused primarily on the European market. The following reflects both that briefing and TBR’s ongoing research and analysis of Fujitsu, the company’s partners and peers, and the overall AI landscape.
 
One highlight that illustrates many facets of Fujitsu’s approach to AI was Levina and Hausdorf’s description of Fujitsu’s customers’ choices between “bring your own data” and “bring your own AI.” The first allows more AI-mature customers to bring their data into a Fujitsu-provided on-premises solution with full support for scaling, maintaining hardware and software, and updating, as needed.
 
The second allows customers to run their own AI stack on a Fujitsu “validated and optimized” platform, developed and maintained by Fujitsu and select technology partners. Critical, in TBR’s view, is Fujitsu’s positioning of these options as responsive to clients’ needs, as determined by all AI stakeholders within an enterprise, including IT, AI and business leaders.
 
Levina and Hausdorf explained, “Together, with our ecosystem of partners, we’re committed to unlock the potential of generative AI for our clients” through “on-premises data sovereign, sustainable private GPT and AI solutions,” focused on “rapid ROI.” Fujitsu is not approaching clients with a technology solution, but rather with options on how to address and solve business problems. As the Fujitsu team said, “Understand the why, know the what, and co-create the how.” The Fujitsu team also noted that the company’s industry expertise resides in the processes and workflows unique and/or critical to an industry.
 

‘Maintaining control of data means owning your AI in the future’

Before diving into Fujitsu’s AI offerings, the Fujitsu team laid out their understanding of the European market, sharing data the company collected around AI adoption, use of AI platforms, and barriers to growth (in Fujitsu’s phrasing, “progress-limiting factors,” which is perhaps a more positive spin on the usual list of barriers and challenges). Fujitsu surveyed or spoke with 400 data and IT professionals across six European countries, and the results indicated that overcoming legacy mindsets continues to be a major impediment to adopting and harnessing the value of AI.
 
TBR’s November 2024 Voice of the Customer Research similarly noted challenges in Europe with “the lack of engagement from employees who are being asked to change.” The Fujitsu team noted that change management, therefore, had to involve all AI stakeholders, including “IT people, business people and AI people” within an enterprise.
 
In TBR’s experience, IT services companies and consultancies continue to find new constituents for change management at their clients as the promise — and disruption — of AI becomes more widespread, reinforcing Fujitsu’s strategy of bringing change management to all AI stakeholders. Lastly, the Fujitsu team noted that within European clients, expectations around AI have heightened, especially as AI initiatives have launched across multiple business units. Again, Fujitsu’s research and TBR’s Voice of the Customer Research align around ROI expectations as AI matures.
 
The Fujitsu team introduced their AI platform by delineating the key performance indicators they believe a successful platform must have: scaling, performance and speed, simplicity, energy efficiency, AI services in data centers, and GPUs.
 
Although TBR is not in a position to evaluate the technological strengths, completeness or complexity of Fujitsu’s platform, the expansive KPIs indicate Fujitsu has considered not only the IT needs behind an AI deployment but also the larger business factors, particularly the financial impacts. Levina and Hausdorf then dove into the details, including the two customer options described above (bring your own data and bring your own AI). They discussed how Fujitsu offers consulting around the technical and business implications of AI platforms and solutions, including an “AI Test Drive,” which allows clients to test AI solutions before investing in new technologies, large language models (LLMs) or other AI components.
 
Notably for TBR, Fujitsu’s presentation extensively highlighted the company’s AI alliance partners, including Intel, NVIDIA, AMD and NetApp, as well as a slew of LLM providers, demonstrating an appreciation for the collaborative and ecosystem-dependent nature of AI at the enterprise level. The Fujitsu team also stressed the European nature of its AI strategy and platform.
 
European clients, Fujitsu noted, had specific requirements related to the European Union’s (EU) General Data Protection Regulation and the EU AI Act, as well as a preference for on-premises solutions. Some of the use cases Levina and Hausdorf described included a law firm using Fujitsu-enabled AI solutions to analyze case data, contracts, corporate and public legal documents, and multiple deployments of Fujitsu-enabled private GPTs.

Additional observations

  • Fujitsu remains focused on targeting customers already aligned with the company around AI, a strategy that TBR believes speeds ROI and increases client retention.
  • In contrast to some peers in the IT services market, Fujitsu has capabilities across the entire AI technology stack — hardware, software and service — which Levina and Hausdorf called “highly appealing,” especially to European clients.
  • Levina and Hausdorf made two comments that, in TBR’s view, neatly sum up AI at present: “AI can be a knowledge management accelerator, but only if well-fed by an enterprise’s own data” and “maintaining control of data means owning your AI in the future.”

Fujitsu’s AI prowess makes it an invaluable partner

TBR has reported extensively on Fujitsu’s evolving AI capabilities and offerings, noting in a special report in May 2024: “TBR appreciates that Fujitsu’s combination of compute power and proven AI expertise makes the company a significant competitor and/or alliance partner for nearly every player fighting to turn GenAI [generative AI] hype into revenue.
 
“Second, Fujitsu’s vision of ‘converging technologies’ aligns exceptionally well with the more tectonic trends TBR has been observing in the technology space, indicating that Fujitsu’s market positioning is more strategic than transactional or opportunistic.” Add in Fujitsu’s deepening experience in delivering AI solutions to AI clients, and TBR continues to see tremendous near-term opportunity and growth for Fujitsu and its ecosystem partners.

Google Cloud Cements Values of Enterprise Readiness, Full-stack AI and Hybrid Cloud at Next 2025

In April Google Cloud hosted its annual Next event to showcase new innovations in AI. Staying true to the theme of “A New Way to Cloud,” Google focused on AI, including how AI can integrate with enterprises’ existing tech landscape, with partners playing the role of orchestrator. After Google CEO Sundar Pichai spoke about the company’s achievements around Gemini, which is integral to Google Cloud’s strategy, Google Cloud CEO Thomas Kurian highlighted the business’s three key attributes: optimized for AI; open and multicloud; and enterprise-ready. Additionally, Google Cloud announced a series of new innovations that highlight how the company is trying to execute on these three areas to be the leader in modern AI development.

Google takes an end-to-end approach to AI

When discussing Google Cloud’s three key attributes, Kurian first highlighted how Google Cloud Platform (GCP) is optimized for AI. Based on our own conversations with IT decision makers, this claim is valid: many customers enlist GCP services purely for functional purposes, as they believe they cannot obtain the same performance with another vendor. This is particularly true of BigQuery, for large-scale data processing and analytics, and increasingly Vertex AI, which now supports over 200 curated foundation models for developers.
 
Within this set of models is, of course, Gemini, Google’s own suite of models, including the new Gemini 2.5 Pro, which has a context window of 1 million tokens and is reportedly now capable of handling advanced reasoning. To be fair, Google still faces stiff competition from other frontier model providers, but Google’s years of AI research through DeepMind and its ability to have models grounded in popular apps like Google Maps, not to mention Google Search, will remain among its key differentiators.
 
With that said, the AI software stack is only as effective as the hardware it runs on. That is why Google has been making some advances in its own custom AI accelerators, and at the event, Google reaffirmed its plans to invest $75 billion in total capex for 2025, despite the current macroeconomic challenges. A large piece of this investment will likely focus on paying for the ramp-up of Google’s sixth-generation TPU (Tensor Processing Unit) — Trillium — which became generally available to Google Cloud customers in December. Additionally, Google is making some big bets on the next wave of AI usage: inference.
 
At the event, Google introduced its seventh-generation TPU, dubbed Ironwood, which reportedly scales up to 9,216 liquid cooling chips linked through a high-powered networking layer, to support the compute-intensive requirements of inference workloads, including proactive AI agents. In 2024 there was a 3x increase in the number of collective TPU and GPU hours consumed by GCP customers, and while this was likely off a small number of hours to begin with, it is clear that customers’ needs and expectations around AI are increasing. These investments in AI hardware help round out key areas of Google’s AI portfolio ― beyond just the developer tools and proprietary Gemini models ― as part of a cohesive, end-to-end approach.
 

Watch now: Cloud market growth will slow in 2025, but will activity follow? Deep dive into generative AI’s impact on the cloud market in 2025 in the below TBR Insights Live session

 

Recognizing the rise of AI inference, Google Cloud reinforces longtime company values of openness and hybrid cloud

With its ties to Kubernetes and multicloud editions of key services like BigQuery and AlloyDB, Google Cloud has long positioned itself as a more open cloud compared to its competitors. However, in recent quarters, the company has seemed to hone this focus more closely, particularly with GDC (Google Distributed Cloud), which is essentially a manifestation of Anthos, Google’s Kubernetes-based control plane that can run in any environment, including at the edge. GDC has been the source of some big wins recently for Google Cloud, including with McDonald’s, which is deploying GDC to thousands of restaurant locations, as well as several international governments running GDC as air-gapped deployments.
 
At Next 2025, Google announced it is making Gemini available on GDC as part of a vision to bring AI to environments outside the central cloud. In our view, this announcement is extremely telling of Google Cloud’s plans to capture the inference opportunity. Per our best estimate, roughly 85% of AI’s usage right now is focused on training, with just 15% in inference, but the inverse could be true in the not-too-distant future. Not only that, but inference will also likely happen in distributed locations for purposes of latency and scale. Letting customers take advantage of Gemini to build applications on GDC — powered by NVIDIA Blackwell GPUs — on premises or at the edge certainly aligns with market trends and will help Google Cloud ensure its services play a role in customers’ AI inference workloads regardless of where they are run.

Boosting enterprise mindshare with security, interoperability and Google-quality search

Kurian mentioned that customers leverage Google Cloud because it is enterprise-ready. In our research, we have found that while Google Cloud is highly compelling for AI and analytics workloads, customers believe the company lacks enterprise-grade capabilities, particularly when compared to Microsoft and Amazon Web Services (AWS). But we believe this perception is changing, and Google Cloud is recognizing that to gain mindshare in the enterprise space, it needs to lead with assets that will work well with customers’ existing IT estates and do so in a secure way. This is why the pending acquisition of Wiz is so important. As highlighted in a recent TBR special report, core Wiz attributes include not only being born in the cloud and able to handle security in a modern way but also connecting to all the leading hyperscalers, as well as legacy infrastructure, such as VMware.
 
Google Cloud has been very clear that it will not disrupt the company’s multihybrid capability. In fact, Google Cloud wants to integrate this value proposition, which suggests Google recognizes its place in the cloud market and the fragmented reality of large enterprises’ IT estates. Onboarding Wiz, which is used by roughly half of the Fortune 500, as a hybrid-multicloud solution could play a sizable role in helping Google Cloud assert itself in more enterprise scenarios. In the meantime, Google Cloud is taking steps to unify disparate assets in the security portfolio.
 
At Next 2025, Google Cloud launched Google Unified Security, which effectively brings Google Threat Intelligence, Security Operations, Security Command Center, Chrome Enterprise and Mandiant into a single platform. By delivering more integrated product experiences, Google helps address clients’ growing preference for “one hand to shake” when it comes to security and lays a more robust foundation for security agents powered by Gemini, such as the alert triage agent within Google Security Operations and the malware analysis agent in Google Threat Intelligence to help determine if code is safe or harmful.
 
One of the other compelling aspects of Google’s enterprise strategy is Agentspace. Launched last year, Agentspace acts as a hub for AI agents that uses Gemini’s multimodal search capabilities to pull information from different storage applications (e.g., Google Drive, Box, SharePoint) and automate common productivity tasks like crafting emails and scheduling meetings. At the event, Google announced that Agentspace is integrated with Chrome, allowing Agentspace users to ask questions about their existing data directly through a search in Chrome. This is another clear example of where Google’s search capabilities come into play and is telling of how Google plans to use Agentspace to democratize agentic AI within the enterprise.

Training and more sales alignment are at the forefront of Google Cloud’s partner priorities

Google Cloud has long maintained a partner-first approach. Attaching partner services on virtually all deals; taking an industry-first approach to AI, particularly in retail and healthcare; and driving more ISV coselling via the Google Cloud Marketplace are a few examples. At Next 2025, Google continued to reaffirm its commitment to partners, implying there will be more alignment between field sales and partners, to ensure customers are matched with the right ISV or global systems integrator (GSI), a strategy many other cloud providers have tried to employ.
 
When it comes to the crucial aspect of training, partners clearly see the role Google Cloud plays in AI, and some of the company’s largest services partners, including Accenture, Cognizant, Capgemini, PwC, Deloitte, KPMG, McKinsey & Co., Kyndryl and HCLTech, have collectively committed to training 200,000 individuals on Google Cloud’s AI technology. Google has invested $100 million in partner training over the past four years, and as highlighted in TBR’s Voice of the Partner research, one of the leading criteria services vendors look for in a cloud partner is the willingness to invest in training and developing certified resources.

Google Cloud wants partners to be the AI agent orchestrators

As previously mentioned, Vertex AI is a key component of Google Cloud’s AI software stack. At Next 2025, Google Cloud introduced a new feature in Vertex called the Agent Development Kit, which is an open-source framework for building multistep agents. Google Cloud is taking steps to ensure these agents can be seamlessly connected regardless of the underlying framework, such as launching Agent2Agent (A2A), which is an open protocol, similar to protocols introduced by model providers like Anthropic.
 
Nearly all of the previously mentioned GSIs, in addition to Boston Consulting Group (BCG), Tata Consultancy Services (TCS) and Wipro, have contributed to the protocol and will be supporting implementations. This broad participation underscores the recognition that AI agents will have a substantial impact on the ecosystem.
 
New use cases will continue to emerge where agents are interacting with one another, not only internally but also across third-party systems and vendors. With the launch of the Agent Development Kit and the related protocol, Google Cloud seems to recognize where agentic AI is headed, and for Google Cloud’s alliance partners, this is an opportune time to ensure they have a solid understanding of multiparty alliance structures and are positioned to scale beyond one-to-one partnerships.

Final thoughts

At Next 2025, Google reportedly announced over 200 new innovations and features, but developments in high-powered compute, hybrid cloud and security, in addition to ongoing support for partners, are particularly telling of the company’s plans to capture more AI workloads within the large enterprise. Taking an end-to-end approach to AI, from custom accelerators to a diverse developer stack that will let customers build their own AI agents for autonomous work, is how Google Cloud aims to protect its already strong position in the market and help lead the shift toward AI inferencing.
 
At the same time, Google Cloud appears to recognize its No. 3 position in the cloud market, significantly lagging behind AWS and Microsoft, which are getting closer to each other in IaaS & PaaS revenue. As such, taking a more active stance on interoperability to ensure AI can work within a customer’s existing IT estate, and guaranteeing partners that have the enterprise relationships are the ones to orchestrate that AI, will help Google Cloud chart its path forward.