Walmart cracks enterprise AI at scale: Thousands of use cases, one framework

Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more


Walmart continues to make strides in cracking the code on deploying agentic AI at enterprise scale. Their secret? Treating trust as an engineering requirement, not some compliance checkbox you tick at the end.

During the “Trust in the Algorithm: How Walmart’s Agentic AI Is Redefining Consumer Confidence and Retail Leadership” session at VB Transform 2025, Walmart’s VP of Emerging Technology Desirée Gosby, explained how the retail giant operationalizes thousands of AI use cases. One of the retailer’s primary objectives is to consistently maintain and strengthen customer confidence among its 255 million weekly shoppers.

“We see this as a pretty big inflection point, very similar to the internet,” Gosby told industry analyst Susan Etlinger during Tuesday’s morning session. “It’s as profound in terms of how we’re actually going to operate, how we actually do work.”

The session delivered valuable lessons learned from Walmart’s AI deployment experiences. Implicit throughout the discussion is the retail giant’s continual search for new ways to apply distributed systems architecture principles, thereby avoiding the creation of technical debt.

>>See all our Transform 2025 coverage here<<

Four-stakeholder framework structures AI deployment

Walmart’s AI architecture rejects horizontal platforms for targeted stakeholder solutions. Each group receives purpose-built tools that address specific operational frictions.

Customers engage Sparky for natural language shopping. Field associates get inventory and workflow optimization tools. Merchants access decision-support systems for category management. Sellers receive business integration capabilities. “And then, of course, we’ve got developers, and really, you know, giving them the superpowers and charging them up with, you know, the new agent of tools,” Gosby explained.

“We have hundreds, if not thousands, of different use cases across the company that we’re bringing to life,” Gosby revealed. The scale demands architectural discipline that most enterprises lack.

The segmentation acknowledges the fundamental need of each team in Walmart to have purpose-built tools for their specific jobs. Store associates managing inventory need different tools from merchants analyzing regional trends. Generic platforms fail because they ignore operational reality. Walmart’s specificity drives adoption through relevance, not mandate.

Trust economics are driving AI adoption at Walmart

Walmart discovered that trust is built through value delivery, not just mandatory training programs that associates, at times, question the value of.

Gosby’s example resonated as she explained her mother’s shopping evolution from weekly store visits to COVID-era deliveries, illustrating exactly how natural adoption works. Each step provided an immediate, tangible benefit. No friction, no forced change management, yet the progression happened faster than anyone could have predicted.

“She’s been interacting with AI through that whole time,” Gosby explained. “The fact that she was able to go to the store and get what she wanted, it was on the shelf. AI was used to do that.”

The benefits customers are getting from Walmart’s predictive commerce vision are further reflected in Gosby’s mother’s experiences. “Instead of having to go weekly, figure out what groceries you need to have delivered, what if it just showed up for you automatically?” That’s the essence of predictive commerce and how it delivers value at scale to every Walmart customer.

“If you’re adding value to their lives, helping them remove friction, helping them save money and live better, which is part of our mission, then the trust comes,” Gosby stated. Associates follow the same pattern. When AI actually improves their work, saves them time and helps them excel, adoption happens naturally and trust is earned.

Fashion cycles compress from months to weeks

Walmart’s Trend to Product system quantifies the operational value of AI. The platform synthesizes social media signals, customer behavior and regional patterns to slash product development from months to weeks.

“Trend to Product has gotten us down from months to weeks to getting the right products to our customers,” Gosby revealed. The system creates products in response to real-time demand rather than historical data.

The months-to-weeks compression transforms Walmart’s retail economics. Inventory turns accelerate. Markdown exposure shrinks. Capital efficiency multiplies. The company maintains price leadership while matching any competitor’s speed-to-market capabilities. Every high-velocity category can benefit from using AI to shrink time-to-market and deliver quantifiable gains.

How Walmart uses MCP Protocol to create a scalable agent architecture

Walmart’s approach to agent orchestration draws directly from its hard-won experience with distributed systems. The company uses Model Context Protocol (MCP) to standardize how agents interact with existing services.

“We break down our domains and really looking at how do we wrap those things as MCP protocol, and then exposing those things that we can then start to orchestrate different agents,” Gosby explained. The strategy transforms existing infrastructure rather than replacing it.

The architectural philosophy runs deeper than protocols. “The change that we’re seeing today is very similar to what we’ve seen when we went from monoliths to distributed systems. We don’t want to repeat those mistakes,” Gosby stated.

Gosby outlined the execution requirements: “How do you decompose your domains? What MCP servers should you have? What sort of agent orchestration should you have?” At Walmart, these represent daily operational decisions, not theoretical exercises.

“We’re looking to take our existing infrastructure, break it down, and then recompose it into the agents that we want to be able to build,” Gosby explained. This standardization-first approach enables flexibility. Services built years ago now power agentic experiences through proper abstraction layers.

Merchant expertise becomes enterprise intelligence

Walmart leverages decades of employee knowledge, making it a core component of its growing AI capabilities. The company systematically captures category expertise from thousands of merchants, creating a competitive advantage no digital-first retailer can match.

“We have thousands of merchants who are excellent at what they do. They are experts in the categories that they support,” Gosby explained. “We have a cheese merchant who knows exactly what wine goes or what cheese pairing, but that data isn’t necessarily captured in a structured way.”

AI operationalizes this knowledge. “With the tools that we have, we can capture that expertise that they have and really bring that to bear for our customers,” Gosby said. The application is specific: “When they’re trying to figure out, hey, I need to throw the party, what kind of appetizers should I have?”

The strategic advantage compounds. Decades of merchant expertise become accessible through natural language queries. Digital-first retailers lack this human knowledge foundation. Walmart’s 2.2 million associates represent proprietary intelligence that algorithms cannot synthesize independently.

New metrics measure autonomous success

Walmart pioneers measurement systems designed for autonomous AI rather than human-driven processes. Traditional funnel metrics fail when agents handle end-to-end workflows.

“In an agentic world, we’re starting to work through this, and it’s going to change,” Gosby said. “The metrics around conversion and things like that, those are not going to change, but we’re going to be looking at goal completion.”

The shift reflects operational reality. “Did we actually achieve what is the ultimate goal that our associate, that our customers, are actually solving?” Gosby asked. The question reframes success measurement.

“At the end of the day, it’s a measure of, are we delivering the benefit? Are we delivering the value that we expect, and then working back from there to basically figure out the right metrics?” Gosby explained. Problem resolution matters more than process compliance. How AI is helping customers achieve their goals is prioritized over conversion funnels.

Enterprise lessons from Walmart’s AI transformation

Walmart’s Transform 2025 session delivers actionable intelligence for enterprise AI deployment. The company’s operational approach provides a framework that has been validated at scale.

  • Apply architectural discipline from day one. The shift from monolithic to distributed systems provided Walmart with the lessons it needed to learn to succeed with AI deployments. The key lesson learned is to build proper foundations before scaling and define a systematic approach that prevents expensive rework.
  • Match solutions to specific user needs. One-size-fits-all AI fails every time. Store associates need different tools than merchants. Suppliers require different capabilities than developers. Walmart’s targeted approach drives adoption.
  • Build trust through proven value. Start with clear wins that deliver measurable results. Walmart moved from basic inventory management to predictive commerce step by step. Each success earns insights and knowledge for the next.
  • Turn employee knowledge into enterprise assets. Decades of specialist expertise exists within your organization. Walmart systematically captures merchant intelligence and operationalizes it across 255 million weekly transactions. This institutional knowledge creates competitive advantage no algorithm can replicate from scratch.
  • Measure what matters in autonomous systems. Conversion rates miss the point when AI handles entire workflows. Focus on problem resolution and value delivery. Walmart’s metrics evolved to match operational reality.
  • Standardize before complexity hits. Integration failures killed more projects than bad code ever did. Walmart’s protocol decisions prevent the chaos that derails most AI initiatives. Structure enables speed.

“It always comes back to basics,” Gosby advised. “Take a step back and first understand what problems do you really need to solve for your customers, for our associates. Where is there friction? Where is there manual work that you can now start to think differently about?”

Walmart’s blueprint scales beyond retail

Walmart demonstrates how enterprise AI succeeds through engineering discipline and systematic deployment. The company processes millions of daily transactions across 4,700 stores by treating each stakeholder group as a distinct challenge requiring tailored, real-time solutions.

“It’s permeating everything it is that we do,” Gosby explained. “But at the end of the day, the way that we look at it is we always start with our customers and our members and really understanding how it’s going to impact them.”

Their framework applies across industries. Financial services organizations balancing customer needs with regulatory requirements, healthcare systems coordinating patient care across providers, manufacturers managing complex supply chains are all facing similar multi-stakeholder challenges. Walmart’s approach provides a tested methodology for addressing this complexity.

“Our customers are trying to solve a problem for themselves. Same thing for our associates,” Gosby stated. “Did we actually solve that problem with these new tools?” This focus on problem resolution rather than technology deployment drives measurable outcomes. Walmart’s scale validates the approach for any enterprise ready to move beyond pilot programs.

Similar Posts

  • RealSense spins out of Intel to scale its stereoscopic imaging technology

    After 14 years of developing inside of semiconductor giant Intel, RealSense is striking out on its own.

    RealSense sells cameras that use stereoscopic imaging, a process that combines two images of the same object from different angles to create depth, enhanced with infrared light. This technology helps machines like robots, drones, and autonomous vehicles have a better perception of the physical world around them. The tech is also used for facial authentication.

    “The common denominator of all of them is they live in the real, physical world,” CEO Nadav Orbach told TechCrunch. “They need to understand the surroundings in 3D and based on that, take and plan actions right in the world. And for that, they need a real-time, high-accuracy ability to understand the surrounding in 3D. And that’s what we do best.”

    Orbach joined Intel back in 2006 as a CPU architect in Israel. He started working on vision technology in 2011 before becoming the general manager of incubation and disruptive innovation in 2022 and moving to San Francisco last year.

    “We knew and understood that 3D perception was going to be big,” Orbach said about the early days of RealSense. “To be honest, we weren’t quite sure in which domain. We tried that across different market segments and different applications, all the way from gesture recognition with computers, phones, until we really found our sweet spot over the years, mostly in robotics.”

    The company works with numerous industries outside of robotics, too. Orbach said they’ve heard from fish farms looking to track the volume inside their pens. Chipotle has also used RealSense cameras, in a partnership with AI restaurant software company PreciTaste, to track when food containers are low.

    RealSense has more than 3,000 customers and has seen a surge in new interest over the last three to four years as AI has improved. With that, the applications for robotics, especially, have scaled.

    Techcrunch event

    Boston, MA
    |
    July 15

    The company realized it may have a better chance keeping up with demand — and scaling itself — if it spun out of Intel and raised its own capital, Orbach said.

    The spinout plans hatched last year and got the approval from former Intel CEO Pat Gelsinger. The company is now independent and raised a $50 million Series A funding round from Intel Capital and other strategic investors to get started on its own.

    “For me, it was exciting, to be honest,” Orbach said. “I’m a veteran executive in the company, but it’s first time that I’m, you know, I was on the other side of the table. It was a very humbling experience for me as a first-time CEO to go and and raise money.”

    RealSense will put the capital toward building out its go-to-market team and making improvements to its technology. The company is particularly focused on improving the tech so it can help improve safety during humans and robot interactions and to improve access control.

    “There is a learning curve of, you know, stepping out,” Orbach said. “I’m extremely excited about that. I’m fortunate to have a very strong team with a lot of people in my team that that have entrepreneurial experience. I feel that with my background, together with with some strong teammates, I think we have the right mix for success. And for me, it’s a dream coming true.”

  • Try on styles with AI, jump on great prices and more

    Whether you’re still on the hunt for the perfect summer maxi skirt, dreaming about a new fall jacket or starting your back to school shopping, our shopping tools can help you explore your personal style and get a good price. Here are a few ways you can use Google’s latest shopping features:Try clothes on, virtuallyAt I/O in May, we introduced our try on tool as a limited experiment in Search Labs, allowing shoppers to upload a photo of themselves and use AI to virtually try on clothes. Today, try on is launching in the U.S., letting you easily try on styles from the billions of apparel items in our Shopping Graph across Search, Google Shopping and even product results on Google Images.

  • Google DeepMind’s new AI can help historians understand ancient Latin inscriptions

    Google DeepMind has unveiled new artificial-intelligence software that could help historians recover the meaning and context behind ancient Latin engravings.  Aeneas can analyze words written in long-weathered stone to say when and where they were originally inscribed. It follows Google’s previous archaeological tool Ithaca, which also used deep learning to reconstruct and contextualize ancient text, in its case Greek. But while Ithaca and Aeneas use some similar systems, Aeneas also promises to give researchers jumping-off points for further analysis. To do this, Aeneas takes in partial transcriptions of an inscription alongside a scanned image of it. Using these, it gives possible dates and places of origins for the engraving, along with potential fill-ins for any missing text. For example, a slab damaged at the start and continuing with … us populusque Romanus would likely prompt Aeneas to guess that Senat comes before us to create the phrase Senatus populusque Romanus, “The Senate and the people of Rome.”  This is similar to how Ithaca works. But Aeneas also cross-references the text with a stored database of almost 150,000 inscriptions, which originated everywhere from modern-day Britain to modern-day Iraq, to give possible parallels—other catalogued Latin engravings that feature similar words, phrases, and analogies. 
    This database, alongside a few thousand images of inscriptions, makes up the training set for Aeneas’s deep neural network. While it may seem like a good number of samples, it pales in comparison to the billions of documents used to train general-purpose large language models like Google’s Gemini. There simply aren’t enough high-quality scans of inscriptions to train a language model to learn this kind of task. That’s why specialized solutions like Aeneas are needed.  The Aeneas team believes it could help researchers “connect the past,” said Yannis Assael, a researcher at Google DeepMind who worked on the project. Rather than seeking to automate epigraphy—the research field dealing with deciphering and understanding inscriptions—he and his colleagues are interested in “crafting a tool that will integrate with the workflow of a historian,” Assael said in a press briefing. 
    Their goal is to give researchers trying to analyze a specific inscription many hypotheses to work from, saving them the effort of sifting through records by hand. To validate the system, the team presented 23 historians with inscriptions that had been previously dated and tested their workflows both with and without Aeneas. The findings, which were published today in Nature, showed that Aeneas helped spur research ideas among the historians for 90% of inscriptions and that it led to more accurate determinations of where and when the inscriptions originated. In addition to this study, the researchers tested Aeneas on the Monumentum Ancyranum, a famous inscription carved into the walls of a temple in Ankara, Turkey. Here, Aeneas managed to give estimates and parallels that reflected existing historical analysis of the work, and in its attention to detail, the paper claims, it closely matched how a trained historian would approach the problem. “That was jaw-dropping,” Thea Sommerschield, an epigrapher at the University of Nottingham who also worked on Aeneas, said in the press briefing.  However, much remains to be seen about Aeneas’s capabilities in the real world. It doesn’t guess the meaning of texts, so it can’t interpret newly found engravings on its own, and it’s not clear yet how useful it will be to historians’ workflows in the long term, according to Kathleen Coleman, a professor of classics at Harvard. The Monumentum Ancyranum is considered to be one of the best-known and most well-studied inscriptions in epigraphy, raising the question of how Aeneas will fare on more obscure samples.  Google DeepMind has now made Aeneas open-source, and the interface for the system is freely available for teachers, students, museum workers, and academics. The group is working with schools in Belgium to integrate Aeneas into their secondary history education.  “To have Aeneas at your side while you’re in the museum or at the archaeological site where a new inscription has just been found—that is our sort of dream scenario,” Sommerschield said.

  • Razer Pro Click V2 Vertical Review: A Hybrid Gaming Mouse

    Switching to a vertical mouse is a hard sell. Having to change how you use a mouse completely can be an intimidating task, especially with how unnatural the new hand position feels at first—you’re going entirely against the muscle memory you’ve spent years building up.One of the largest challenges to the switch is the initial loss of pointer accuracy. If you’re in an office setting, you may find yourself wandering around a bit or struggling to move your new mouse as quickly as you did before. But in a slow-paced setting like that, all you struggle with is a few mis-clicks or slightly slower navigation. If you try to make this transition with gaming, it’s far more jarring, and the consequences are much more immediately noticeable.But even if it’s difficult to adapt to, could vertical mice be the future of gaming? Razer’s new Pro Click V2 Vertical Edition is a hybrid productivity and gaming vertical mouse. Vertical mice typically cater to office workers, but the focus on gaming performance makes the $120 Pro Click V2 one of a kind.Desk PresenceThe Pro Click V2 Vertical looks, more than anything else, like a modern gaming mouse. It has the textured exterior, metallic highlights, and slightly organic, H.R. Giger-esque curvature typical of Razer’s design language. But everything has been shifted around. The curved, cutting thumb rest sits on top of the mouse instead of on the side. A flare juts out from the right side as a place to rest the underside of your hand. The gunmetal highlight sits at the peak of the mouse rather than between the two buttons. Even the USB port is vertical, a humorous attention to detail.It’s intentionally designed as a gaming mouse that just happens to be vertical. Aesthetically, the only downside is the minimal RGB lighting. With only one section of lighting that runs along the bottom of the mouse, RGB lighting fans might feel disappointed. Still, it’s bright, reactive, and has great color accuracy. It’s more than enough for me, especially with how customizable it is with Razer’s Chroma software.The Pro Click V2 Vertical has the same specs as the standard Pro Click V2, with a 1,000-Hz polling rate, a 2.4-GHz dongle that can be stored on the underside, Bluetooth multi-device connectivity, and a reprogrammable button on top. The only features lost are the mouse wheel’s horizontal scrolling and toggleable non-ratcheted rotation.This mouse includes two major productivity features: app-specific profiles and multi-device connectivity, and both work effortlessly. Razer Synapse immediately detected different software and changed the active profile in response, and pressing the button on the underside of the mouse swapped between paired devices instantaneously.Beyond that, Razer Synapse is as impressive as always. I consistently find the software to be one of the best and most intuitive on the market, and that’s the case here. All of the menus are simple and efficient, the settings can be changed in real time, and the adjustments all have tooltips and explanations to tell you exactly what you’re changing.Annoyingly, Razer Synapse has advertisements on the homepage, something I’ve complained about when reviewing SteelSeries products in the past. However, unlike Steelseries GG, these “recommendations” can be permanently disabled in the app’s settings.Performance and PracticeThe overall hand position of the Pro Click V2 Vertical is natural, but incredibly upright. While some vertical mice, like those from Logitech or Hansker, find a middle ground between a standard and truly “vertical” hand position, Razer opted for a nearly perpendicular shape. While this is technically an ideal ergonomic shape, it will be harder to adapt if you’re moving directly from a standard mouse, and might not be as comfortable during the adjustment period.It felt unnatural for the first week or so, and required practice to use comfortably and confidently. Once I had acclimated, my speed and accuracy were nearly at the same level as a standard mouse, although consistent use still felt clunky and unfamiliar compared to the horizontal mice I’d been using for most of my life.

  • Our first long-duration energy storage partnership

    Electricity powers modern life. And we’re accelerating a wide range of technologies, from enhanced geothermal to advanced nuclear to even fusion technologies, that can enable a future where on-demand electricity needs are met with clean energy, every hour of every day.Today, we’re adding another technology to our portfolio: long duration energy storage (LDES). Through a new long-term partnership with Energy Dome, we plan to support multiple commercial projects globally to deploy their LDES technology.Energy Dome’s novel CO₂ Battery can store excess clean energy and then dispatch it back to the grid for 8-24 hours, bridging the gap between when renewable energy is generated and when it is needed. With this commercial partnership, as well as an investment in the company, we believe these projects can unlock new clean energy for grids where we operate before 2030, helping meet near-term electricity system needs and moving us closer to our 24/7 carbon-free energy goal.By bringing this first-of-a-kind LDES technology to market faster, we aim to rapidly bring its potential to communities everywhere — making reliable, affordable electricity available around the clock and supporting the resilience of grids as they integrate growing amounts of renewable energy sources.Why it’s importantLithium-ion batteries, which typically store and dispatch power for 4 hours or less, have been critical for adding electricity capacity to grids and managing short-term fluctuations in renewable generation — when the sun isn’t shining or the wind isn’t blowing. Google’s support for these shorter-duration batteries has helped the grids we rely on, from Belgium to Nevada, meet peak electricity demand and reduce the need to ramp up fossil fuel power plants.But what if we could store and dispatch clean energy for more than a few hours, or even a full day? Studies by the Electric Power Research Institute show that LDES technologies can cost-effectively integrate a growing volume of renewables onto power systems and contribute to more flexible, reliable grids. The LDES Council estimates that deploying up to 8 terawatts (TW) of LDES by 2040 could result in $540 billion in annual savings globally, thanks in part to their ability to optimize grids.How the technology worksEnergy Dome’s novel approach to energy storage uses carbon dioxide (CO₂) held in a unique dome-shaped battery. When there’s an abundance of renewable energy on the grid, the system uses that power to compress CO₂ gas into a liquid. When the grid needs more clean power, the liquid CO₂ expands back into a hot gas under pressure, creating a powerful force — much like steam escaping a pressure cooker — which spins a turbine. This spinning turbine generates carbon-free energy that can flow directly back into the grid for durations ranging from 8 to 24 hours.Energy Dome has already signed contracts to build commercial scale projects in Italy, the U.S., and India. And their technology has already proven successful, having injected electrons into the Italian grid for more than three years, thanks to their commercial demonstration facility and now with their full-scale 20 megawatt (MW) commercial plant in Sardinia, Italy.Why scale is crucialLDES has the potential to commercialize much faster than some of the other advanced clean energy technologies in our portfolio. This means we can use it in the near term to help the electricity system grow more flexibly and reliably, alongside other tools we’re developing such as data center demand response.By supporting multiple commercial deployments of Energy Dome’s technology globally, we aim to bring this technology to scale faster and at lower costs. Beyond our long-term collaboration with Energy Dome, we plan to support a growing range of LDES technologies under development through both commercial agreements that can catalyze wider market adoption of more mature technologies, like Energy Dome’s, as well as earlier-stage investments.To remove barriers to the deployment and commercialization of LDES and other advanced carbon-free energy technologies, we’re also advocating for clean energy policies, ensuring that energy markets fully value firm, flexible carbon-free technologies, and advancing policy measures that enable infrastructure essential for grid decarbonization and energy security.We’re excited to take this first step with Energy Dome to unlock the full potential of LDES. Our partnership will strengthen grid resilience while enabling us to power our technologies, grow our economies and keep the lights on in our homes with 24/7 clean energy.

Leave a Reply

Your email address will not be published. Required fields are marked *