The hidden scaling cliff that’s about to break your agent rollouts

Join the event trusted by enterprise leaders for nearly two decades. VB Transform brings together the people building real enterprise AI strategy. Learn more


Enterprises that want to build and scale agents also need to embrace another reality: agents aren’t built like other software. 

Agents are “categorically different” in how they’re built, how they operate, and how they’re improved, according to Writer CEO and co-founder May Habib. This means ditching the traditional software development life cycle when dealing with adaptive systems.

“Agents don’t reliably follow rules,” Habib said on Wednesday while on stage at VB Transform. “They are outcome-driven. They interpret. They adapt. And the behavior really only emerges in real-world environments.”

Knowing what works — and what doesn’t work — comes from Habib’s experience helping hundreds of enterprise clients build and scale enterprise-grade agents. According to Habib, more than 350 of the Fortune 1000 are Writer customers, and more than half of the Fortune 500 will be scaling agents with Writer by the end of 2025.

Using non-deterministic tech to produce powerful outputs can even be “really nightmarish,” Habib said — especially when trying to scale agents systemically. Even if enterprise teams can spin up agents without product managers and designers, Habib thinks a “PM mindset” is still needed for collaborating, building, iterating and maintaining agents.

“Unfortunately or fortunately, depending on your perspective, IT is going to be left holding the bag if they don’t lead their business counterparts into that new way of building.”

>>See all our Transform 2025 coverage here<<

Why goal-based agents is the right approach 

One of the shifts in thinking includes understanding the outcome-based nature of agents. For example, she said that many customers request agents to assist their legal teams in reviewing or redlining contracts. But that’s too open-ended. Instead, a goal-oriented approach means designing an agent to reduce the time spent reviewing and redlining contracts.

“In the traditional software development life cycle, you are designing for a deterministic set of very predictable steps,” Habib said. “It’s input in, input out in a more deterministic way. But with agents, you’re seeking to shape agentic behavior. So you are seeking less of a controlled flow and much more to give context and guide decision-making by the agent.”

Another difference is building a blueprint for agents that instructs them with business logic, rather than providing them with workflows to follow. This includes designing reasoning loops and collaborating with subject experts to map processes that promote desired behaviors.

While there’s a lot of talk about scaling agents, Writer is still helping most clients with building them one at a time. That’s because it’s important first to answer questions about who owns and audits the agent, who makes sure it stays relevant and still checks if it’s still producing desired outcomes.

“There is a scaling cliff that folks get to very, very quickly without a new approach to building and scaling agents,” Habib said. “There is a cliff that folks are going to get to when their organization’s ability to manage agents responsibly really outstrips the pace of development happening department by department.”

QA for agents vs software

Quality assurance is also different for agents. Instead of an objective checklist, agentic evaluation includes accounting for non-binary behavior and assessing how agents act in real-world situations. That’s because failure isn’t always obvious — and not as black and white as checking if something broke. Instead, Habib said it’s better to check if an agent behaved well, asking if fail-safes worked, evaluating outcomes and intent: “The goal here isn’t perfection It is behavioral confidence, because there is a lot of subjectivity in this here.”

Businesses that don’t understand the importance of iteration end up playing “a constant game of tennis that just wears down each side until they don’t want to play anymore,” Habib said. It’s also important for teams to be okay with agents being less than perfect and more about “launching them safely and running fast and iterating over and over and over.”

Despite the challenges, there are examples of AI agents already helping bring in new revenue for enterprise businesses. For example, Habib mentioned a major bank that collaborated with Writer to develop an agent-based system, resulting in a new upsell pipeline worth $600 million by onboarding new customers into multiple product lines.

New version controls for AI agents

Agentic maintenance is also different. Traditional software maintenance involves checking the code when something breaks, but Habib said AI agents require a new kind of version control for everything that can shape behavior. It also requires proper governance and ensuring that agents remain useful over time, rather than incurring unnecessary costs.

Because models don’t map cleanly to AI agents, Habib said maintenance includes checking prompts, model settings, tool schemas and memory configuration. It also means fully tracing executions across inputs, outputs, reasoning steps, tool calls and human interactions. 

“You can update a [large language model] LLM prompt and watch the agent behave completely differently even though nothing in the git history actually changed,” Habib said. “The model links shift, retrieval indexes get updated, tool APIs evolve and suddenly the same prompt does not behave as expected…It can feel like we are debugging ghosts.”

Similar Posts

  • Web Guide: An experimental AI-organized search results page

    We’re launching Web Guide, a Search Labs experiment that uses AI to intelligently organize the search results page, making it easier to find information and web pages.Web Guide groups web links in helpful ways — like pages related to specific aspects of your query. Under the hood, Web Guide uses a custom version of Gemini to better understand both a search query and content on the web, creating more powerful search capabilities that better surface web pages you may not have previously discovered. Similar to AI Mode, Web Guide uses a query fan-out technique, concurrently issuing multiple related searches to identify the most relevant results.For example, try it for open-ended searches like “how to solo travel in Japan.” Or try detailed queries in multiple sentences like, “My family is spread across multiple time zones. What are the best tools for staying connected and maintaining close relationships despite the distance?”

  • RealSense spins out of Intel to scale its stereoscopic imaging technology

    After 14 years of developing inside of semiconductor giant Intel, RealSense is striking out on its own.

    RealSense sells cameras that use stereoscopic imaging, a process that combines two images of the same object from different angles to create depth, enhanced with infrared light. This technology helps machines like robots, drones, and autonomous vehicles have a better perception of the physical world around them. The tech is also used for facial authentication.

    “The common denominator of all of them is they live in the real, physical world,” CEO Nadav Orbach told TechCrunch. “They need to understand the surroundings in 3D and based on that, take and plan actions right in the world. And for that, they need a real-time, high-accuracy ability to understand the surrounding in 3D. And that’s what we do best.”

    Orbach joined Intel back in 2006 as a CPU architect in Israel. He started working on vision technology in 2011 before becoming the general manager of incubation and disruptive innovation in 2022 and moving to San Francisco last year.

    “We knew and understood that 3D perception was going to be big,” Orbach said about the early days of RealSense. “To be honest, we weren’t quite sure in which domain. We tried that across different market segments and different applications, all the way from gesture recognition with computers, phones, until we really found our sweet spot over the years, mostly in robotics.”

    The company works with numerous industries outside of robotics, too. Orbach said they’ve heard from fish farms looking to track the volume inside their pens. Chipotle has also used RealSense cameras, in a partnership with AI restaurant software company PreciTaste, to track when food containers are low.

    RealSense has more than 3,000 customers and has seen a surge in new interest over the last three to four years as AI has improved. With that, the applications for robotics, especially, have scaled.

    Techcrunch event

    Boston, MA
    |
    July 15

    The company realized it may have a better chance keeping up with demand — and scaling itself — if it spun out of Intel and raised its own capital, Orbach said.

    The spinout plans hatched last year and got the approval from former Intel CEO Pat Gelsinger. The company is now independent and raised a $50 million Series A funding round from Intel Capital and other strategic investors to get started on its own.

    “For me, it was exciting, to be honest,” Orbach said. “I’m a veteran executive in the company, but it’s first time that I’m, you know, I was on the other side of the table. It was a very humbling experience for me as a first-time CEO to go and and raise money.”

    RealSense will put the capital toward building out its go-to-market team and making improvements to its technology. The company is particularly focused on improving the tech so it can help improve safety during humans and robot interactions and to improve access control.

    “There is a learning curve of, you know, stepping out,” Orbach said. “I’m extremely excited about that. I’m fortunate to have a very strong team with a lot of people in my team that that have entrepreneurial experience. I feel that with my background, together with with some strong teammates, I think we have the right mix for success. And for me, it’s a dream coming true.”

  • Trump’s AI Action Plan is a distraction

    On Wednesday, President Trump issued three executive orders, delivered a speech, and released an action plan, all on the topic of continuing American leadership in AI.  The plan contains dozens of proposed actions, grouped into three “pillars”: accelerating innovation, building infrastructure, and leading international diplomacy and security. Some of its recommendations are thoughtful even if incremental, some clearly serve ideological ends, and many enrich big tech companies, but the plan is just a set of recommended actions.  The three executive orders, on the other hand, actually operationalize one subset of actions from each pillar:  One aims to prevent “woke AI” by mandating that the federal government procure only large language models deemed “truth-seeking” and “ideologically neutral” rather than ones allegedly favoring DEI. This action purportedly accelerates AI innovation. A second aims to accelerate construction of AI data centers. A much more industry-friendly version of an order issued under President Biden, it makes available rather extreme policy levers, like effectively waiving a broad swath of environmental protections, providing government grants to the wealthiest companies in the world, and even offering federal land for private data centers. A third promotes and finances the export of US AI technologies and infrastructure, aiming to secure American diplomatic leadership and reduce international dependence on AI systems from adversarial countries. This flurry of actions made for glitzy press moments, including an hour-long speech from the president and onstage signings. But while the tech industry cheered these announcements (which will swell their coffers), they obscured the fact that the administration is currently decimating the very policies that enabled America to become the world leader in AI in the first place.
    To maintain America’s leadership in AI, you have to understand what produced it. Here are four specific long-standing public policies that helped the US achieve this leadership—advantages that the administration is undermining.  Investing federal funding in R&D  Generative AI products released recently by American companies, like ChatGPT, were developed with industry-funded research and development. But the R&D that enables today’s AI was actually funded in large part by federal government agencies—like the Defense Department, the National Science Foundation, NASA, and the National Institutes of Health—starting in the 1950s. This includes the first successful AI program in 1956, the first chatbot in 1961, and the first expert systems for doctors in the 1970s, along with breakthroughs in machine learning, neural networks, backpropagation, computer vision, and natural-language processing.
    American tax dollars also funded advances in hardware, communications networks, and other technologies underlying AI systems. Public research funding undergirded the development of lithium-ion batteries, micro hard drives, LCD screens, GPS, radio-frequency signal compression, and more in today’s smartphones, along with the chips used in AI data centers, and even the internet itself. Instead of building on this world-class research history, the Trump administration is slashing R&D funding, firing federal scientists, and squeezing leading research universities. This week’s action plan recommends investing in R&D, but the administration’s actual budget proposes cutting nondefense R&D by 36%. It also proposed actions to better coordinate and guide federal R&D, but coordination won’t yield more funding. Some say that companies’ R&D investments will make up the difference. However, companies conduct research that benefits their bottom line, not necessarily the national interest. Public investment allows broad scientific inquiry, including basic research that lacks immediate commercial applications but sometimes ends up opening massive markets years or decades later. That’s what happened with today’s AI industry. Supporting immigration and immigrants Beyond public R&D investment, America has long attracted the world’s best researchers and innovators. Today’s generative AI is based on the transformer model (the T in ChatGPT), first described by a team at Google in 2017. Six of the eight researchers on that team were born outside the US, and the other two are children of immigrants.  This isn’t an exception. Immigrants have been central to American leadership in AI. Of the 42 American companies included in the 2025 Forbes ranking of the 50 top AI startups, 60% have at least one immigrant cofounder, according to an analysis by the Institute for Progress. Immigrants also cofounded or head the companies at the center of the AI ecosystem: OpenAI, Anthropic, Google, Microsoft, Nvidia, Intel, and AMD. “Brain drain” is a term that was first coined to describe scientists’ leaving other countries for the US after World War II—to the Americans’ benefit. Sadly, the trend has begun reversing this year. Recent studies suggest that the US is already losing its AI talent edge through the administration’s anti-immigration actions (including actions taken against AI researchers) and cuts to R&D funding. Banning noncompetes Attracting talented minds is only half the equation; giving them freedom to innovate is just as crucial.

    Silicon Valley got its name because of mid-20th-century companies that made semiconductors from silicon, starting with the founding of Shockley Semiconductor in 1955. Two years later, a group of employees, the “Traitorous Eight,” quit to launch a competitor, Fairchild Semiconductor. By the end of the 1960s, successive groups of former Fairchild employees had left to start Intel, AMD, and others collectively dubbed the “Fairchildren.”  Software and internet companies eventually followed, again founded by people who had worked for their predecessors. In the 1990s, former Yahoo employees founded WhatsApp, Slack, and Cloudera; the “PayPal Mafia” created LinkedIn, YouTube, and fintech firms like Affirm. Former Google employees have launched more than 1,200 companies, including Instagram and Foursquare. AI is no different. OpenAI has founders that worked at other tech companies and alumni who have gone on to launch over a dozen AI startups, including notable ones like Anthropic and Perplexity. This labor fluidity and the innovation it has created were possible in large part, according to many historians, because California’s 1872 constitution has been interpreted to prohibit noncompete agreements in employment contracts—a statewide protection the state originally shared only with North Dakota and Oklahoma. These agreements bind one in five American workers. Last year, the Federal Trade Commission under President Biden moved to ban noncompetes nationwide, but a Trump-appointed federal judge has halted the action. The current FTC has signaled limited support for the ban and may be comfortable dropping it. If noncompetes persist, American AI innovation, especially outside California, will be limited. Pursuing antitrust actions One of this week’s announcements requires the review of FTC investigations and settlements that “burden AI innovation.” During the last administration the agency was reportedly investigating Microsoft’s AI actions, and several big tech companies have settlements that their lawyers surely see as burdensome, meaning this one action could thwart recent progress in antitrust policy. That’s an issue because, in addition to the labor fluidity achieved by banning noncompetes, antitrust policy has also acted as a key lubricant to the gears of Silicon Valley innovation.  Major antitrust cases in the second half of the 1900s, against AT&T, IBM, and Microsoft, allowed innovation and a flourishing market for semiconductors, software, and internet companies, as the antitrust scholar Giovanna Massarotto has described. William Shockley was able to start the first semiconductor company in Silicon Valley only because AT&T had been forced to license its patent on the transistor as part of a consent decree resolving a DOJ antitrust lawsuit against the company in the 1950s. 
    The early software market then took off because in the late 1960s, IBM unbundled its software and hardware offerings as a response to antitrust pressure from the federal government. As Massarotto explains, the 1950s AT&T consent decree also aided the flourishing of open-source software, which plays a major role in today’s technology ecosystem, including the operating systems for mobile phones and cloud computing servers. Meanwhile, many attribute the success of early 2000s internet companies like Google to the competitive breathing room created by the federal government’s antitrust lawsuit against Microsoft in the 1990s. 
    Over and over, antitrust actions targeting the dominant actors of one era enabled the formation of the next. And today, big tech is stifling the AI market. While antitrust advocates were rightly optimistic about this administration’s posture given key appointments early on, this week’s announcements should dampen that excitement.  I don’t want to lose focus on where things are: We should want a future in which lives are improved by the positive uses of AI.  But if America wants to continue leading the world in this technology, we must invest in what made us leaders in the first place: bold public research, open doors for global talent, and fair competition.  Prioritizing short-term industry profits over these bedrock principles won’t just put our technological future at risk—it will jeopardize America’s role as the world’s innovation superpower.  Asad Ramzanali is the director of artificial intelligence and technology policy at the Vanderbilt Policy Accelerator. He previously served as the chief of staff and deputy director of strategy of the White House Office of Science and Technology Policy under President Biden.

  • Google DeepMind’s new AI can help historians understand ancient Latin inscriptions

    Google DeepMind has unveiled new artificial-intelligence software that could help historians recover the meaning and context behind ancient Latin engravings.  Aeneas can analyze words written in long-weathered stone to say when and where they were originally inscribed. It follows Google’s previous archaeological tool Ithaca, which also used deep learning to reconstruct and contextualize ancient text, in its case Greek. But while Ithaca and Aeneas use some similar systems, Aeneas also promises to give researchers jumping-off points for further analysis. To do this, Aeneas takes in partial transcriptions of an inscription alongside a scanned image of it. Using these, it gives possible dates and places of origins for the engraving, along with potential fill-ins for any missing text. For example, a slab damaged at the start and continuing with … us populusque Romanus would likely prompt Aeneas to guess that Senat comes before us to create the phrase Senatus populusque Romanus, “The Senate and the people of Rome.”  This is similar to how Ithaca works. But Aeneas also cross-references the text with a stored database of almost 150,000 inscriptions, which originated everywhere from modern-day Britain to modern-day Iraq, to give possible parallels—other catalogued Latin engravings that feature similar words, phrases, and analogies. 
    This database, alongside a few thousand images of inscriptions, makes up the training set for Aeneas’s deep neural network. While it may seem like a good number of samples, it pales in comparison to the billions of documents used to train general-purpose large language models like Google’s Gemini. There simply aren’t enough high-quality scans of inscriptions to train a language model to learn this kind of task. That’s why specialized solutions like Aeneas are needed.  The Aeneas team believes it could help researchers “connect the past,” said Yannis Assael, a researcher at Google DeepMind who worked on the project. Rather than seeking to automate epigraphy—the research field dealing with deciphering and understanding inscriptions—he and his colleagues are interested in “crafting a tool that will integrate with the workflow of a historian,” Assael said in a press briefing. 
    Their goal is to give researchers trying to analyze a specific inscription many hypotheses to work from, saving them the effort of sifting through records by hand. To validate the system, the team presented 23 historians with inscriptions that had been previously dated and tested their workflows both with and without Aeneas. The findings, which were published today in Nature, showed that Aeneas helped spur research ideas among the historians for 90% of inscriptions and that it led to more accurate determinations of where and when the inscriptions originated. In addition to this study, the researchers tested Aeneas on the Monumentum Ancyranum, a famous inscription carved into the walls of a temple in Ankara, Turkey. Here, Aeneas managed to give estimates and parallels that reflected existing historical analysis of the work, and in its attention to detail, the paper claims, it closely matched how a trained historian would approach the problem. “That was jaw-dropping,” Thea Sommerschield, an epigrapher at the University of Nottingham who also worked on Aeneas, said in the press briefing.  However, much remains to be seen about Aeneas’s capabilities in the real world. It doesn’t guess the meaning of texts, so it can’t interpret newly found engravings on its own, and it’s not clear yet how useful it will be to historians’ workflows in the long term, according to Kathleen Coleman, a professor of classics at Harvard. The Monumentum Ancyranum is considered to be one of the best-known and most well-studied inscriptions in epigraphy, raising the question of how Aeneas will fare on more obscure samples.  Google DeepMind has now made Aeneas open-source, and the interface for the system is freely available for teachers, students, museum workers, and academics. The group is working with schools in Belgium to integrate Aeneas into their secondary history education.  “To have Aeneas at your side while you’re in the museum or at the archaeological site where a new inscription has just been found—that is our sort of dream scenario,” Sommerschield said.

  • Obvio’s stop sign cameras use AI to root out unsafe drivers

    American streets are incredibly dangerous for pedestrians. A San Carlos, California-based startup called Obvio thinks it can change that by installing cameras at stop signs — a solution the founders also say won’t create a panopticon. 

    That’s a bold claim at a time when other companies like Flock have been criticized for how its license plate-reading cameras have become a crucial tool in an overreaching surveillance state. 

    Obvio founders Ali Rehan and Dhruv Maheshwari believe they can build a big enough business without indulging those worst impulses. They’ve designed the product with surveillance and data-sharing limitations to ensure they can follow through with that claim.

    They’ve found deep pockets willing to believe them, too. The company has just completed a $22 million Series A funding round led by Bain Capital Ventures. Obvio plans to use those funds to expand beyond the first five cities where it’s currently operating in Maryland. 

    Rehan and Maheshwari met while working at Motive, a company that makes dashboard cameras for the trucking industry. While there, Maheshwari told TechCrunch the pair realized “a lot of other normal passenger vehicles are awful drivers.” 

    The founders said they were stunned the more they looked into road safety. Not only were streets and crosswalks getting more dangerous for pedestrians, but in their eyes, the U.S. was also falling behind on enforcement. 

    [embedded content]

    “Most other countries are actually pretty good at this,” Maheshwari said. “They have speed camera technology. They have a good culture of driving safety. The U.S. is actually one of the worst across all the modern nations.”

    Maheshwari and Rehan began studying up on road safety by reading books and attending conferences. They found that people in the industry gravitated toward three general solutions: education, engineering, and enforcement. 

    In their eyes, those approaches were often too separated from each other. It’s hard to quantify the impact of educational efforts. Local officials may try to fix a problematic intersection by, say, installing a roundabout, but that can take years of work and millions of dollars. And law enforcement can’t camp out at every stop sign.

    Rehan and Maheshwari saw promise in combining them. 

    The result is a pylon (often brightly-colored) topped with a solar-powered camera that can be installed near almost any intersection. It’s designed not to blend in — part of the education and awareness aspect — and it’s also carefully engineered to be cheap and easy to install.

    The on-device AI is trained to spot the worst types of stop sign or other infractions. (The company also claims on its website it can catch speeding, crosswalk violations, illegal turns, unsafe lane changes, and even distracted driving.) When one of these things happen, the system matches a car’s license plate to the state’s DMV database. 

    All of that information — the accuracy of the violation, the license plate — is verified by either Obvio staff or contractors before it’s sent to law enforcement, which then has to review the infractions before issuing a citation.

    Obvio gives the tech to municipalities for free and makes money from the citations. Exactly how that citation revenue will get split between Obvio and the governments will vary from place to place, as Maheshwari said regulations about such agreements differ by state.

    That clearly creates an incentive for increasing the number of citations. But Rehan and Maheshwari said they can build a business around stopping the worst offenses across a wide swath of American cities. They also said they want Obvio to remain present in — and responsive to — the communities that use their tech.

    “Automated enforcement should be used in conjunction with community advocacy and community support, it shouldn’t be this camera that you put up that does revenue grab[s] and gotchas,” Maheshwari said. The goal is to “start using these cameras in a way to warn and deter the most egregious drivers [so] you can actually create communitywide support and behavior change.”

    Cities and their citizens “need to trust us,” Maheshwari said. 

    There’s also a technological explanation for why Obvio’s cameras may not become an overpowered surveillance tool for law enforcement beyond their intended use.

    Obvio’s camera pylon records and processes its footage locally. It’s only when a violation is spotted that the footage leaves the device. Otherwise, all other footage of vehicles and pedestrians passing through a given intersection stays on the device for about 12 hours before it gets deleted. (The footage is also technically owned by the municipalities, which have remote access.)

    This doesn’t eliminate the chance that law enforcement will use the footage to surveil citizens in other ways. But it does reduce that chance.

    That focus is what drove Bain Capital Ventures partner Ajay Agarwal to invest in Obvio.

    “Yes, in the short term, you can maximize profits, and erode those values, but I think over time, it will limit the ability of this company to be ubiquitous. It’ll create enemies or create people who don’t want this,” he told TechCrunch. “Great founders are willing to sacrifice entire lines of business, frankly, and lots of revenue, in pursuit of the ultimate mission.”

Leave a Reply

Your email address will not be published. Required fields are marked *