
PromptQL Partners with UC Berkeley to Develop New Data Agent Benchmark for Reliability of Enterprise AI Agents
BERKELEY, Calif., June 04, 2025 (GLOBE NEWSWIRE) — PromptQL , a platform for reliable AI, today announced a strategic research collaboration with the University of California, Berkeley to develop the first comprehensive data agent benchmark for enterprise reliability specifically designed to evaluate general-purpose AI data agents in enterprise environments.
A recent McKinsey study revealed that 78% of organizations use AI in at least one business function, however, more than 80% say their organization hasn't seen a tangible impact on enterprise-level Earnings Before Interest and Taxes (EBIT). The partnership – led by Aditya Parameswaran, Professor and Co-Director of UC Berkeley's EPIC Data Lab , along with his students – addresses this fundamental challenge organizations face when deploying AI systems in business-critical environments.
While existing agentic data benchmarks like GAIA, Spider, and FRAMES test specific AI tasks, they overlook the complexity, reliability demands, and messy, siloed data that define real business environments. The forthcoming data agent benchmark aims to offer a solution by creating a framework that reflects real-world complexities.
'Our customer conversations reveal a clear pattern—they're ready to move from proof-of-concepts to production AI, yet they lack the evaluation tools to make confident deployment decisions,' said Tanmai Gopal, CEO of PromptQL. 'The data agent benchmark changes that by using representative datasets from our work in telecom, healthcare, finance, retail, and anti-money laundering to reflect the real complexity of enterprise AI.'
UC Berkeley's EPIC Data Lab brings expertise to this collaboration. Professor Parameswaran is a leading authority on the use of AI for next-gen usable data analysis tools and has received numerous prestigious awards. His research group has created widely-adopted data tools with tens of millions of downloads.
'Current benchmarks suffer from what I call the '1% problem'—they're built for tech giants and ignore the 99% of organizations grappling with real-world data complexity,' Parameswaran said. 'The data agent benchmark marks a shift toward evaluating AI based on the reliability, transparency, and practical value enterprises actually need. This collaboration bridges academic rigor with the production insights PromptQL brings from real deployments.'
The data agent benchmark beta will be revealed later this year. Organizations interested in early access or contributing use-cases or datasets can reach out to the research team at [email protected] .
PromptQL will be at AI Engineer World's Fair , June 3-6 in San Francisco. Tanmai Gopal, PromptQL's co-founder and CEO, will present a session, 'Al Automation that Actually Works: $100M Impact on Messy Data with Zero Surprises,' on June 4 at 11:15 a.m. PT. To learn more or schedule a demo at the PromptQL booth, visit https://hasura.io/events/ai-engineer-worlds-fair-2025 .
About PromptQL
PromptQL is a next-generation AI platform from the makers of Hasura, the company behind the pioneering GraphQL Engine. Built for enterprise-grade reliability, PromptQL enables natural language analysis and automation on internal business data — with an industry-first accuracy SLA. By learning the unique language of your business and planning tasks before executing them deterministically, PromptQL brings human-level precision to AI agents.
About UC Berkeley EPIC Data Lab
The EPIC Data Lab at UC Berkeley develops low-code and no-code interfaces for data work, powered by Gen AI. Co-Led by Professor Aditya Parameswaran, the lab follows Berkeley's tradition of multidisciplinary systems research with emphasis on real-world impact and practical deployment. The lab's tools, including DocETL and other widely-adopted systems, demonstrate Berkeley's leadership in democratizing data science capabilities.
Media Contact:Erica Anderson Offleash for PromptQL
[email protected]
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Entrepreneur
25 minutes ago
- Entrepreneur
Kosmc AI Raises Pre-Seed Round to Power AI-Driven Social Commerce
The fresh funding will be used to enhance Kosmc AI's no-code product suite, deepen affiliate integrations, and scale operations across India and Southeast Asia. You're reading Entrepreneur India, an international franchise of Entrepreneur Media. Delhi-based startup Kosmc AI has secured USD 200,000 in pre-seed funding from undisclosed angel investors across India and the Middle East to expand its AI-led social commerce infrastructure. The funding will be used to enhance Kosmc AI's no-code product suite, deepen affiliate integrations, and scale operations across India and Southeast Asia. The startup is on a mission to simplify how creators and direct-to-consumer (D2C) brands monetise content and engagement across platforms like Instagram, WhatsApp, and Telegram. Founded in 2024 by Ankur Gupta and Manavta Narula, Kosmc AI is building infrastructure that bridges the gap between social media engagement and actual commerce. The startup claims to have already supported over 160,000 monthly active users in more than 25 countries. "Social platforms are now where discovery and engagement begin, but monetisation still relies on traditional e-commerce systems," said Ankur Gupta, Founder and CEO of Kosmc AI. "Kosmc is building infrastructure that allows individuals and brands to convert conversations and content into commerce—without needing code, commissions, or technical complexity." Kosmc AI offers a powerful suite of four tools: Link-in-bio storefronts for selling directly via social media for selling directly via social media Chat automation tools for handling customer conversations for handling customer conversations Smartlinks to simplify mobile checkout journeys to simplify mobile checkout journeys Affiliate infrastructure that connects creators with D2C brands The company's no-code approach empowers digital-first entrepreneurs, content creators, and early-stage brands to own their commerce journey in a mobile-first, creator-led economy. Co-founder and COO Manavta Narula added, "The company's mission is to make the digital economy more accessible. Commerce today needs to be creator-first, mobile-native, and frictionless. We are building tools that allow anyone to own their narrative, engage meaningfully, and earn from their presence online without heavy dependencies." As the creator economy continues to surge, Kosmc AI positions itself at the forefront—enabling seamless monetisation where content, community, and commerce converge.


Fast Company
an hour ago
- Fast Company
5 dark facts to remember in the face of AI hype
Emily Bender is a Professor of Linguistics at the University of Washington where she is also Faculty Director of the Computational Linguistics Master of Science program, affiliate faculty in the School of Computer Science and Engineering, and affiliate faculty in the Information School. Alex Hanna is Director of Research at the Distributed AI Research Institute and a lecturer in the School of Information at the University of California Berkeley. She has been featured in articles for the Washington Post, Financial Times, The Atlantic, and Time. What's the big idea? The AI Con is an exploration of the hype around artificial intelligence, whose interests it serves, and the harm being done under this umbrella. Society has options when it comes to pushing back against AI hype, so there is still hope that we can collectively resist and prevent tech companies from mortgaging humanity's future. Below, co-authors Emily Bender and Alex Hanna share five key insights from their new book, The AI Con: How to Fight Big Tech's Hype and Create the Future We Want. Listen to the audio version—read by Emily and Alex—in the Next Big Idea App. 1. The tech that's driving the current wave of AI hype is built on a parlor trick Chatbots like ChatGPT are impressive technology, but maybe not in the way you think. They cannot perform the range of functions they purportedly fulfill, but rather, they are designed to impress us. The key to their parlor trick lies in how people utilize language. You might think it's a simple matter of decoding what the words say, but the process is both far more complex and far more social. We interpret language by relying on everything we know (or guess) about the person who said the words, and whatever common ground we share with them. Then we make inferences about what they must have been trying to convey. We do this instinctively and reflexively. So, when we encounter synthetic text of the kind that comes out of ChatGPT and its ilk, we interpret it by imagining a mind behind the text, even though there is no mind there. In other words, the linguistic and social skills we wrap around AI outputs are what make it so easy for the purveyors of chatbots to fool us into perceiving chatbots as reasoning entities. 2. AI is not going to take your job, but it will make your job a lot worse Much of the purpose of AI technology serves to remove humans from the equation at work. The story of the Writers Guild of America strike is instructive here. In 2023, the Writers Guild of America East and West (or the WGA), the labor union representing Hollywood writers, went on strike for several reasons, including a demand to raise the pay rate that writers receive from streaming services. They also wanted to ensure that they wouldn't be reduced to babysitters for chatbots tasked to write scripts based on harebrained ideas from movie and television producers. John Lopez, a member of the WGA's AI working group, noted that writers could be paid the rewrite rate for dealing with AI-generated content, which is much less than the pay rate for an original script. We've seen the threat of image and text generators drastically reduce the number of job opportunities for graphic designers, video game artists, and journalists. This is not because these tools can adequately perform the tasks of these professionals, but they perform well enough for careers to be cut short and for workers to be rehired at a fraction of what they had been paid before, just so that they can fix the sloppy outputs of AI. 'They perform well enough for careers to be cut short and for workers to be rehired at a fraction of what they had been paid before.' Furthermore, systems that get called 'AI' are often a thin veneer that hides the tried-and-true corporate strategy of outsourcing labor to people in the Majority World, also called the Global South. Many of these workers moderate online content, test chatbots for toxic outputs, and even remotely drive vehicles that are advertised as being fully automated. Luckily, workers have been able to push back, both by concerned labor action, industrial sabotage (especially through creative tools for artists, like Nightshade and Glaze, which prevent their work from being used for training image generation models), and political education. 3. The purpose of the AI con is to disconnect people from social services Because we use language in just about every sphere of activity, and because the synthetic text extruding from machines can be trained to mimic language, it can seem like we are about to have technology that can provide medical diagnoses, personalized tutoring, wise decision making in the allocation of government services, legal representation, and more—all for just the cost of electricity (plus whatever the companies making the chatbots want to charge). But in all these cases, it's not the words that matter, but the actual thought that goes into them and the relationships they help us build and maintain. AI systems are only good for those who want to redirect funding away from social services and justify austerity measures. Meanwhile, those in power will be sure to get services from actual people, while foisting the shoddy facsimiles off on everyone else. The head of Health AI at Google, Greg Corrado, said he wouldn't want Google's Med-PaLM system to be part of his family's health care journey. That didn't stop him from bragging about how it supposedly passed a medical licensing exam. It didn't. But more to the point, designing systems to pass multiple-choice exams about medical situations is not an effective way to build useful medical technology. In these domains, AI hype takes the form of specious claims of technological solutions to social problems, based, at best, on spurious and unfounded evaluations of the systems being sold. 4. AI is not going to kill us all, but climate change might There was a time in Silicon Valley and Washington D.C. when an idiosyncratic, yet serious, question was posed to people working on technology or tech policy: 'What is your p(doom)?' p(doom) refers to probability of doom, or the likelihood that AI would somehow kill all of humanity. This doomerism is predicated on the development of artificial general intelligence (or AGI). AGI is poorly defined, but the basic idea is a system which can do a variety of tasks as well as or better than humans. Unfortunately, doomerism has serious purchase with some technologists and policymakers, and is predicated on a body of unseemly ideologies, including effective altruism, longtermism, and rationalism. These ideologies take the moral philosophy of utilitarianism to the extreme, suggesting that we need to discount harm in the present to save the billions of trillions of humans who will live in some undefined future. These ideologies are eugenicist in their origins and implications. 'Doomerism has serious purchase with some technologists and policymakers.' Meanwhile, we are likely to fail to meet the Paris Agreement's goal to limit the increase in global average temperature to well below 2 degrees Celsius above pre-industrial levels, and AI is making this problem worse. The data centers that host these tools are generating vast amounts of excess carbon, semiconductors used for their parts are leeching forever chemicals into the ground, and backup generators are projected to cause more respiratory illnesses in the poorest parts of the U.S. and elsewhere. Not only are robots not going to take over the world, but their production is going to make the climate crisis much worse. 5. None of this is inevitable The people selling AI systems and the hype around them would like us to voluntarily give up our agency in these matters. They tell us that AI, or even AGI, is inevitable, or at least that systems like ChatGPT are 'here to stay.' But none of this is inevitable. We do have agency, both collectively and individually. Collectively, we can push for regulations that prevent AI tech from being used on us and for labor contracts that keep us in control of our work. On an individual level, we can refuse to use AI systems. We can be critical consumers of automation, being sure we understand what's being automated, how it was evaluated, and why it's being automated. We can also be critical consumers of journalism about technology, looking for and supporting work that holds power to account. And finally, we can and should engage in ridicule as praxis, meaning having fun pointing out all the ways in which synthetic media extruding machines are janky and tacky.
Yahoo
an hour ago
- Yahoo
WhatsApp 都可以自己寫 AI 機器人?官方確認正在測試
ChatGPT、DeepSeek 的使用介面都有一個共通之處,就是以對話方式呈現,而本來就是通訊軟件的 WhatsApp 也相當有潛質可以成為一個 AI 平台。WABetaInfo 就發現 WhatsApp 正向小部分 beta 測試者開放試用 AI Studio 的新功能,讓用戶可以自行創作 AI 對話機器人。 母公司 Meta 旗下的 Messenger 和 Instagram 其實都已經在海外加入了 AI Studio,也有專屬的網頁版本,不過這次就是給 WhatsApp 這個不一樣的用戶群體也可以使用。在 AI Studio 裡面,用戶可以設計對話機器人的性格、角色,包括是熱情、直接的旅遊顧問;具自信、有魄力的教練,當然背後會是由 Meta AI 模型驅動。 相對於 OpenAI 的 GPTs 和 Google Gemini Gem 服務,WhatsApp 的版本更著重於私人使用,因為使用者需要自行把 AI 機器人的連結分享出去,而不像其他服務會有個市集般的地方讓其他人發掘使用。 目前未知 AI Studio 什麼時候才會正式推出,更未知會否開放到我們本地用戶,但反正 Meta 是在密密加強 WhatsApp 的能力。 更多內容: Introducing GPTs Build custom experts with Gems WhatsApp beta for iOS 25.17.10.75: what's new? WhatsApp to let users build their own AI chatbots to use in the app Perplexity AI 加入 WhatsApp,香港可用、免登入、免 VPN!吉卜力風格、中文對答都可以! 遲到好過冇到!iPad 版 WhatsApp 終於登陸 App Store 緊貼最新科技資訊、網購優惠,追隨 Yahoo Tech 各大社交平台! 🎉📱 Tech Facebook: 🎉📱 Tech Instagram: 🎉📱 Tech WhatsApp 社群: 🎉📱 Tech WhatsApp 頻道: 🎉📱 Tech Telegram 頻道: