
The ROI Trap: How Marketing Metrics Can Mislead The C-Suite
Imagine walking through a retail district where every screen reacts to your presence, delivering messages tailored just for you.
That's the world Minority Report envisioned two decades ago. It felt futuristic then. Today, it's not far from reality as consumers seamlessly hop across physical and digital environments: see a brand on a digital billboard, check reviews on Instagram and finally purchase after a Google search.
The journey is far from linear. It's multi-touch, multi-screen and multi-moment. Yet, despite this evolution, I see our marketing measurement frameworks as stuck in the past.
We're Optimizing What's Measurable, Not What's Effective
Modern marketers are under immense pressure to prove performance. Understandably, most rely on metrics that are easy to track: clicks, impressions, views, conversions.
But these are often surface-level indicators. This has led to what I call the ROI trap, where attribution is disproportionately skewed to the last click or the final touchpoint. Channels like search, retargeting and programmatic display appear to outperform others not because they're more impactful, but because they're easier to measure.
Here's the reality: A consumer may have seen your brand five times: on a billboard, YouTube ad, podcast mention, influencer's story and finally a search result. The last one gets all the credit, doesn't it? But what about the brand lift? The emotional nudge? The decision-making groundwork that took place earlier? We're rewarding closers instead of recognizing the entire sales team.
I think marketing ROI today is broken, not because we lack data, but because we measure what's convenient instead of what's consequential.
The Old World: Predictable Planning, Coherent Strategy
For decades, media planning revolved around reach and frequency models. Marketers built annual plans and allocated budgets across TV, radio, print and outdoor based on marketing mix models (MMM). It wasn't perfect, but it worked. Audiences were easier to define, media was finite and stable, and ROI could be projected with relative confidence.
Today, those guardrails are gone. We've traded predictability for precision and lost perspective along the way.
The Modern Challenge: Fragmentation, Siloed Metrics And Misleading Signals
We now operate in an ecosystem dominated by platforms like TikTok, YouTube, CTV, Instagram, programmatic digital out-of-home (pDOOH) and more. Each comes with its own performance dashboards: YouTube prioritizes watch time, TikTok counts views, our displays track click-through rates, and social media chases engagement.
Although each channel is optimized in isolation, the consumer moves fluidly across them all. This siloed optimization results in inefficiency, overlapping spend and ultimately, underperformance. Marketers are stuck chasing platform KPIs while missing the holistic journey that actually drives purchase behavior.
A Personal Turning Point: When The Screen Started Listening
A few years ago, I encountered a digital billboard that allowed me to change its content using my phone. It was simple but profound. That moment rewired how I saw advertising. If a screen can respond, it can track. If it can track, it can measure.
That realization formed the foundation of my media technology company. I think it's important to find ways to bridge the physical and digital media gap and allow brands to plan, automate and measure across every environment consumers touch.
Toward A Smarter ROI: Three Shifts That Matter
To evolve beyond vanity metrics, marketing leaders must embrace a new ROI model, one that reflects causality, journey orchestration and adaptability.
Don't ask "What ad drove the conversion?" Instead, seek: "What impact did the campaign create that wouldn't have happened otherwise?" Incrementality isolates true lift, helping marketers quantify what exposure really moved the needle.
Marketing is now a matter of choreography. Out-of-home creates awareness. Social drives validation. Search captures intent. Understanding the order in which consumers encounter messages is key to unlocking influence, with the moment of influence often more powerful than the medium itself. Therefore, look to prioritize message flow over channel focus.
AI is transforming media from a static buy to a dynamic design. By analyzing audience behavior in real time, AI can help recommend which message should follow which, where it should appear and how the creatives will convert. Historical planning still has a place, but this should take a back seat to real-time learning.
A New Formula For Marketing Success
Let's move beyond outdated ROI definitions. Instead of just dividing return by spend, consider a smarter equation where ROI equals incremental lift multiplied by optimal sequence and learning velocity. To operationalize this, business leaders need to connect data across all channels (digital and offline), break down internal media silos and link media goals to business outcomes (over vanity metrics).
The Future Of ROI: From Impressions To Intelligence
Every screen is now a performance medium, from smart TVs to digital billboards to checkout kiosks. But performance won't be unlocked through only impressions or cheaper clicks.
I think we need to stop asking "Which platform worked?" and start asking "What journey delivered value?" The future belongs to brands that can answer this question and orchestrate conversations across every moment, touchpoint and screen.
Forbes Business Council is the foremost growth and networking organization for business owners and leaders. Do I qualify?
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles
Yahoo
10 minutes ago
- Yahoo
Knox Now FedRAMP High Listed, Unlocking Secure AI and Cloud for U.S. Government
WASHINGTON and NEW YORK, Aug. 18, 2025 /PRNewswire/ -- Knox Systems today announced it is now FedRAMP High listed, opening the door for commercial SaaS and AI platforms to securely serve the federal government in just 90 days. Modernization as National SecurityAmerica's most critical systems still run on outdated infrastructure, leaving them vulnerable to increasingly sophisticated foreign cyber threats. FedRAMP High makes the Knox Cloud a trusted foundation for agencies to modernize securely, with AI-driven compliance and real-time threat monitoring baked in. Secure AI and Cloud for the Federal Government and DoD Without the BureaucracyAI is only as powerful as the infrastructure that runs it. By eliminating years of red tape, Knox allows agencies to adopt AI and cloud tools at mission speed - faster, cheaper, and safer than ever before. "FedRAMP High isn't just compliance - it's the key to unlocking secure innovation for government," said Irina Denisenko, CEO of Knox Systems. "With Knox, agencies finally get the guardrails they need and the speed they've been missing." About Knox SystemsKnox ( operates the largest federal managed cloud, trusted by top agencies and partners across defense and civilian sectors. Built for speed, resilience, and compliance, Knox delivers FedRAMP authorization in 90 days - turning the biggest bottleneck in government IT into the fastest path to modernization. For media inquiries contact media@ View original content to download multimedia: SOURCE Knox Systems, Inc Error in retrieving data Sign in to access your portfolio Error in retrieving data Error in retrieving data Error in retrieving data Error in retrieving data


Tom's Guide
11 minutes ago
- Tom's Guide
Forget AirPods — Samsung's new earbuds put Gemini and live translation in your ears for just $149
While Apple struggles to bring us the fully upgraded Siri we were promised — wait until 2026! — Samsung just got the jump on Apple with a pair of $149 earbuds that feature Gemini and live translation integration. Even worse for Apple, it's at a price that beats out the ANC-equipped AirPods 4, rather than the flagship AirPods Pro 2. The Samsung Galaxy Buds3 FE) offer a similar style to the Galaxy Buds 3 Pro but undercut them by $100. And they might just be the affordable earbuds that you've been waiting for. The Buds3 FE look very similar to their more expensive cousins, all the way down to an almost identical shape and charging case with a clear lid. They're not quite as shiny, though, coming in just Black or Gray. But they pack in some serious specs, especially for their relatively affordable price. There's spatial audio on board, called "360 audio." That feature makes your music sound like it's coming from all around you, as opposed to just the left and right. On top of that, there's improved ANC over the previous model, along with better call quality thanks to "a machine learning model" that's trained to isolate your voice. You'll remain clear to your call partner in even busy and noisy environments — always a bonus. Sound quality is apparently going to be "rich, powerful," while delivering "deep bass and clear treble." Those are all things I like my earbuds to do, but I'll have to get them in my ears before I can tick the "sound quality" box in my review. The Galaxy Buds3 FE feature a voice assistant through Gemini, so you don't need to pull your phone out of your pocket to ask the assistant a question or get help. So you can say something like "Hey, Google. Start a 20-minute yoga workout." Then there's live, conversational translation. That's a big one. You can have a chat with someone speaking a different language, and understand what they're saying as the buds translate what they're saying. That's something the AirPods Pro 3 are pegged to have, so it's interesting to see this on something much cheaper. Get instant access to breaking news, the hottest reviews, great deals and helpful tips. Finally, there are the expected specs. We've got 6 hours of ANC on time in the buds, with 24 hours in the case. That's nothing remarkable. Finally, there's Bluetooth 5.4 for a solid connection and AAC and SBC codecs. That's slightly disappointing — LDAC or AptX would have been nice to see. IP54 water resistance is going to serve fitness fanatics well, but there are better options out there if you're looking for a pair of the best earbuds for running. The Galaxy Buds3 FE are coming out September 4, 2025. Stay tuned for our full review. Follow Tom's Guide on Google News to get our up-to-date news, how-tos, and reviews in your feeds. Make sure to click the Follow button.


CNET
11 minutes ago
- CNET
Google's AI Filmmaker Program Flow Helped Creators Make 100 Million Videos
Google's generative AI filmmaking program Flow has reached a milestone. The tech giant confirmed exclusively to CNET that Flow creators have produced over 100 million AI videos in the program. Thanks in part to its advanced AI video model, Veo 3, Flow allows users to generate video clips and edit them together to create scenes. It's been 90 days since Google surprised us with Flow at its annual I/O developers conference. According to Elias Roman, senior director of product management for Flow in Google Labs, much of the time since has been spent "hustling just to keep up with the demand." Flow is a departure from Google's previous generative AI work. For years, the company's AI efforts have been focused on Gemini, its all-in-one chatbot. It has flooded its products with AI, like with Search's AI overviews and Gmail's AI-generated summaries. Its research assistant tool, NotebookLM, with its AI audio generator that can transform documents into personal podcasts, continually rolls out new features. The industry leader has spent billions of dollars trying to win the race to develop the most advanced AI for average Google searchers, developers and, yes, even artists and creators. 100 million AI videos is a significant milestone for the company, and it helps show us what the future of AI-enabled creation might look like. Getting in the AI Flow To compete with Midjourney and Stable Diffusion, Google created a crop of AI image models, originally named ImageFX and now known as Imagen (pronounced "imagine"). Its previous generative media models were better suited for amateur or enthusiast creators, not professionals, and they didn't dominate the AI creative space. That all changed with Veo 3. Google dropped Veo 3, its newest AI video model, at May's I/O conference. Veo 3 leapfrogged the competition with a somewhat obvious but first-in-industry advancement: AI videos with synchronized, AI-generated audio. The model garnered a ton of attention online, and Google reported over 40 million AI videos just seven weeks later. "What Veo 3 allowed was a much wider set of people to create very compelling videos, engaging all the senses out of the box. You didn't have to stitch together a toolkit," said Roman. "To be able to do the Foley [ambient sounds], the sound effects, the soundtrack, the dialog, all of that, and not make the user think about each of those modalities in a specific way, I think, is a big unlock, too." My Attempt at AI-Created Irish Dancers with Google Flow My Attempt at AI-Created Irish Dancers with Google Flow Click to unmute Video Player is loading. Play Video Pause Skip Backward Skip Forward Next playlist item Unmute Current Time 0:00 / Duration 0:12 Loaded : 47.46% 0:00 Stream Type LIVE Seek to live, currently behind live LIVE Remaining Time - 0:12 Share Fullscreen This is a modal window. Beginning of dialog window. Escape will cancel and close the window. Text Color White Black Red Green Blue Yellow Magenta Cyan Opacity Opaque Semi-Transparent Text Background Color Black White Red Green Blue Yellow Magenta Cyan Opacity Opaque Semi-Transparent Transparent Caption Area Background Color Black White Red Green Blue Yellow Magenta Cyan Opacity Transparent Semi-Transparent Opaque Font Size 50% 75% 100% 125% 150% 175% 200% 300% 400% Text Edge Style None Raised Depressed Uniform Drop shadow Font Family Proportional Sans-Serif Monospace Sans-Serif Proportional Serif Monospace Serif Casual Script Small Caps Reset Done Close Modal Dialog End of dialog window. Close Modal Dialog This is a modal window. This modal can be closed by pressing the Escape key or activating the close button. Close Modal Dialog This is a modal window. This modal can be closed by pressing the Escape key or activating the close button. My Attempt at AI-Created Irish Dancers with Google Flow Although you can use Veo 3 in Flow, it's an entirely separate program, built for professional creators and filmmakers. Google intentionally moved away from the ImageFX anchor, Roman said, and wanted to combine the more advanced Imagen and Veo models with Gemini, which was used in the training of Veo and "basically speaks native Veo." Flow is one way to combine all those AI models and pieces, uniting Google's different generative AI tools for seamless video creation and editing. What makes Flow different from Veo and Imagen Flow was built to focus on consistency, that is, the ability to maintain visual identity from one clip to the next. If you have a 90-second video of your character drinking coffee in a cafe, you don't want their hair length or eye color changing every 8 seconds between scenes. That consistency is important for professional projects, and it's also difficult to attain. Roman called it the "Achilles heel of AI video." Flow has several tools to help you maintain that consistency, and in my testing, they do give you a new level of control over your work that was previously lacking from Google AI tools. The best way I can describe Flow is an upscaled version of simple video generator interfaces, with the option to export multiple clips into a simplified version of a Premiere Pro-like timeline. AI tools often get upgraded with the hope that they'll become more useful for professional creators, though the target audience isn't automatically drawn to using them. Generative AI is a contentious issue in creative industries, especially when it comes to wholesale creation of text, image and video. AI enthusiasts might laud the creativity and speed of AI models, but creators continue to voice legitimate concerns about how AI is trained and deployed. It's why publishers and artists have filed lawsuits against AI companies alleging copyright infringement. It's why workers in data-rich industries face job security concerns as executives look to cut costs. Whatever Google claims in its privacy policies, AI is also under fire for the type of imagery it's empowered to create. Last year, users found Gemini could produce images of people of color in Nazi soldier uniforms. Google apologized for what the company called "inaccuracies in some historical image generation depictions" and said it was working to improve those depictions immediately. (Google's guidelines prohibit the creation of abusive and illegal AI content. Roman said that improving the enforcement of its safety policies is aided by technological updates and real-life usage and reports.) Going forward, Roman said Flow is working on expanding Veo 3's capabilities, improving consistency, and adding new features like bespoke voices for character work. The project's north star is making creation more accessible to people. "We can lower the barriers that prevent a much wider set of people from telling stories through video, and we can raise the ceiling on what kind of stories can be told through video," said Roman. "Some of them are going to be funny and silly, like wild street interviews or Yeti ASMR bloggers, and some of them are gonna be really powerful." How to use Google's Flow for AI videos Flow, which is part of Google Labs and accessible via its AI Test Kitchen, is available to paying Google AI subscribers in its $20 per month Pro plan and $250 per month Ultra plan (currently discounted for $125 for three months). Google Labs's privacy notice says that "human reviewers read, annotate and process" your Labs interactions and tool outputs to improve its AI models. (Your Labs data is stored for up to 18 months by default, and the company advises you not to upload or submit confidential information. Google's general privacy center has more info.) I spent some time testing Flow, generating clips and stitching them together using its scenebuilder. Several tools are only available to Flow users. Ingredients-to-video: There are a couple of ways you can prompt to generate video clips, including the self-explanatory text-to-video and image/frames-to-video. Ingredients-to-video is a new one worth exploring. With this method, you upload specific pictures and add a text prompt, and Flow will piecemeal the parts together. For example, you can upload a picture of a man, a product photo of a specific jacket, and a scenic background, and then Flow can combine them and animate the video. Extending clips and smoothing transitions: Extend can help you lengthen clips. In the scenebuilder timeline, drag the end of one clip's frame out to your desired length. If you're going to generate a new video and want a smooth transition, I recommend going to the end of the first clip and hitting the plus button at the top of the marker to save the final frame to your library. You can then use that image in a frames-to-video prompt to maintain that consistency from clip to clip. Doodling and making edits: If you're editing a frame or image in a separate document, you can upload your marked-up image to Flow and instruct the model to implement the changes. You can also do that with images you've drawn on, and it can bring those doodles to life. This is a developing feature -- a new prototype for this is in the works now -- but it's definitely fun to stretch Flow's capabilities like that. Prompting with Gemini: There's no way to have Gemini automatically create and/or improve your prompts directly in Flow (something I hope changes in a future update), but you can use the chatbot to help you craft the perfect prompt. If you're struggling to bring more detail-oriented ideas to life, try letting Gemini help you out. For more, check out the top AI image generators and a guide to writing the best AI image prompts.