
How AI is driving the convergence of networking and security
The impact of AI cannot be overestimated. It is acting like a new kind of gravity, pulling together the previously distinct worlds of networking and security, while reshaping traffic flows across every network layer. We have become accustomed to the changes that AI is bringing about within the data center, and the overheads it is placing on data center infrastructure. But it is now apparent that its influence is being felt much more widely, reshaping requirements across the entire ICT landscape, from the WAN down to the campus.
AI is turning security on its head, believes Mauricio Sanchez, Senior Director of Enterprise Security and Networking Research with independent analyst firm Dell'Oro Group: "AI has been transforming the hacker community, enabling even mediocre players to become super hackers," he believes. "The positive aspects of AI are being turned on their head and used for evil."
Add in the perennial demand of enterprises for five nines of uptime, and a cogent argument is emerging for a convergence of networking and security: "Network operations and security operations can simply no longer exist in parallel silos," notes Sanchez.
Enterprises are all at different stages on this journey, working out how to unite AI with converged networking and security at their own pace. Some are simply applying AI to legacy network and security operations. "They are finding, as they try to move on from this point, that they are getting into what I would call AI-strained infrastructure," says Sanchez. "This is where administrators are sweating bullets, not really empowered or able to react appropriately."
Some progressive companies are in a better place than this when it comes to embracing AI and bringing it to bear on both networking and security, according to Dell'Oro's findings, moving towards what Sanchez describes as the AI augmented-network stage. "Beyond that is a level where both infrastructural change and operational change lead to an AI-empowered network that is being run cleanly and efficiently, able to use AI and serve AI applications well," he concludes. "That is the objective over the next five years for many enterprises."
New challenges, new solutions
To further explore these themes, Sanchez was invited by Unified SASE as a Service provider Aryaka to join a 'fireside chat'. Along with Sanchez was Renuka Nadkarni, Chief Product Officer, Aryaka and Kevin Deierling SVP Marketing, Networking, NVIDIA.
All three agree that AI is changing cybersecurity in many different ways. "AI is using a ton of data, and that creates some opacity," observes NVIDIA's Deierling. "It's hard to see what's happening when AIs are talking to other Ais, and that creates new challenges."
Agility and flexibility are essential responses: "The amount of data that's being created by AI is massive, and the networking performance needed is incredible," he says. "We're shipping 400 gigabit per second networks today, moving to 800 gigabits per second, with 1.6 terabits right around the corner. You can't just statically create a set of rules and hope for the best. It's about being dynamic and responsive in the face of all these new challenges."
Nadkarni of Aryaka believes it all comes down to an age old problem, that of reconciling performance with security: "Back in the day, you had separate networking and security teams making separate decisions," she points out. "The security people were often getting in the way of the business, with frequent conflict between the two. Now our customers are migrating heavily towards a converged networking and security play. And it's not easy, because the whole industry has been divided into networking vendors and security vendors. The whole unified SASE as a service that we are trying to bring to the table was architected from the start to bring things together."
An added challenge, according to Nadkarni, is that AI introduces a certain amount of undeterministic behaviour, both on the networking and the security side: "Customer network architecture and network design used to be about a point to point link," she says. "It was deterministic, because people would typically buy from service providers in increments of 10 Mbps or 100 Mbps, defined as between offices and data centers. But now users are everywhere. Applications are hosted in public clouds, accessed via SaaS. We're seeing a lot of AI applications coming in as SaaS. Traffic patterns have changed drastically, but the need for security is something that hasn't changed."
AI has certainly spelled the end for the static workloads of yesterday where it was easy for networking and security managers to keep tabs on what is happening. In an era of agentic AI workflows, where AIs are talking to Ais, and then interacting with humans, the pace has picked up and complexity can be overwhelming, often at the expense of security.
Deierling says that to assist here, NVIDIA has developed a platform called Morpheus that characterizes behaviours: "It characterizes devices as well as people, and we developed a digital fingerprint using AI to do that," he explains. "We stream the data in real time to these powerful AI engines that can detect anomalous behaviours. If suddenly a human being is firing passwords at the speed of a computer, we can detect that in real time and actually isolate that traffic. We accelerate things with our networking hardware, and we stream telemetry data so we can perform AI very quickly. And then we provide those solutions to partners to build something that customers can use."
Nadkarni says Morpheus works well with Aryaka's platform: "In most security implementations, you take a subset of your traffic and share it with a security vendor," she says. "But if you already knew what traffic needed to be processed you wouldn't need to do that. Because of Morpheus, we have the ability to process all traffic through our system, and we don't have to make choices."
She invites a comparison between the fast-evolving AI we see today and the recent emergence of DevSecOps: "It touches so many aspects of a customer's activities. Many of our customers are telling us they are creating an AI adoption team. We advise them to break down the problem into smaller pieces. Identify all the stakeholders who are accountable for it. For example, who owns the data? Security of data is really important when it comes to AI."
AI is here, it's massive and it's going to transform every industry, concludes Deierling: "Every enterprise should focus on their core expertise, and use AI to accelerate that," he advises. "AI is fast, and it uses huge amounts of data. It's a different type of challenge than what we've seen, but I agree with Renuka that it's an evolution of DevSecOps. Call it AISecOps. You need to protect models, you need to protect data, you need to protect users."
Nadkarni believes we are in 'very exciting times': "We've already seen adoption of cloud, and of different SaaS applications," she says. "AI will have a bigger impact than that. But as Kevin was saying, enterprises should focus on the most important things for their business and then leverage the latest AI technologies that are available, as well as make sure that their network is modernized. It's the industry's job to focus on providing the best technology, offering the best solutions, making it easier to adopt AI and go through the changes that are coming. It's a privilege to be around at this time, seeing all the benefits of this new technology as they unfold."
Sanchez from Dell'Oro Group concludes with the advice that all enterprises go back to basics, focussing on things like visibility: "In order to do that, they need to have the right infrastructure in place, the right foundational elements, because you can't build a strong house on weak foundations," he claims. "Don't try to figure this out for yourself. There are smart people, from companies like NVIDIA and Aryaka, that can help on this journey to make sure that that you don't stumble."
Hashtags

Try Our AI Features
Explore what Daily8 AI can do for you:
Comments
No comments yet...
Related Articles


Techday NZ
12 hours ago
- Techday NZ
Ingram Micro launches Xvantage AI support for partners in ANZ
Ingram Micro has introduced the XvantageTM Enable | AI programme across Australia and New Zealand, aiming to provide structured support for customers and vendors seeking to adopt and scale AI-powered services and solutions. The XvantageTM Enable | AI offering comes as part of Ingram Micro's digital platform and features a combination of assessments, educational resources and vendor-specific enablement to assist technology partners through various stages of artificial intelligence adoption. The initiative is designed to help businesses improve customer experience, productivity and security using AI-enhanced technologies. According to Hope McGarry, Executive Managing Director, Australia at Ingram Micro, the programme is intended to demystify the process of bringing AI into existing operations while supporting partners with hands-on guidance and resources. McGarry stated: "At Ingram Micro, we're proud to support our partners through XvantageTM Enable | AI, a dynamic enablement program designed to accelerate AI adoption and drive real business outcomes. Whether you're just beginning your AI journey or scaling advanced solutions, our platform empowers you with curated learning paths, hands-on use cases, and seamless access to leading technologies. From intelligent automation to AI-enhanced cybersecurity and productivity, we're here to help you take advantage of the full potential of AI." The programme delivers several key resources for partners within the channel ecosystem, including AI Maturity Assessments, which help businesses evaluate their preparedness for adopting AI and uncover skill or capability gaps. Additionally, AI Basecamps offer foundational knowledge to enable effective delivery, and AI Growth Tracks provide step-by-step enablement alongside major technology vendors. Partners also have access to mission-specific data and AI services for secure and relevant data strategies, and immediate access to global centres of excellence for showcasing AI value to clients. Vanessa Sorenson, Chief Partner Officer ANZ at Microsoft, commented on the integration of the programme into partner offerings and its applicability across different experience levels and use cases, stating: "Ingram Micro's XvantageTM Enable | AI takes channel partner enablement to a new level with automated learning journeys and high-impact AI use cases across a variety of domains and lines of business. Regardless of the level of experience, Ingram Micro's program enables partners to deploy AI solutions from workplace productivity with Microsoft 365 Copilot to building AI Agents in Copilot Studio or implementing Intelligent Document Processing with Microsoft Azure AI services." The XvantageTM Enable | AI programme also covers AI-driven security solutions, providing a channel for partners to access, sell and deploy security offerings from leading vendors. Janet Docherty, Director, Channel and Alliances, ANZ, Palo Alto Networks, said: "XvantageTM Enable | AI provides Ingram Micro partners with a streamlined path to successfully sell and implement Palo Alto Networks' comprehensive AI-driven security portfolio. This program is key to empowering partners of all experience levels to maximise their opportunities and drive significant growth in the AI security market." With the broader release of XvantageTM Enable | AI in the Australian and New Zealand markets, Ingram Micro is targeting technology providers across the region, offering tailored engagement regardless of a partner's AI maturity or experience level. The programme is structured to address initial understanding through foundational learning, as well as advanced deployment of AI-enhanced business processes and security models. The initiative underscores the growing emphasis on AI as a driver of productivity, operational efficiency and risk management in the channel sector, with Ingram Micro's platform providing partners access to a blend of technical education, vendor collaboration and direct support in advancing market offerings. Follow us on: Share on:


Techday NZ
12 hours ago
- Techday NZ
CoachHub launches global AI coaching tool AIMY for tailored professional growth
CoachHub has announced the global availability of AIMYTM, its AI-driven coaching solution, now accessible to all customers. AIMYTM has been developed to provide continuous, personalised coaching to employees at all levels within an organisation. The digital tool, based on advanced behavioural science, supports scalable professional development through features such as goal setting, role play and situational coaching. The company stated that AIMYTM offers a 360° approach to coaching, allowing organisations to extend development opportunities to their entire workforce and integrate learning and development (L&D) initiatives seamlessly. The solution can be tailored to suit various organisational requirements and learning objectives, supporting both individual and group development. A key capability of AIMYTM is its customisability. Coachees are able to select the voice, gender, coaching style, and language, with the aim of making sessions more human and engaging. "Offering coachees the ability to select a voice, avatar, language and coaching style makes the entire experience feel more human, relevant and engaging. This builds trust and encourages deeper reflection," says Hilary Aylesworth, Chief Product and Technology Officer at CoachHub. "When coachees can see themselves in the coach, through language, tone, style, and approach, they're more likely to open up, reflect and take action." The launch follows a beta programme involving early adopters including Texas Tech University and Bolton. According to CoachHub, feedback from these organisations has been positive, particularly regarding roleplay-based feedback practice and leadership communication. Analysis of user statistics indicates that 80% of those practising roleplay did so to rehearse delivering feedback. The scenarios addressed were varied: 40% of users practised with conversations directed towards their manager, 32% toward a peer, and 28% toward a direct report. Topics rehearsed most frequently included conflict management (30%), collaboration (27%), and communication (18%). AIMYTM is designed to help users address both immediate workplace challenges and longer-term professional goals. At the beginning of each session, users can choose from different coaching modes, including goal-based coaching, situational coaching and role-play. They also have the option to revisit previous interactions, supporting ongoing development. The tool's applications span a variety of common business needs. These include preparing for performance reviews, giving and receiving feedback, clarifying expectations in new roles, developing workplace communication skills, fostering collaboration, managing tasks and deadlines, handling stress, and adapting to change or new organisational priorities. "AIMYTM is not just an AI assistant; it is a true AI coach. Employees can leverage AIMYTM anytime, anywhere and in the language they need," said Matti Niebelschütz, CEO of CoachHub. "AIMYTM transforms professional development on a global scale by connecting coaching to what matters most - your values and strategic vision. The result: individuals who thrive, organisations that retain their best talent, and sustained, lasting success." A user at Texas Tech University described their experience with the platform, stating, "AIMYTM was easy to use and I had no trouble navigating the application. The clear guidance on its capabilities and limitations was particularly helpful, as it provided a good framework for understanding how it could support me. When I explored the scenario and other goal-related topics, the questions AIMYTM posed were appropriate and effectively prompted me to think through my objectives in greater detail. I could readily see the value it adds." Similarly, a user at Bolton highlighted AIMYTM's role in workplace development, saying, "AIMYTM has been an interesting addition to our development toolkit. Its intuitive design offers quick access to a coaching conversation and an untiring sparring partner for common workplace challenges. It helps structure thoughts during initial goal-setting and problem-solving. Overall, it's an innovative tool to get people's growth journeys started." CoachHub reports that over the past six years, its digital coaching initiatives have served more than 1,000 global organisations and have included a network of over 3,500 business and executive coaches. AIMYTM's development has drawn on this experience, in addition to research partnerships with certified coaches, external experts, CoachHub's Science Council, and the International Coaching Federation Germany Charter Chapter.


Techday NZ
12 hours ago
- Techday NZ
Tracer unveils AI tool to shield brands from ChatGPT fraud
Tracer AI has launched a product designed to monitor and protect brands from fraudulent activities that are perpetrated via AI chatbots, focusing on OpenAI's ChatGPT. The introduction of Tracer Protect for ChatGPT comes as enterprise concern grows about the ways in which generative AI, particularly chatbots, are being exploited to commit brand fraud, distribute counterfeits and conduct executive impersonation scams. Tracer AI's system is built to track mentions of specific brands, products, services, and executives within ChatGPT responses and to proactively identify and mitigate schemes targeting both businesses and consumers. Changing search habits The escalation in generative AI use, including ChatGPT, which reportedly has approximately 400 million weekly active users worldwide, has resulted in consumers rapidly shifting from traditional search engines such as Google to AI chatbots for information and product recommendations. Tracer AI states that this mainstream adoption and the public's general trust in chatbot answers has opened effective new channels for bad actors to carry out brand abuse campaigns. These can manifest as phishing schemes, fraudulent product recommendations, or the promotion of unsafe counterfeit goods. Mechanisms of attack According to Tracer AI, generative engine optimisation (GEO) is being used to promote fraudulent content within chatbot outputs, which often remains hidden from traditional search engines and detection tools. The result is a growing incidence of social engineering, impersonation, and narrative poisoning, in which misleading stories about brands are intentionally promoted to influence consumers and even affect how AI systems respond to future queries regarding those brands. Tracer's approach "OpenAI is already taking important steps to battle nefarious activity in ChatGPT. With Tracer Protect for ChatGPT, Tracer will now be a key pillar of a robust solution to this problem, proactively partnering with brands to remediate infringement activity on OpenAI and any links to websites, mobile apps and marketplaces engineered to prey on consumers," said Rick Farnell, CEO of Tracer. The system's architecture features Flora, Tracer's agentic AI platform, which uses experience-based learning to improve its threat detection and enforcement capabilities over time. The company claims its integration of advanced automation and human analysis allows for continuous monitoring, drastically reducing the time required to identify and respond to brand misuse, and filtering out irrelevant threats. Collaboration and platform integration Tracer Protect for ChatGPT is powered by the Universal AI Platform from Dataiku, aiming to deliver fast and accurate brand threat detection at scale within generative AI environments. "Tracer AI is leading from the front, proving that building and controlling advanced AI agents can deliver a transformative and durable business advantage, including moving towards better genAI control. By combining its Flora agent with the right analytics and its Marlin vision model through The Universal AI Platform, Tracer has shown how to translate frontier AI into real-world outcomes," explained Sophie Dionnet, Senior Vice President of Product and Business Solutions at Dataiku. "This is exactly what Dataiku was built for: enabling visionary teams to create, govern and connect AI agents that solve real-world challenges in entirely new ways while achieving measurable business impact." The product leverages proprietary Human-in-the-Loop AI (HITL) methodologies, combining algorithmic speed with review by expert analysts to ensure that enforcement decisions are legally robust and aligned with client-specific requirements. Industry perspectives "The emergence of AI chatbots as a new vector for brand manipulation is a pressing concern for enterprise organisations," said Sawyer Ramsey, Strategic Account Executive at Snowflake. "Using the Snowflake AI Data Cloud, Tracer's platform is able to achieve incredible response times to provide right-time insights for their customers. The company's proactive approach to monitoring brand reputation in AI outputs represents exactly the kind of forward-thinking protection the cybersecurity industry needs as we navigate our new digital environment and is a testament to the powerful things companies can do on our unified platform." Farnell further emphasised the heightened need for advanced digital brand protection, stating, "The urgency to get ahead of this threat cannot be overstated given the increasing ease with which fraudulent content can be generated and the unprecedented consumer shift from using search engines to now using AI chatbots. Organisations must adopt equally advanced countermeasures to protect their digital presence." "This escalating technological arms race between malicious actors and brand defenders necessitates a proactive and nuanced approach to digital security, which is why we built Tracer Protect for ChatGPT. With our first-of-its-kind brand protection product that actively monitors and analyses ChatGPT outputs to detect and neutralise brand infringements, we're helping enterprises use AI for good and get ahead of these dangerous new threats posing grave risks to brands and their customers." Expansion plans Tracer Protect for ChatGPT is the first of several planned products targeting generative AI ecosystems. Tracer AI intends to extend its monitoring and enforcement solutions to other platforms, including Claude, Perplexity, and Gemini, later in the year. The company highlights its ongoing efforts in proactive brand protection, asserting that live, targeted monitoring of large language model (LLM) outputs represents a shift from reactive to anticipatory security measures. Tracer AI's continued development in this area is marked by its use of its HITL AI, Flora, and Tracer Graph technologies. The firm has recently been recognised within the technology industry, receiving awards for both product development and leadership.