Latest news with #AMDInstinctGPUs


Hamilton Spectator
22-07-2025
- Business
- Hamilton Spectator
Embedded LLM Launches First-of-its-Kind Monetisation Platform for AMD AI GPUs
SINGAPORE, July 21, 2025 (GLOBE NEWSWIRE) — Embedded LLM today announced the global launch of TokenVisor, its monetisation and management platform for Graphics Processing Units (GPUs). The platform was first unveiled alongside AMD at the Advancing AI 2025 conference (Santa Clara, California) held in June. As organizations race to build 'AI factories,' they face a universal challenge: how to translate massive hardware investment into measurable revenue or departmental value. Without turnkey tools for billing, usage tracking, and multi-tenant management, the path from capital expenditure to positive ROI is often slow and complex. TokenVisor addresses this challenge head-on, providing the first such commercialization layer specifically for the growing AMD AI ecosystem. 'TokenVisor brings powerful new capabilities to the AMD GPU neocloud ecosystem, helping providers efficiently manage and monetise LLM workloads,' said Mahesh Balasubramanian, Senior Director of Product Marketing, Data Center GPU Business, AMD. The platform's impact is already being recognized by major industry hardware leaders. 'TokenVisor flips the economics of AI infrastructure,' said Kumar Mitra, general manager and managing director of Lenovo in Greater Asia Pacific. 'By pairing Lenovo ThinkSystem servers with AMD Instinct GPUs and TokenVisor's turnkey monetisation layer, our customers are launching revenue-generating LLM services at unprecedented speed and scale, providing the financial guardrails and chargeback capabilities that CIOs and CFOs require to confidently greenlight AI investments at scale. It's the key to unlocking the full economic potential of the AI factory.' Engineered to simplify operations, TokenVisor enables GPU owners to: 'The spirit of open collaboration we saw at Advancing AI 2025 is what drives us. TokenVisor is the hypervisor for the AI Token era , born from that spirit and engineered with insights from the AMD neocloud community. It's the tool that unlocks the potential of decentralized GPU computing by making it commercially viable for everyone. This launch is a critical milestone, turning the vision of an open AI ecosystem into a reality for the entire AMD community.' Early Adoption Highlights Global Demand Early adopters consistently report that TokenVisor 'took the guesswork out of commercialization.' By providing built-in tools to benchmark capacity, control resource supply, and automate billing, providers can confidently model their business and launch revenue-generating services in a matter of days—a process that previously took months of custom development. The platform's comprehensive support for popular LLM models and responsive technical support are also cited as critical factors in this rapid deployment and return on investment (ROI). For media enquiries, please contact: Jiaqi Lim Head of PR & Marketing pr@ About Embedded LLM Embedded LLM PTE. LTD. creates innovative Large Language Model (LLM) platforms, empowering organisations with generative AI. Inspired by its mission to build essential AI infrastructure for the knowledge economy, the company delivers robust and secure solutions. A significant open-source contributor, notably enhancing vLLM for AMD ROCm, Embedded LLM also offers open source tools like its JamAI Base no-/low-code LLM orchestration platform. The company is committed to making LLM technology accessible and fostering innovation within the open ecosystem. About AMD For more than 55 years, AMD has driven innovation in high-performance computing, graphics, and visualization technologies. Hundreds of millions of consumers, Fortune 500 businesses, and leading scientific research facilities around the world rely on AMD technology to improve how they live, work, and play. AMD employees are focused on building leadership high-performance and adaptive products that push the boundaries of what is possible. For more information about how AMD is enabling today and inspiring tomorrow, visit . About Lenovo Lenovo is a US$69 billion revenue global technology powerhouse, ranked #248 in the Fortune Global 500, and serving millions of customers every day in 180 markets. Focused on a bold vision to deliver Smarter Technology for All, Lenovo has built on its success as the world's largest PC company with a full-stack portfolio of AI-enabled, AI-ready, and AI-optimized devices (PCs, workstations, smartphones, tablets), infrastructure (server, storage, edge, high performance computing and software defined infrastructure), software, solutions, and services. Lenovo's continued investment in world-changing innovation is building a more equitable, trustworthy, and smarter future for everyone, everywhere. Lenovo is listed on the Hong Kong stock exchange under Lenovo Group Limited (HKSE: 992) (ADR: LNVGY). To find out more visit . A photo accompanying this announcement is available at
Yahoo
17-06-2025
- Business
- Yahoo
Watch These AMD Price Levels as Stock Hits 5-Month High Following Last Week's AI Showcase
AMD shares continued gaining ground Tuesday, boosted by upbeat Wall Street commentary following the chipmaker's "Advancing AI" event last week. The stock broke out from a pennant pattern earlier this month and closed above the closely watched 200-day moving average in Monday's trading session. Investors should watch crucial overhead areas on AMD's chart around $145, $160 and $175, while also monitoring support levels near $115 and $ Micro Devices (AMD) shares continued gaining ground Tuesday, boosted by upbeat Wall Street commentary following the chipmaker's "Advancing AI" event last week. Piper Sandler on Monday raised its price target for the stock and expressed enthusiasm for AMD's recently unveiled Helios server rack architecture, which will combine the company's next-generation AMD MI400 chips into one larger system. The investment bank pointed out that the hardware, anticipated for release in 2026, is "pivotal" for the growth of AMD Instinct GPUs. Meanwhile, analysts at Bank of America speculate that the chipmaker could announce Amazon (AMZN) as a partner after the tech giant's cloud unit, Amazon Web Services (AWS), was a key sponsor of last week's event. AMD shares rose 0.6% to around $127 on Tuesday, after surging nearly 8% yesterday to pace S&P 500 advancers. The stock is up 66% from its early-April low, though has gained just 5% since the start of 2025 amid uncertainty over chip export curbs and the company's ability to capture a greater share of the lucrative AI chip market that's now dominated by Nvidia (NVDA). Below, we break down the technicals on AMD's chart and point out crucial price levels worth watching out for. After hitting their May high, AMD shares formed a pennant, a chart pattern that signals a continuation of the stock's uptrend that started in early April. Indeed, the stock broke out from the pattern earlier this month and staged a volume-backed close above the closely watched 200-day moving average in Monday's trading session. Moreover, the relative strength index indicates bullish momentum, generating a reading just below the indicator's overbought threshold. Let's identify three crucial overhead areas on AMD's chart to watch and also point out support levels worth monitoring. Follow-through buying could trigger an initial rally toward $145. This area may provide overhead selling pressure near several peaks and troughs that formed on the chart between April and December last year. A decisive close above this crucial area may see the shares test resistance around $160. Investors could seek to lock in profits in this location near a trendline that connects a range of corresponding trading activity that developed on the chart from April to October last year. The next overhead area to watch sits at $175. The shares may run into sellers in this region near prominent peaks that emerged in May and October last year. During retracements in the stock, it's firstly worth monitoring the $115 level. The stock could encounter support near last week's retest of the pennant pattern's breakout point, which also closely aligns with a range of price action stretching back to mid-January. Finally, selling below this level opens the door for a drop to lower support around $108. Investors could seek to accumulate AMD shares in this location near the low of the pennant pattern and early-February low. The comments, opinions, and analyses expressed on Investopedia are for informational purposes only. Read our warranty and liability disclaimer for more info. As of the date this article was written, the author does not own any of the above securities. Read the original article on Investopedia
Yahoo
16-06-2025
- Business
- Yahoo
AMD Stock Soars as Piper Sandler Raises Price Target After 'Advancing AI' Event
Advanced Micro Devices shares jumped nearly 10% to lead S&P 500 gainers Monday as Piper Sandler raised their price target for the chipmaker's stock. The move comes after AMD's "Advancing AI" event last week, which saw the reveal of next-generation server rack architecture. Bank of America analysts expect a partnership to be announced between AMD and Amazon Web Micro Devices (AMD) shares popped nearly 10% to lead S&P 500 gainers Monday as Piper Sandler analysts raised their price target for the stock coming out of the chipmaker's "Advancing AI" event. Piper raised its target to $140 from $125 and maintained an "overweight" rating for AMD stock. Shares of AMD were at about $127 in recent trading, making Piper's target a roughly 10% premium. The analysts came away "enthused" by the firm's newly unveiled Helios server rack architecture, which it called "pivotal" for the growth of AMD Instinct GPUs. Helios will combine next-generation AMD MI400 chips into one larger system, the company said, and is expected in 2026. AMD highlighted its partnerships with ChatGPT maker OpenAI, Meta Platforms (META), Oracle (ORCL), Microsoft (MSFT), and others at the event. Bank of America analysts believe there's another high-profile partner announcement to come: Amazon (AMZN). Amazon Web Services (AWS) was "a key sponsor for the event," BofA said. However, AWS typically uses its own events to announce new engagements, making a future reveal likely, the bank added. BofA maintained a "buy" rating and price target of $130 following the event. For comparison, the analyst consensus price target from Visible Alpha is about $124. Read the original article on Investopedia Error while retrieving data Sign in to access your portfolio Error while retrieving data Error while retrieving data Error while retrieving data Error while retrieving data


Channel Post MEA
16-06-2025
- Business
- Channel Post MEA
Oracle And AMD Collaborate To Deliver Breakthrough Performance In AI Workloads
Oracle and AMD have announced that AMD Instinct MI355X GPUs will be available on Oracle Cloud Infrastructure (OCI) to give customers more choice and more than 2X better price-performance for large-scale AI training and inference workloads compared to the previous generation. Oracle will offer zettascale AI clusters accelerated by the latest AMD Instinct processors with up to 131,072 MI355X GPUs to enable customers to build, train, and inference AI at scale. 'To support customers that are running the most demanding AI workloads in the cloud, we are dedicated to providing the broadest AI infrastructure offerings,' said Mahesh Thiagarajan, executive vice president, Oracle Cloud Infrastructure. 'AMD Instinct GPUs, paired with OCI's performance, advanced networking, flexibility, security, and scale, will help our customers meet their inference and training needs for AI workloads and new agentic applications.' To support new AI applications that require larger and more complex datasets, customers need AI compute solutions that are specifically designed for large-scale AI training. The zettascale OCI Supercluster with AMD Instinct MI355X GPUs meets this need by providing a high-throughput, ultra-low latency RDMA cluster network architecture for up to 131,072 MI355X GPUs. AMD Instinct MI355X delivers nearly triple the compute power and a 50 percent increase in high-bandwidth memory than the previous generation. 'AMD and Oracle have a shared history of providing customers with open solutions to accommodate high performance, efficiency, and greater system design flexibility,' said Forrest Norrod, executive vice president and general manager, Data Center Solutions Business Group, AMD. 'The latest generation of AMD Instinct GPUs and Pollara NICs on OCI will help support new use cases in inference, fine-tuning, and training, offering more choice to customers as AI adoption grows.' AMD Instinct MI355X Coming to OCI AMD Instinct MI355X-powered shapes are designed with superior value, cloud flexibility, and open-source compatibility—ideal for customers running today's largest language models and AI workloads. With AMD Instinct MI355X on OCI, customers will be able to benefit from: Significant performance boost: Helps customers increase performance for AI deployments with up to 2.8X higher throughput. To enable AI innovation at scale, customers can expect faster results, lower latency, and the ability to run larger AI workloads. Helps customers increase performance for AI deployments with up to 2.8X higher throughput. To enable AI innovation at scale, customers can expect faster results, lower latency, and the ability to run larger AI workloads. Larger, faster memory: Allows customers to execute large models entirely in memory, enhancing inference and training speeds for models that require high memory bandwidth. The new shapes offer 288 gigabytes of high-bandwidth memory 3 (HBM3) and up to eight terabytes per second of memory bandwidth. Allows customers to execute large models entirely in memory, enhancing inference and training speeds for models that require high memory bandwidth. The new shapes offer 288 gigabytes of high-bandwidth memory 3 (HBM3) and up to eight terabytes per second of memory bandwidth. New FP4 support: Allows customers to deploy modern large language and generative AI models cost-effectively with the support of the new 4-bit floating point compute (FP4) standard. This enables ultra-efficient and high-speed inference. Allows customers to deploy modern large language and generative AI models cost-effectively with the support of the new 4-bit floating point compute (FP4) standard. This enables ultra-efficient and high-speed inference. Dense, liquid-cooled design: Enables customers to maximize performance density at 125 kilowatts per rack for demanding AI workloads. With 64 GPUs per rack at 1,400 watts each, customers can expect faster training times with higher throughput and lower latency. Enables customers to maximize performance density at 125 kilowatts per rack for demanding AI workloads. With 64 GPUs per rack at 1,400 watts each, customers can expect faster training times with higher throughput and lower latency. Built for production-scale training and inference: Supports customers deploying new agentic applications with a faster time-to-first token (TTFT) and high tokens-per-second throughput. Customers can expect improved price performance for both training and inference workloads. Supports customers deploying new agentic applications with a faster time-to-first token (TTFT) and high tokens-per-second throughput. Customers can expect improved price performance for both training and inference workloads. Powerful head node: Assists customers in optimizing their GPU performance by enabling efficient job orchestration and data processing with an AMD Turin high-frequency CPU with up to three terabytes of system memory. Assists customers in optimizing their GPU performance by enabling efficient job orchestration and data processing with an AMD Turin high-frequency CPU with up to three terabytes of system memory. Open-source stack: Enables customers to leverage flexible architectures and easily migrate their existing code with no vendor lock-in through AMD ROCm. AMD ROCm is an open software stack that includes popular programming models, tools, compilers, libraries, and runtimes for AI and HPC solution development on AMD GPUs. Enables customers to leverage flexible architectures and easily migrate their existing code with no vendor lock-in through AMD ROCm. AMD ROCm is an open software stack that includes popular programming models, tools, compilers, libraries, and runtimes for AI and HPC solution development on AMD GPUs. Network innovation with AMD Pollara: Provides customers with advanced RoCE functionality that enables innovative network fabric designs. Oracle will be the first to deploy AMD Pollara AI NICs on backend networks, providing advanced RoCE functions such as programmable congestion control and support for open industry standards from the Ultra Ethernet Consortium (UEC) for high-performance and low latency networking.


Techday NZ
13-06-2025
- Business
- Techday NZ
Oracle unveils AMD-powered zettascale AI cluster for OCI cloud
Oracle has announced it will be one of the first hyperscale cloud providers to offer artificial intelligence (AI) supercomputing powered by AMD's Instinct MI355X GPUs on Oracle Cloud Infrastructure (OCI). The forthcoming zettascale AI cluster is designed to scale up to 131,072 MI355X GPUs, specifically architected to support high-performance, production-grade AI training, inference, and new agentic workloads. The cluster is expected to offer over double the price-performance compared to the previous generation of hardware. Expanded AI capabilities The new announcement highlights several key hardware and performance enhancements. The MI355X-powered cluster provides 2.8 times higher throughput for AI workloads. Each GPU features 288 GB of high-bandwidth memory (HBM3) and eight terabytes per second (TB/s) of memory bandwidth, allowing for the execution of larger models entirely in memory and boosting both inference and training speeds. The GPUs also support the FP4 compute standard, a four-bit floating point format that enables more efficient and high-speed inference for large language and generative AI models. The cluster's infrastructure includes dense, liquid-cooled racks, each housing 64 GPUs and consuming up to 125 kilowatts per rack to maximise performance density for demanding AI workloads. This marks the first deployment of AMD's Pollara AI NICs to enhance RDMA networking, offering next-generation high-performance and low-latency connectivity. Mahesh Thiagarajan, Executive Vice President, Oracle Cloud Infrastructure, said: "To support customers that are running the most demanding AI workloads in the cloud, we are dedicated to providing the broadest AI infrastructure offerings. AMD Instinct GPUs, paired with OCI's performance, advanced networking, flexibility, security, and scale, will help our customers meet their inference and training needs for AI workloads and new agentic applications." The zettascale OCI Supercluster with AMD Instinct MI355X GPUs delivers a high-throughput, ultra-low latency RDMA cluster network architecture for up to 131,072 MI355X GPUs. AMD claims the MI355X provides almost three times the compute power and a 50 percent increase in high-bandwidth memory over its predecessor. Performance and flexibility Forrest Norrod, Executive Vice President and General Manager, Data Center Solutions Business Group, AMD, commented on the partnership, stating: "AMD and Oracle have a shared history of providing customers with open solutions to accommodate high performance, efficiency, and greater system design flexibility. The latest generation of AMD Instinct GPUs and Pollara NICs on OCI will help support new use cases in inference, fine-tuning, and training, offering more choice to customers as AI adoption grows." The Oracle platform aims to support customers running the largest language models and diverse AI workloads. OCI users leveraging the MI355X-powered shapes can expect significant performance increases—up to 2.8 times greater throughput—resulting in faster results, lower latency, and the capability to run larger models. AMD's Instinct MI355X provides customers with substantial memory and bandwidth enhancements, which are designed to enable both fast training and efficient inference for demanding AI applications. The new support for the FP4 format allows for cost-effective deployment of modern AI models, enhancing speed and reducing hardware requirements. The dense, liquid-cooled infrastructure supports 64 GPUs per rack, each operating at up to 1,400 watts, and is engineered to optimise training times and throughput while reducing latency. A powerful head node, equipped with an AMD Turin high-frequency CPU and up to 3 TB of system memory, is included to help users maximise GPU performance via efficient job orchestration and data processing. Open-source and network advances AMD emphasises broad compatibility and customer flexibility through the inclusion of its open-source ROCm stack. This allows customers to use flexible architectures and reuse existing code without vendor lock-in, with ROCm encompassing popular programming models, tools, compilers, libraries, and runtimes for AI and high-performance computing development on AMD hardware. Network infrastructure for the new supercluster will feature AMD's Pollara AI NICs that provide advanced RDMA over Converged Ethernet (RoCE) features, programmable congestion control, and support for open standards from the Ultra Ethernet Consortium to facilitate low-latency, high-performance connectivity among large numbers of GPUs. The new Oracle-AMD collaboration is expected to provide organisations with enhanced capacity to run complex AI models, speed up inference times, and scale up production-grade AI workloads economically and efficiently.