logo
AMD brings Stable Diffusion 3.0 Medium AI to Ryzen laptops

AMD brings Stable Diffusion 3.0 Medium AI to Ryzen laptops

Techday NZ24-07-2025
AMD has launched the first BF16 Stable Diffusion 3.0 Medium model for local image generation, optimised for AMD XDNA 2 NPUs and available within the Amuse 3.1 release.
The new model, created in collaboration with Stability AI, supports high-quality image generation while reducing memory usage, enabling operation on laptops with as little as 24GB of RAM by consuming just 9GB for execution. This development means that users can run advanced AI image tasks on consumer devices equipped with Ryzen AI 300 Series or Ryzen AI Max processors with more than 50 NPU TOPS.
Technical details
According to AMD, the BF16 precision format, also referenced as block FP16, blends the accuracy characteristic of FP16 with the computational performance of INT8. This combination supports more sophisticated generative AI workloads while maintaining efficiency within hardware capabilities. The Amuse 3.1 software release, which enables access to this model, further extends these benefits to compatible consumer hardware.
The model features a 2-stage processing pipeline utilising the XDNA 2 NPU, upscaling initial 2MP (1024x1024) images to a 4MP (2048x2048) resolution. This enhancement, as stated, is intended to deliver "print quality images tailored to your specifications on-the-go."
AMD provided further details about practical usage and potential applications, referencing capabilities for graphics design and rapid creation of custom marketing assets. These tasks can be accomplished offline, without the need for a persistent Wi-Fi connection or recurring subscription fees for image generation.
Usage and compatibility
The new BF16 SD 3.0 Medium model is designed to function on systems featuring Ryzen AI 300 series or Ryzen AI Max processors, each paired with an AMD XDNA 2 NPU delivering at least 50 TOPS of performance. AMD highlighted that this enables high precision AI image generation to be accessible on a wider range of laptop devices than previously possible, without requiring machines equipped with 32GB or more RAM.
AMD explained how users can access the new capabilities: "Try SD 3.0 Medium (with BF16 precision) right now on an AMD Ryzen AI 300 series or Ryzen AI MAX+ laptop with at least 24GB memory right now by following three simple steps: SD3 Medium (NPU) Hardware Requirements: AMD Ryzen AI 300 series or Ryzen AI MAX+ laptop equipped with a 50 TOPs or higher AMD XDNA 2 NPU and at least 24GB of system RAM. Download and Install the latest Adrenalin Driver. Download and Install Amuse 3.1 Beta. In EZ Mode, move the slider all the way to HQ and toggle 'XDNA 2 Stable Diffusion Offload'."
Image generation and prompting
AMD noted that prompt engineering remains crucial for optimal results with Stable Diffusion 3.0 Medium. They advised: "Stable Diffusion 3.0 Medium is an extremely capable model that is very sensitive to the prompt content, structure and order. Here are some prompting tips in order to get the best quality: The same size, seed, steps, sampler, scheduler and model combo should yield the same image. You generally want to start by describing the type of image then the structural components of an image and then transition to details and other context. Not every seed will yield a perfect result. Typically, you want to iterate your prompt till you get to the visual structure you want (typos are fine) and then automate a batch with 25-30 seeds. Even spaces and full stops make a difference. You can utilize negative prompts to remove elements you don't want from an image but doing this excessively will have a quality impact on the image."
The company also supplied sample prompts, seeds, and settings that allow users to recreate test images, including detailed instructions for users aiming to replicate specific outputs.
Further guidance and considerations
AMD reminded users that Amuse 3.1 is classified as beta software, supplied by a third-party provider, and may show instability or bugs. They also commented on applicable licencing: "Image generation using SD 3 Medium is regulated by the Stability AI Community Licence and is free for personal use and for SMEs under $1 million in annual revenue. Licensing requirements may change at the sole discretion of the third party. For complete licensing requirements please refer to: LICENSE.md · stabilityai/stable-diffusion-3-medium at main. Internet connection is required to download the model and other configuration files."
The company clarified feature availability based on processor compatibility, noting that AMD Ryzen AI includes a combination of an AI engine, Radeon graphics engine, and Ryzen processor cores subject to enablement from both OEM and ISV partners. They advised customers to check system compatibility prior to purchase. At Computex 2024, AMD introduced the world's first block FP16 stable diffusion model: the SDXL Turbo. The breakthrough model combined the accuracy of FP16 with the performance of INT8 and was a collaboration between AMD and Stability AI.
The BF16-enabled SD 3.0 Medium model is intended to address the expanding demand for on-device AI capability while managing the memory constraints present in most mainstream laptops, and is available immediately with the release of Amuse 3.1 for supported hardware.
Orange background

Try Our AI Features

Explore what Daily8 AI can do for you:

Comments

No comments yet...

Related Articles

AMD brings 128B LLMs to Windows PCs with Ryzen AI Max+ 395
AMD brings 128B LLMs to Windows PCs with Ryzen AI Max+ 395

Techday NZ

time6 days ago

  • Techday NZ

AMD brings 128B LLMs to Windows PCs with Ryzen AI Max+ 395

AMD has announced a free software update enabling 128 billion parameter Large Language Models (LLMs) to be run locally on Windows PCs powered by AMD Ryzen AI Max+ 395 128GB processors, a capability previously only accessible through cloud infrastructure. With this update, AMD is allowing users to access and deploy advanced AI models locally, bypassing the need for third-party infrastructure, which can provide greater control, lower ongoing costs, and improved privacy. The company says this shift addresses growing demand for scalable and private AI processing at the client device level. Previously, models of this scale, such as those approaching the size of ChatGPT 3.0, were operable only within large-scale data centres. The new functionality comes through an upgrade to AMD Variable Graphics Memory, included with the upcoming Adrenalin Edition 25.8.1 WHQL drivers. This upgrade leverages the 96GB Variable Graphics Memory available on the Ryzen AI Max+ 395 128GB machine, supporting the execution of memory-intensive LLM workloads directly on Windows PCs. A broader deployment This update also marks the AMD Ryzen AI Max+ 395 (128GB) as the first Windows AI PC processor to run Meta's Llama 4 Scout 109B model - specifically with full vision and multi-call processing (MCP) support. The processor can manage all 109 billion parameters in memory, although the mixture-of-experts (MoE) architecture means only 17 billion parameters are active at any given time. The company reports output rates of up to 15 tokens per second for this model. According to AMD, the ability to handle such large models locally is important for users who require high-capacity AI assistants on-the-go. The system also supports flexible quantisation and can facilitate a range of LLMs, from compact 1B parameter models to Mistral Large, using the GGUF format. This isn't just about bringing cloud-scale compute to the desktop; it's about expanding the range of options for how AI can be used, built, and deployed locally. The company further states that performance in MoE models like Llama 4 Scout correlates with the number of active parameters, while dense models depend on the total parameter count. The memory capacity of the AMD Ryzen AI Max+ platform allows users to opt for higher-precision models, supporting up to 16-bit models through when trade-offs between quality and performance are warranted. Context and workflow AMD also highlights the importance of context size when working with LLMs. The AMD Ryzen AI Max+ 395 (128GB), equipped with the new driver, can run Meta's Llama 4 Scout at a context length of 256,000 (with Flash Attention ON and KV Cache Q8), significantly exceeding the standard 4,096 token window default in many applications. Examples provided include demonstrations where an LLM summarises extensive documents, such as an SEC EDGAR filing, requiring over 19,000 tokens to be held in context. Another example cited the summarisation of a research paper from the ARXIV database, needing more than 21,000 tokens from query initiation to final output. AMD notes that more complex workflows might require even greater context capacity, particularly for multi-tool and agentic scenarios. AMD states that while occasional users may manage with a context length of 32,000 tokens and a lightweight model, more demanding use cases will benefit from hardware and software that support expansive contexts, as offered by the AMD Ryzen AI Max+ 395 128GB. Looking ahead, AMD points to an expanding set of agentic workflows as LLMs and AI agents become more widely adopted for local inferencing. Industry trends indicate that model developers, including Meta, Google, and Mistral, are increasingly integrating tool-calling capabilities into their training runs to facilitate local personal assistant use cases. AMD also provides guidance on maintaining caution when enabling tool access for large language models, noting the potential for unpredictable system behaviour and outcomes. Users are advised to install LLM implementations only from trusted sources. The AMD Ryzen AI Max+ 395 (128GB) is now positioned to support most models available through and other tools, offering flexible deployment and model selection options for users with high-performance local AI requirements.

AMD brings Stable Diffusion 3.0 Medium AI to Ryzen laptops
AMD brings Stable Diffusion 3.0 Medium AI to Ryzen laptops

Techday NZ

time24-07-2025

  • Techday NZ

AMD brings Stable Diffusion 3.0 Medium AI to Ryzen laptops

AMD has launched the first BF16 Stable Diffusion 3.0 Medium model for local image generation, optimised for AMD XDNA 2 NPUs and available within the Amuse 3.1 release. The new model, created in collaboration with Stability AI, supports high-quality image generation while reducing memory usage, enabling operation on laptops with as little as 24GB of RAM by consuming just 9GB for execution. This development means that users can run advanced AI image tasks on consumer devices equipped with Ryzen AI 300 Series or Ryzen AI Max processors with more than 50 NPU TOPS. Technical details According to AMD, the BF16 precision format, also referenced as block FP16, blends the accuracy characteristic of FP16 with the computational performance of INT8. This combination supports more sophisticated generative AI workloads while maintaining efficiency within hardware capabilities. The Amuse 3.1 software release, which enables access to this model, further extends these benefits to compatible consumer hardware. The model features a 2-stage processing pipeline utilising the XDNA 2 NPU, upscaling initial 2MP (1024x1024) images to a 4MP (2048x2048) resolution. This enhancement, as stated, is intended to deliver "print quality images tailored to your specifications on-the-go." AMD provided further details about practical usage and potential applications, referencing capabilities for graphics design and rapid creation of custom marketing assets. These tasks can be accomplished offline, without the need for a persistent Wi-Fi connection or recurring subscription fees for image generation. Usage and compatibility The new BF16 SD 3.0 Medium model is designed to function on systems featuring Ryzen AI 300 series or Ryzen AI Max processors, each paired with an AMD XDNA 2 NPU delivering at least 50 TOPS of performance. AMD highlighted that this enables high precision AI image generation to be accessible on a wider range of laptop devices than previously possible, without requiring machines equipped with 32GB or more RAM. AMD explained how users can access the new capabilities: "Try SD 3.0 Medium (with BF16 precision) right now on an AMD Ryzen AI 300 series or Ryzen AI MAX+ laptop with at least 24GB memory right now by following three simple steps: SD3 Medium (NPU) Hardware Requirements: AMD Ryzen AI 300 series or Ryzen AI MAX+ laptop equipped with a 50 TOPs or higher AMD XDNA 2 NPU and at least 24GB of system RAM. Download and Install the latest Adrenalin Driver. Download and Install Amuse 3.1 Beta. In EZ Mode, move the slider all the way to HQ and toggle 'XDNA 2 Stable Diffusion Offload'." Image generation and prompting AMD noted that prompt engineering remains crucial for optimal results with Stable Diffusion 3.0 Medium. They advised: "Stable Diffusion 3.0 Medium is an extremely capable model that is very sensitive to the prompt content, structure and order. Here are some prompting tips in order to get the best quality: The same size, seed, steps, sampler, scheduler and model combo should yield the same image. You generally want to start by describing the type of image then the structural components of an image and then transition to details and other context. Not every seed will yield a perfect result. Typically, you want to iterate your prompt till you get to the visual structure you want (typos are fine) and then automate a batch with 25-30 seeds. Even spaces and full stops make a difference. You can utilize negative prompts to remove elements you don't want from an image but doing this excessively will have a quality impact on the image." The company also supplied sample prompts, seeds, and settings that allow users to recreate test images, including detailed instructions for users aiming to replicate specific outputs. Further guidance and considerations AMD reminded users that Amuse 3.1 is classified as beta software, supplied by a third-party provider, and may show instability or bugs. They also commented on applicable licencing: "Image generation using SD 3 Medium is regulated by the Stability AI Community Licence and is free for personal use and for SMEs under $1 million in annual revenue. Licensing requirements may change at the sole discretion of the third party. For complete licensing requirements please refer to: · stabilityai/stable-diffusion-3-medium at main. Internet connection is required to download the model and other configuration files." The company clarified feature availability based on processor compatibility, noting that AMD Ryzen AI includes a combination of an AI engine, Radeon graphics engine, and Ryzen processor cores subject to enablement from both OEM and ISV partners. They advised customers to check system compatibility prior to purchase. At Computex 2024, AMD introduced the world's first block FP16 stable diffusion model: the SDXL Turbo. The breakthrough model combined the accuracy of FP16 with the performance of INT8 and was a collaboration between AMD and Stability AI. The BF16-enabled SD 3.0 Medium model is intended to address the expanding demand for on-device AI capability while managing the memory constraints present in most mainstream laptops, and is available immediately with the release of Amuse 3.1 for supported hardware.

Vultr launches early access to AMD Instinct MI355X GPU for AI
Vultr launches early access to AMD Instinct MI355X GPU for AI

Techday NZ

time18-06-2025

  • Techday NZ

Vultr launches early access to AMD Instinct MI355X GPU for AI

Vultr has announced the availability of the AMD Instinct MI355X GPU as part of its cloud infrastructure services. As one of the first cloud providers to integrate the new AMD Instinct MI355X GPU, Vultr is now taking pre-orders for early access, with global availability scheduled for the third quarter of the year. The GPU forms part of AMD's latest focus on high-capacity computational demands, catering to artificial intelligence (AI) workloads as well as enterprise-scale applications. Product features The AMD Instinct MI355X GPU is based on AMD's 4th Generation CDNA architecture. According to Vultr, this GPU features 288 GB of HBM3E memory, delivers up to 8 TB/s of memory bandwidth, and supports expanded datatypes such as FP6 and FP4. These improvements are designed to address complex tasks ranging from AI training and inference to scientific simulations within high-performance computing (HPC) environments. For customers operating within higher-density data environments, the Instinct MI355X supports direct liquid cooling (DLC). This enhancement offers increased thermal efficiency, which is intended to unlock greater computing performance per rack and facilitate advanced, scalable cooling strategies. The GPU is also supported by the latest version of AMD's ROCm software, which further optimises tasks related to AI inference, training, and compatibility with various frameworks. This results in improved throughput and reduced latency for critical operations. AMD and Vultr partnership Vultr's portfolio already includes other AMD offerings, such as the AMD EPYC 9004 Series and EPYC 7003 Series central processing units (CPUs), as well as previous GPU models like the Instinct MI325X and MI300X. Customers using the MI355X in combination with AMD EPYC 4005 Series CPUs will benefit from a fully supported computing stack across both processing and acceleration functions, streamlining high-powered workloads from end to end. Negin Oliver, Corporate Vice President of Business Development, Data Centre GPU Business at AMD, stated: "AMD is the trusted AI solutions provider of choice, enabling customers to tackle the most ambitious AI initiatives, from building large-scale AI cloud deployments to accelerating AI-powered scientific discovery. AMD Instinct MI350 series GPUs paired with AMD ROCm software provide the performance, flexibility, and security needed to deliver tailored AI solutions that meet the diverse demands of the modern AI landscape." The collaboration builds on Vultr's efforts to support a range of AMD solutions tailored for enterprise, HPC, and AI sectors, reinforcing the company's capacity to cater to evolving customer workloads. Cloud market implications J.J. Kardwell, Chief Executive Officer of Vultr, highlighted the alignment of the new GPU with market requirements. Kardwell commented: "AMD MI355X GPUs are designed to meet the diverse and complex demands of today's AI workloads, delivering exceptional value and flexibility. As AI development continues to accelerate, the scalability, security, and efficiency these GPUs deliver are more essential than ever. We are proud to be among the first cloud providers worldwide to offer AMD MI355X GPUs, empowering our customers with next-generation AI infrastructure." AMD is recognised as a member of the Vultr Cloud Alliance, which supports a collaborative ecosystem of technology providers focused on offering integrated cloud computing solutions. The introduction of the MI355X GPU follows a period of upgrades across AMD's GPU lineup, including a greater emphasis on catering to both inferencing and enterprise-scale workloads. Vultr's offering is aimed at organisations seeking advanced compute resources for AI-driven applications and scientific tasks requiring significant computational capacity. Vultr's global network reportedly serves hundreds of thousands of customers across 185 countries, supplying services in cloud compute, GPU, bare metal infrastructure and cloud storage. The addition of AMD's latest GPU to its infrastructure underlines Vultr's commitment to providing a variety of options for businesses and developers pursuing AI and HPC advancements.

DOWNLOAD THE APP

Get Started Now: Download the App

Ready to dive into a world of global content with local flavor? Download Daily8 app today from your preferred app store and start exploring.
app-storeplay-store