• Live Crypto Prices
  • Crypto News
    • Worldwide
      • Bitcoin
      • Ethereum
      • Altcoin
      • Blockchain
      • Regulation
    • Australian Crypto News
  • Education
    • Cryptocurrency For Beginners
    • Where to Buy Cryptocurrency
    • Where to Store Cryptos
    • Cryptocurrency Tax in Australia 2021
No Result
View All Result
CryptoABC.net
No Result
View All Result

AMD Radeon PRO GPUs and ROCm Software Expand LLM Inference Capabilities

August 31, 2024
in Blockchain
Reading Time: 3min read
0 0
A A
0
Llama 3.1 Now Optimized for AMD Platforms from Data Center to AI PCs
0
SHARES
7
VIEWS
ShareShareShareShareShare


Felix Pinkston
Aug 31, 2024 01:52

AMD’s Radeon PRO GPUs and ROCm software enable small enterprises to leverage advanced AI tools, including Meta’s Llama models, for various business applications.





AMD has announced advancements in its Radeon PRO GPUs and ROCm software, enabling small enterprises to leverage Large Language Models (LLMs) like Meta’s Llama 2 and 3, including the newly released Llama 3.1, according to AMD.com.

New Capabilities for Small Enterprises

With dedicated AI accelerators and substantial on-board memory, AMD’s Radeon PRO W7900 Dual Slot GPU offers market-leading performance per dollar, making it feasible for small firms to run custom AI tools locally. This includes applications such as chatbots, technical documentation retrieval, and personalized sales pitches. The specialized Code Llama models further enable programmers to generate and optimize code for new digital products.

The latest release of AMD’s open software stack, ROCm 6.1.3, supports running AI tools on multiple Radeon PRO GPUs. This enhancement allows small and medium-sized enterprises (SMEs) to handle larger and more complex LLMs, supporting more users simultaneously.

Expanding Use Cases for LLMs

While AI techniques are already prevalent in data analysis, computer vision, and generative design, the potential use cases for AI extend far beyond these areas. Specialized LLMs like Meta’s Code Llama enable app developers and web designers to generate working code from simple text prompts or debug existing code bases. The parent model, Llama, offers extensive applications in customer service, information retrieval, and product personalization.

Small enterprises can utilize retrieval-augmented generation (RAG) to make AI models aware of their internal data, such as product documentation or customer records. This customization results in more accurate AI-generated outputs with less need for manual editing.

Local Hosting Benefits

Despite the availability of cloud-based AI services, local hosting of LLMs offers significant advantages:

  • Data Security: Running AI models locally eliminates the need to upload sensitive data to the cloud, addressing major concerns about data sharing.
  • Lower Latency: Local hosting reduces lag, providing instant feedback in applications like chatbots and real-time support.
  • Control Over Tasks: Local deployment allows technical staff to troubleshoot and update AI tools without relying on remote service providers.
  • Sandbox Environment: Local workstations can serve as sandbox environments for prototyping and testing new AI tools before full-scale deployment.

AMD’s AI Performance

For SMEs, hosting custom AI tools need not be complex or expensive. Applications like LM Studio facilitate running LLMs on standard Windows laptops and desktop systems. LM Studio is optimized to run on AMD GPUs via the HIP runtime API, leveraging the dedicated AI Accelerators in current AMD graphics cards to boost performance.

Professional GPUs like the 32GB Radeon PRO W7800 and 48GB Radeon PRO W7900 offer sufficient memory to run larger models, such as the 30-billion-parameter Llama-2-30B-Q8. ROCm 6.1.3 introduces support for multiple Radeon PRO GPUs, enabling enterprises to deploy systems with multiple GPUs to serve requests from numerous users simultaneously.

Performance tests with Llama 2 indicate that the Radeon PRO W7900 offers up to 38% higher performance-per-dollar compared to NVIDIA’s RTX 6000 Ada Generation, making it a cost-effective solution for SMEs.

With the evolving capabilities of AMD’s hardware and software, even small enterprises can now deploy and customize LLMs to enhance various business and coding tasks, avoiding the need to upload sensitive data to the cloud.

Image source: Shutterstock


Credit: Source link

ShareTweetSendPinShare
Previous Post

NVIDIA Introduces Fast Inversion Technique for Real-Time Image Editing

Next Post

Exploring the Future of Real World Assets in DeFi

Next Post
Sui Introduces Secure Native Randomness for Testnet Applications

Exploring the Future of Real World Assets in DeFi

You might also like

LTC Price Prediction: Targeting $87-$95 Range as Technical Indicators Signal Further Decline Through November 2025

LTC Price Prediction: Targets $62-65 by April 2026 as Technical Indicators Signal Neutral Momentum

March 10, 2026
Crypto Price Prediction Today 4 March – XRP, Bitcoin, Ethereum

Crypto Price Prediction Today 4 March – XRP, Bitcoin, Ethereum

March 4, 2026
Bitcoin On-Chain Data Identifies Unusual Market Cap Behavior

Bitcoin On-Chain Data Identifies Unusual Market Cap Behavior

March 7, 2026
Bitcoin LTH Stay Calm Amid Volatile Market Swings

Bitcoin LTH Stay Calm Amid Volatile Market Swings

March 8, 2026
Bitcoin At The Bottom? The 23-Month Cycle That Has Never Failed

Bitcoin At The Bottom? The 23-Month Cycle That Has Never Failed

March 9, 2026
Analyst Says It’s Time For Bitcoin, But What’s Important About $58,000?

Analyst Says It’s Time For Bitcoin, But What’s Important About $58,000?

March 4, 2026
CryptoABC.net

This is an Australian online news/education portal that aims to provide the latest crypto news, real-time updates, education and reviews within Australia and around the world. Feel free to get in touch with us!

What's New Here!

Circle CEO Allaire Supports Binance Stablecoin Decision

Circle Deploys USDC and CCTP on Morph Layer-2 Network

March 11, 2026
Insurance Giant Aon Partners With Coinbase and Paxos in Trialing Use of Stablecoins for Premium Payments

Insurance Giant Aon Partners With Coinbase and Paxos in Trialing Use of Stablecoins for Premium Payments

March 11, 2026

Subscribe Now

  • Contact Us
  • Privacy Policy
  • Terms of Use
  • DMCA

© 2021 cryptoabc.net - All rights reserved!

No Result
View All Result
  • Live Crypto Prices
  • Crypto News
    • Worldwide
      • Bitcoin
      • Ethereum
      • Altcoin
      • Blockchain
      • Regulation
    • Australian Crypto News
  • Education
    • Cryptocurrency For Beginners
    • Where to Buy Cryptocurrency
    • Where to Store Cryptos
    • Cryptocurrency Tax in Australia 2021

© 2021 cryptoabc.net - All rights reserved!

Welcome Back!

Login to your account below

Forgotten Password?

Create New Account!

Fill the forms below to register

All fields are required. Log In

Retrieve your password

Please enter your username or email address to reset your password.

Log In
Please enter CoinGecko Free Api Key to get this plugin works.