Tech

Here’s why 100TB+ SSDs will play a huge role in ultra large language models in the near future

Share
Share

  • Kioxia reveals new project called AiSAQ which wants to substitute RAM with SSDs for AI data processing
  • Bigger (read: 100TB+) SSDs could improve RAG at a lower cost than using memory only
  • No timeline has been given, but expect Kioxia’s rivals to offer similar tech

Large language models often generate plausible but factually incorrect outputs – in other words, they make stuff up. These “hallucination”s can damage reliability in information-critical tasks such as medical diagnosis, legal analysis, financial reporting, and scientific research.

Retrieval-Augmented Generation (RAG) mitigates this issue by integrating external data sources, allowing LLMs to access real-time information during generation, reducing errors, and, by grounding outputs in current data, improving contextual accuracy. Implementing RAG effectively requires substantial memory and storage resources, and this is particularly true for large-scale vector data and indices. Traditionally, this data has been stored in DRAM, which, while fast, is both expensive and limited in capacity.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
New hires are cybersecurity goldmines for hackers, and most companies don’t even realize they’re making it easy
Tech

New hires are cybersecurity goldmines for hackers, and most companies don’t even realize they’re making it easy

Most phishing incidents happen before new employees even understand how internal systems...

Analytical model evaluates performance of grant-free communication in densely populated IoT environment
Tech

Analytical model evaluates performance of grant-free communication in densely populated IoT environment

Credit: Pixabay/CC0 Public Domain Imagine a world where every smart device, from...