Tech

Here’s why 100TB+ SSDs will play a huge role in ultra large language models in the near future

Share
Share

  • Kioxia reveals new project called AiSAQ which wants to substitute RAM with SSDs for AI data processing
  • Bigger (read: 100TB+) SSDs could improve RAG at a lower cost than using memory only
  • No timeline has been given, but expect Kioxia’s rivals to offer similar tech

Large language models often generate plausible but factually incorrect outputs – in other words, they make stuff up. These “hallucination”s can damage reliability in information-critical tasks such as medical diagnosis, legal analysis, financial reporting, and scientific research.

Retrieval-Augmented Generation (RAG) mitigates this issue by integrating external data sources, allowing LLMs to access real-time information during generation, reducing errors, and, by grounding outputs in current data, improving contextual accuracy. Implementing RAG effectively requires substantial memory and storage resources, and this is particularly true for large-scale vector data and indices. Traditionally, this data has been stored in DRAM, which, while fast, is both expensive and limited in capacity.

Share

Leave a comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Related Articles
When the school bell rings, the bandwidth drops: How post-15:40 internet surges affect UK broadband quality
Tech

When the school bell rings, the bandwidth drops: How post-15:40 internet surges affect UK broadband quality

Half of parents work after school, causing a broadband battle with streaming-addicted...

You can put Google Gemini right on your smartphone home screen – here’s how
Tech

You can put Google Gemini right on your smartphone home screen – here’s how

Google has launched Gemini home screen widgets for Android and iOS devices...

You can now fact check anybody’s post in WhatsApp – here’s how
Tech

You can now fact check anybody’s post in WhatsApp – here’s how

Perplexity AI’s new WhatsApp integration offers instant fact-checking without leaving the app...