New AI News | MIT Fixes AI Chatbots #1 Drawback!

3 min read

AI News Latest: Researchers from MIT and other institutions have found a solution to prevent chatbots, like the ones driving AI assistants, from crashing during long conversations. The problem was related to the key-value cache, a sort of memory for the chatbot, which could get overloaded, causing the model to fail. The new method, called StreamingLLM, ensures that the initial data points in the cache are preserved, allowing the chatbot to handle conversations of over 4 million words without crashing.

Imagine you’re talking to a computer-based assistant, like Siri or Gemini, and sometimes these assistants struggle when conversations go on for a long time. Researchers from MIT and other places discovered why this happens and came up with a fix.

AI News Chatbot Breaks down after Long Conversations

They found that the computer’s memory, which is like its brain, can get overwhelmed during lengthy chats. This can make the assistant slow down or even crash. The solution they developed is called StreamingLLM, and it makes sure that the important bits of the conversation are always kept in the computer’s memory.

To understand this, think of the memory like a backpack. When it gets full, the computer has to decide what to keep and what to throw away. The new method ensures that the computer always keeps the first few things you talked about, even if the backpack is full. This helps the computer stay fast and reliable, even during super long chats.

This improvement means that these computer assistants, which use large language models, can work for a long time without needing a break. They can be useful for writing, editing, or doing other tasks without constantly needing to start over.

The researchers figured out that the computer pays extra attention to the first things said in a conversation. Even if those first things seem unrelated to the later parts of the talk, the computer uses them to keep everything running smoothly. They call these first things “attention sinks,” and by making sure they stay in the computer’s memory, the chatbot can handle really long conversations without any problems.

This method is so good that it’s now part of a library used by powerful computers, like the ones from NVIDIA, which helps them perform better when dealing with lots of words. In simpler terms, it’s like giving your computer assistant a magic trick to handle super long conversations without getting tired or confused.

ai generated, woman, robot-7874140.jpg

This improvement makes AI assistants more efficient for tasks like copywriting, editing, or generating code, as they can work continuously without frequent reboots. The researchers achieved this by addressing a phenomenon called “attention sinks” and optimizing the cache’s structure. The method has been incorporated into NVIDIA’s library for large language models.

You May Also Like

More From Author

+ There are no comments

Add yours