In 2010, Media Lab students Karthik Dinakar SM ’12, PhD ’17, and Birago Jones SM ’12 wanted to develop a tool to assist content moderation teams at companies like Twitter and YouTube. The project prompted excitement, earning them a demo at a White House cyberbullying summit. When Dinakar struggled to create a working demo, Jones…
Researchers from MIT and other locations have developed a solution to an issue with chatbot performance deterioration following continuous dialogue with a human - a problem attributed to the memory degradation in large language machine-learning models. Their solution, termed StreamingLLM, works by retaining key data points in the memory cache, enabling a chatbot to continue…
In 2010, Media Lab students Karthik Dinakar SM ’12, PhD ’17 and Birago Jones SM ’12 began a project to develop a tool to help content moderators at Twitter (now X) and YouTube spot troubling posts. While preparing for a cyberbullying summit at the White House, they realized that their model was missing nuances in…
A team of researchers, including those from the Massachusetts Institute of Technology (MIT), have created a system called StreamingLLM that allows chatbots to maintain ongoing dialogues without suffering from performance issues. The method involves a reconfiguration of the model's key-value cache—a form of memory storage—that commonly leads to models failing when the cache is overloaded…
In 2010, Media Lab students Karthik Dinakar and Birago Jones began a class project to create a tool for content moderation teams at major corporations like Twitter and YouTube. The technology, which generated significant interest, was presented at a White House cyberbullying summit as a mechanism for identifying concerning posts on social media platforms. During…
Researchers from MIT and other institutions have proposed a solution to the challenge of AI systems losing the context of conversation in extended dialogues. Large language models such as ChatGPT, which enable the functioning of chatbots, often struggle to retain information from long conversations, resulting in rapid performance deterioration.
The team has developed a method…
In 2010, Karthik Dinakar and Birago Jones began a project to develop a tool helping content moderation teams at companies like Twitter and YouTube. The aim was to assist these teams in identifying inappropriate or harmful content. The project created considerable interest and the researchers were invited to present their work at a cyberbullying summit…
A group of researchers from MIT and other institutions have pinpointed a key issue that causes performance degradation in AI chatbots during long conversations and have developed a simple solution to rectify it. Large language machine-learning models such as the ChatGPT use key-value cache to store data. However, when the cache needs to hold more…
In 2010, Karthik Dinakar and Birago Jones, while working on a class project at Media Lab, developed a tool targeting content moderation for social media. The project, aimed at identifying harmful posts on platforms such as Twitter and YouTube, landed them a presentation at a White House cyberbullying summit. They realized before the event, however,…
Researchers from MIT and other institutions have devised an innovative solution to prevent chatbots from crashing during prolonged dialogues. The method, known as StreamingLLM, makes a simple adjustment to the key-value cache, essentially the 'conversation memory,' of well-developed machine-learning models. By ensuring the first few data points don't get bumped out, the chatbot can maintain…
In 2010, Karthik Dinakar and Birago Jones, two students from the Media Lab at the Massachusetts Institute of Technology, collaborated on a project aimed at supporting content moderation teams for social media companies such as Twitter and YouTube. Their ambition was to develop a tool capable of identifying concerning posts online. Despite some initial struggles,…
