Researchers from Stanford University have developed a new model to investigate the contributions of individual data points to machine learning processes. This allows an understanding of how the value of each data point changes as the scale of the dataset grows, illustrating that some points are more useful in smaller datasets, while others become more…
MIT President Sally Kornbluth and Provost Cynthia Barnhart launched a call for papers last summer relating to generative AI, with the aim of collecting effective strategies, policy suggestions, and calls to action in this expansive field. The response was overwhelming, with a total submission of 75 proposals, out of which 27 were chosen for seed…
Polkadot, a blockchain platform with a market cap of $8 billion, released its H1 2024 Treasury report revealing massive spending in marketing and promotional activities. The details expose the company's high expenses, with the Treasury spending $87 million in the first half of 2024 alone. Most notable was the allocation of $37 million in "Outreach",…
In this week's round-up of artificial intelligence (AI) news, quite a few intriguing developments have been highlighted. For starters, the question of AI security is heavily discussed, largely sparked by Microsoft's revelation of a "Skeleton Key Jailbreak". The discovery raises concerns over the potential misbehavior of AI models, emphasizing the need for safe integrations and…
Overfitting is a prevalent problem when training large neural networks on limited data. It indicates a model's strong performance on the training data but its failure to perform comparably on unseen test data. This issue arises when the network’s feature detectors become overly specialized to the training data, building complex dependencies that do not apply…
The computer vision sector is currently dominated by large-scale models that offer remarkable performance but demand high computational resources, making them impractical for real-world applications. To address this, the Google Research Team has opted to reduce these models into smaller, more efficient architectures via model pruning and knowledge distillation. The team's focus is on knowledge…
Last year, a request from MIT President Sally Kornbluth and Provost Cynthia Barnhart for research proposals concerning generative AI initiatives resulted in 75 submissions. Out of these, 27 were selected for initial funding. Inspired by the impressive response, Kornbluth and Barnhart issued a second call for papers last fall. This resulted in 53 more proposals,…
Workplace safety is experiencing a revolution with the adoption of Artificial Intelligence (AI) and automation technologies that can anticipate and prevent accidents before they occur. Unlike the reactive safety protocols of the past which responded to accidents after they happened, these technologies make safety proactive, reducing risks and increasing the overall efficiency of the workplace.
Proactive…
Boston University researchers have developed an AI system that can predict with nearly 80% accuracy whether someone with mild cognitive impairment will develop Alzheimer's disease within six years, based on the analysis of speech patterns. The study, published in the journal Alzheimer’s & Dementia, leverages AI to extract diagnostic information from cognitive assessments, thus expediting…
Developing and fine-tuning language model systems is a challenging process that usually consumes a significant amount of time and resources even for tech giants like Google and Meta. It involves an iterative process of supervised fine-tuning, aligning with human preferences, distillation, and continuous adjustment until a certain quality threshold is met. This process can take…
The evolution of Large Language Models (LLMs) in artificial intelligence has spawned several sub-groups, including Multi-Modal LLMs, Open-Source LLMs, Domain-specific LLMs, LLM Agents, Smaller LLMs, and Non-Transformer LLMs.
Multi-Modal LLMs, such as OpenAI's Sora, Google's Gemini, and LLaVA, consolidate various types of input like images, videos, and text to perform more sophisticated tasks. OpenAI's Sora…
The creation and implementation of effective AI agents have become a vital point of interest in the Language Learning Model (LLM) field. AI company, Anthropic, recently spotlighted several successful design patterns being employed in practical applications. Discussed in relation to Claude's models, these patterns offer transferable insights for other LLMs. Five key design patterns examined…