Research conducted by institutions like FAIR, Meta AI, Datashape, and INRIA explores the emergence of Chain-of-Thought (CoT) reasoning in Language Learning Models (LLMs). CoT enhances the capabilities of LLMs, enabling them to perform complex reasoning tasks, even though they are not explicitly designed for it. Even as LLMs are primarily trained for next-token prediction, they…
Monte Carlo (MC) methods are popularly used for modeling complex real-world systems, particularly those related to financial mathematics, numerical integration, and optimization problems. However, these models demand a large number of samples to achieve high precision, especially with complex issues.
As a solution, researchers from the Massachusetts Institute of Technology (MIT), the University of Waterloo, and…
In the rapidly evolving field of artificial intelligence (AI), large language models (LLMs) play a crucial role in processing vast amounts of information. However, to ensure their efficiency and reliability, certain techniques and tools are necessary. Some of these fundamental methodologies include Retrieval-Augmented Generation (RAG), agentic functions, Chain of Thought (CoT) prompting, few-shot learning, prompt…
Deep learning models have significantly affected the evolution of audio classification. Originally, Convolutional Neural Networks (CNNs) monopolized this field, but it has since shifted to transformer-based architectures that provide improved performance and unified handling of various tasks. However, the computational complexity associated with transformers presents a challenge for audio classification, making the processing of long…
Microsoft's AI courses offer robust education in AI and machine learning across a range of skill levels. By emphasizing practical usage, advanced techniques, and ethical AI practices, students learn how to develop and deploy AI solutions effectively and responsibly.
The "Fundamentals of machine learning" course provides a grounding in machine learning's core concepts along with deep…
In 2010, Karthik Dinakar and Birago Jones, students at the Media Lab, collaborated to develop a tool that could assist content moderation teams in identifying concerning posts on platforms like Twitter and YouTube. Their innovation received extensive attention, leading to an invitation to present their creation at a cyberbullying seminar at the White House. However,…
In the growing field of warehouse automation, managing hundreds of robots zipping through a large warehouse is a logistical challenge. Delivery paths, potential collisions and congestion all pose significant issues, making the task a complex problem that even the best algorithms find hard to manage. To solve this, a team of MIT researchers has developed…
Matrix multiplication (MatMul) is a fundamental process in most neural network topologies. It is commonly used in vector-matrix multiplication (VMM) by dense layers in neural networks, and in matrix-matrix multiplication (MMM) by self-attention mechanisms. Significant reliance on MatMul can be attributed to GPU optimization for these tasks. Libraries like cuBLAS and the Compute Unified Device…
Researchers have identified cultural accumulation as a crucial aspect of human success. This practice refers to our capacity to learn skills and accumulate knowledge over generations. However, currently used artificial learning systems, like deep reinforcement learning, frame the learning question as happening within a single "lifetime." This approach does not account for the generational and…
Language Learning Models (LLMs) can come up with good answers and even be honest about their mistakes. However, they often provide simplified estimations when they haven't seen certain questions before, and it's crucial to develop ways to draw reliable confidence estimations from them. Traditionally, both training-based and prompting-based approaches have been used, but these often…
Stanford University researchers have developed a new method called Demonstration ITerated Task Optimization (DITTO) designed to align language model outputs directly with users' demonstrated behaviors. This technique was introduced to address the challenges language models (LMs) face - including the need for big data sets for training, generic responses, and mismatches between universal style and…