ChatGPT, an AI system by OpenAI, is making waves in the artificial intelligence field with its advanced language capabilities. Capable of performing tasks such as drafting emails, conducting research, and providing detailed information, such tools are transforming the way office tasks are conducted. They contribute to more efficient and productive workplaces. As with any technological…
Optimal transport is a mathematical field focused on the most effective methods for moving mass between probability distributions. It has a broad range of applications in disciplines such as economics, physics, and machine learning. However, the optimization of probability measures in optimal transport frequently faces challenges due to complex cost functions influenced by various factors…
Large language models (LLMs) are exceptional at generating content and solving complex problems across various domains. Nevertheless, they struggle with multi-step deductive reasoning — a process requiring coherent and logical thinking over extended interactions. The existing training methodologies for LLMs, based on next-token prediction, do not equip them to apply logical rules effectively or maintain…
Harnessing high-dimensional clinical data (HDCD) – health care datasets with significantly higher variables than patients – for genetic discovery and disease prediction poses a considerable challenge. HDCD analysis and processing demands immense computational resources due to its rapidly expanding data space. This further complicates interpreting models based on this data, potentially hindering clinical decisions. Traditional…
The evaluation of large language models (LLMs) has always been a daunting task due to the complexity and versatility of these models. However, researchers from Google DeepMind, Google, and UMass Amherst have introduced FLAMe, a new family of evaluation models developed to assess the reliability and accuracy of LLMs. FLAMe stands for Foundational Large Autorater…
Large Language Models (LLMs) have become increasingly important in AI and data processing tasks, but their superior size leads to substantial memory requirements and bandwidth consumption. Standard procedures such as Post-Training Quantization (PTQ) and Quantized Parameter-Efficient Fine-Tuning (Q-PEFT) can often compromise accuracy and performance, and are impractical for larger networks. To combat this, researchers have…
Language models (LMs), used in applications such as autocomplete and language translation, are trained on a vast amount of text data. Yet, these models also face significant challenges in relation to privacy and copyright concerns. In some cases, the inadvertent inclusion of private and copyrighted content in training datasets can lead to legal and ethical…
Researchers at the University of Texas (UT) in Austin have introduced a new benchmark designed to evaluate the effectiveness of artificial intelligence in solving complex mathematical problems. PUTNAMBENCH is aimed at solving a key issue facing the sector as current benchmarks are not sufficiently rigorous and mainly focus on high-school level mathematics.
Automating mathematical reasoning…
Large Language Models (LLMs) are vital for tasks in natural language processing but they encounter issues when it comes to deployment. This is due to their substantial computational and memory requirements during inference. Current research studies are focused on boosting LLM efficiency by applying methods such as quantization, pruning, distillation, and improved decoding. One of…
Large language models (LLMs) are being extensively used in multiple applications. However, they have a significant limitation: they struggle to process long-context tasks due to the constraints of transformer-based architectures. Researchers have explored various approaches to boost LLMs' capabilities in processing extended contexts, including improving softmax attention, reducing computational costs and refining positional encodings. Techniques…
Researchers from the Shanghai AI Laboratory and Tsinghua University have developed NeedleBench, a novel framework to evaluate the retrieval and reasoning capabilities of large language models (LLMs) in exceedingly long contexts (up to 1 million tokens). The tool is critical for real-world applications such as legal document analysis, academic research, and business intelligence, which rely…
The task of translating natural language queries (text-to-SQL) into SQL has been historically challenging due to the complexity of understanding user questions, database schemas, and SQL production. Recent innovations have seen the integration of Pre-trained Language Models (PLMs) into text-to-SQL systems, which have displayed much promise. However, they can generate incorrect SQL due to growing…