Reinforcement Learning (RL) expands beyond its origins in gaming and finds innovative applications across various industries such as finance, healthcare, robotics, autonomous vehicles, and smart infrastructure.
In finance, RL algorithms are reinventing investment strategies and risk management by making sequential decisions, observing market conditions, and adjusting strategies based on rewards. Despite their potential, these algorithms struggle…
Language model-based machine learning systems, or LLMs, are reaching beyond their previous role in dialogue systems and are now actively participating in real-world applications. There is an increasing belief that many web interactions will be facilitated by systems driven by these LLMs. However, due to the complexities involved, humans are presently needed to verify the…
Large Language Models (LLMs) like those used in Microsoft Bing or Google Search are capable of providing natural language responses to user queries. Traditional search engines often struggle to provide cohesive responses, only offering relevant page results. LLMs improve upon this by compiling results into understandable answers. Yet, issues arise with keeping LLMs current with…
Pretrained language models (LMs) are essential tools in the realm of machine learning, often used for a variety of tasks and domains. But, adapting these models, also known as finetuning, can be expensive and time-consuming, especially for larger models. Traditionally, the solution to this issue has been to use Parameter-efficient finetuning (PEFT) methods such as…
The rapid improvement of large language models and their role in natural language processing has led to challenges in incorporating less commonly spoken languages. Embedding the majority of artificial intelligence (AI) systems in well-known languages inevitably forces a technological divide across linguistic communities that remains mostly unaddressed.
This paper introduces the SambaLingo system, a novel…
Large-scale language models (LLMs) have made substantial progress in understanding language by absorbing information from their environment. However, while they excel in areas like historical knowledge and insightful responses, they struggle when it comes to real-time comprehension. Embodied AI, integrated into items like smart glasses or home robots, aims to interact with humans using everyday…
Memory is a crucial component of intelligence, facilitating the recall and application of past experiences to current situations. However, both traditional Transformer models and Transformer-based Large Language Models (LLMs) have limitations related to context-dependent memory due to the workings of their attention mechanisms. This primarily concerns the memory consumption and computation time of these attention…
Scientific research, despite its vital role in improving human well-being, often grapples with challenges due to its complexities and the slow progress it typically makes. This often necessitates specialized expertise. The application of artificial intelligence (AI), especially large language models (LLMs) is identified as a potential game-changer in the process of scientific research. LLMs have…
Recent research in Artificial Intelligence (AI) has shown a growing interest in the capabilities of large language models (LLMs) due to their versatility and adaptability. These models, traditionally used for tasks in natural language processing, are now being explored for potential use in computational tasks, such as regression analysis. The idea behind this exploration is…
Chain-of-thought (CoT) prompting, an instruction method for language models (LMs), seeks to improve a model's performance across arithmetic, commonsense, and symbolic reasoning tasks. However, it falls short in larger models (with over 100 billion parameters) due to its repetitive rationale and propensity to produce unaligned rationales and answers.
Researchers from Penn State University and Amazon AGI…
MixedBread.ai, known for its work in artificial intelligence, has come up with a novel method called Binary Matryoshka Representation Learning (Binary MRL) for reducing the size of the memory footprint of embeddings used in natural language processing (NLP) applications. Embeddings are crucial to various functions in NLP such as recommendation systems, retrieval processes, and similarity…
Large Language Models (LLMs), known for their key role in advancing natural language processing tasks, continue to be polished to better comprehend and execute complex instructions across a range of applications. However, a standing issue is the tendency for LLMs to only partially follow given instructions, a shortcoming that results in inefficiencies when the models…