Skip to content Skip to sidebar Skip to footer

Large Language Model

The AI study from China presents MiniCPM: Unveiling progressive minimal language models via scalable teaching methods.

In recent years, there has been increasing attention paid to the development of Small Language Models (SLMs) as a more efficient and cost-effective alternative to Large Language Models (LLMs), which are resource-heavy and present operational challenges. In this context, researchers from the Department of Computer Science and Technology at Tsinghua University and Modelbest Inc. have…

Read More

This academic paper from Meta and MBZUAI introduces a systematic AI structure designed to investigate precise scaling interactions related to model size and its knowledge storage capacity.

Researchers from Meta/FAIR Labs and Mohamed bin Zayed University of AI have carried out a detailed exploration into the scaling laws for large language models (LLMs). These laws delineate the relationship between factors such as a model's size, the time it takes to train, and its overall performance. While it’s commonly held that larger models…

Read More

Eagle (RWKV-5) and Finch (RWKV-6): Realizing Significant Advancements in Repetitive Neural Networks-Based Language Models through the Incorporation of Multiheaded Matrix-Valued States and Dynamic Data-Driven Recurrence Processes.

The field of Natural Language Processing (NLP) has witnessed a radical transformation following the advent of Large Language Models (LLMs). However, the prevalent Transformer architecture used in these models suffers from quadratic complexity issues. While techniques such as sparse attention have been developed to lower this complexity, a new generation of models is making headway…

Read More

Researchers from Hong Kong Polytechnic University and Chongqing University Have Developed a Tool, CausalBench, for Evaluating Logical Machine Learning in AI Advancements.

Causal learning plays a pivotal role in the effective operation of artificial intelligence (AI), helping improve AI models' ability to rationalize decisions, adapt to new data, and visualize hypothetical scenarios. However, the evaluation of large language models' (LLM) proficiency in processing causality, such as GPT-3 and its variants, remains a challenge due to the need…

Read More

Google AI Debuts Patchscopes: A Machine Learning Method Teaching LLMs to Yield Natural Language Explanations of Their Concealed Interpretations.

To overcome the challenges in interpretability and reliability of Large Language Models (LLMs), Google AI has introduced a new technique, Patchscopes. LLMs, based on autoregressive transformer architectures, have shown great advancements but their reasoning process and decision-making are opaque and complex to understand. Current methods of interpretation involve intricate techniques that dig into the models'…

Read More

Samba-CoE v0.3: Transforming AI Efficiency through Enhanced Routing Abilities.

SambaNova has unveiled its latest Composition of Experts (CoE) system, the Samba-CoE v0.3, marking a significant advancement in the effectiveness and efficiency of machine learning models. The Samba-CoE v0.3 demonstrates industry-leading capabilities and has outperformed competitors such as DBRX Instruct 132B and Grok-1 314B on the OpenLLM Leaderboard. Samba-CoE v0.3 unveils a new and efficient routing…

Read More

Cohere AI introduces Rerank 3: An innovative base model created to enhance enterprise search and enhance Retrieval Augmented Generation (RAG) systems.

Artificial Intelligence (AI) company Cohere has launched Rerank 3, an advanced foundation model designed to enhance enterprise search and Retrieval Augmented Generation (RAG) systems, promising superior efficiency, accuracy, and cost-effectiveness than its earlier versions. The key beneficiaries of Rerank 3 are enterprises grappling with vast and diverse semi-structured data, such as emails, invoices, JSON documents,…

Read More

Researchers from UC Berkeley have introduced ThoughtSculpt, a novel system that improves the reasoning capabilities of large language models. This system uses advanced Monte Carlo Tree Search methods and unique revision techniques.

Large language models (LLMs), crucial for various applications such as automated dialog systems and data analysis, often struggle in tasks necessitating deep cognitive processes and dynamic decision-making. A primary issue lies in their limited capability to engage in significant reasoning without human intervention. Most LLMs function on fixed input-output cycles, not permitting mid-process revisions based…

Read More

This Chinese AI paper presents a reflection on search Trees (RoT): An LLM Reflection Framework with the intention of enhancing the efficiency of tree-search-inspired prompting techniques.

Large language models (LLMs) paired with tree-search methodologies have been leading advancements in the field of artificial intelligence (AI), particularly for complex reasoning and planning tasks. These models are revolutionizing decision-making capabilities across various applications. However, a notable imperfection lies in their inability to learn from prior mistakes and frequent error repetition during problem-solving. Improving the…

Read More

SpeechAlign: Improving Speech Synthesis through Human Input to Increase Realism and Expressivity in Tech-Based Communication

Speech synthesis—the technological process of creating artificial speech—is no longer a sci-fi fantasy but a rapidly evolving reality. As interactions with digital assistants and conversational agents become commonplace in our daily lives, the demand for synthesized speech that accurately mimics natural human speech has escalated. The main challenge isn't simply to create speech that sounds…

Read More

AutoWebGLM: An Automated Web Navigation Agent, Superior to GPT-4, Based on ChatGLM3-6B

Large Language Models (LLMs) have taken center stage in many intelligent agent tasks due to their cognitive abilities and quick responses. Even so, existing models often fail to meet demands when negotiating and navigating the multitude of complexities on webpages. Factors such as versatility of actions, HTML text-processing constraints, and the intricacy of on-the-spot decision-making…

Read More

CT-LLM: A Compact LLM Demonstrating the Important Move to Prioritize Chinese Language in LLM Development

Natural Language Processing (NLP) has traditionally centered around English language models, thereby excluding a significant portion of the global population. However, this status quo is being challenged by the Chinese Tiny LLM (CT-LLM), a groundbreaking development aimed at a more inclusive era of language models. CT-LLM, innovatively trained on the Chinese language, one of the…

Read More