Tsinghua University's Knowledge Engineering Group (KEG) has introduced GLM-4 9B, an innovative, open-source language model that surpasses other models like GPT-4 and Gemini in different benchmark tests. Developed by the Tsinghua Deep Model (THUDM) team, GLM-4 9B signals an important development in the sphere of natural language processing.
At its core, GLM-4 9B is a colossal…
The advancement of natural language processing (NLP) capabilities has been to a large extent, dependent on developing large language models (LLMs). Although these models deliver high performance, they also pose challenges due to their need for immense computational resources and related costs, making them hard to scale up without incurring substantial expenses.
These challenges, therefore, create…
Snowflake recently introduced the Polaris Catalog, a new open-source catalog for Apache Iceberg designed to boost data interoperability across multiple engines and cloud services. The release illustrates Snowflake's commitment to granting businesses more control, flexibility, and security in their data management.
The data sector has grown increasingly fond of open-source file and table formats due to…
A research team from IEIT Systems has recently developed a new model, Yuan 2.0-M32, which uses the Mixture of Experts (MoE) architecture. This complex model is built on the same foundation as the Yuan-2.0 2B, but with utilization of 32 experts, only two of whom are active at any given time, resulting in its unique…
Artificial Intelligence (AI) is increasingly being used in legal research and document drafting, aimed at improving efficiency and accuracy. However, concerns regarding the reliability of these tools persist, especially given the potential for the creation of false or misleading information, referred to as "hallucinations". This issue is of particular concern given the high-stakes nature of…
Language and Large Model (LLM) research has shifted focus to steerability and persona congruity with complexities, challenging previous research simply based on one-dimensional personas or multiple-choice formats. A persona's intricacy and its potential to multiply biases in LLM simulations when there's lack of alignment with typical demographic views is now recognized.
A recent research by…