Skip to content Skip to sidebar Skip to footer

Artificial Intelligence

A more efficient and improved method to inhibit AI chatbots from producing harmful responses.

AI chatbots like ChatGPT, trained on vast amounts of text from billions of websites, have a broad potential output which includes harmful or toxic material, or even leaking personal information. To maintain safety standards, large language models typically undergo a process known as red-teaming, where human testers use prompts to elicit and manage unsafe outputs.…

Read More

A fresh approach to artificial intelligence measures ambiguity in health-related imagery.

Biomedical segmentation pertains to marking pixels from significant structures in a medical image like cells or organs which is crucial for disease diagnosis and treatment. Generally, a single answer is provided by most artificial intelligence (AI) models while making these annotations, but such a process is not always straightforward. In a recent paper, Marianne Rakic, an…

Read More

Scientists at the University of Auckland have presented ChatLogic, an advanced tool for multi-step reasoning in large language models, which improves precision in complex tasks by over half.

Large language models (LLMs) are exceptional at generating content and solving complex problems across various domains. Nevertheless, they struggle with multi-step deductive reasoning — a process requiring coherent and logical thinking over extended interactions. The existing training methodologies for LLMs, based on next-token prediction, do not equip them to apply logical rules effectively or maintain…

Read More

Google Research introduces a new AI strategy for genetic exploration which can utilize concealed information in highly dimensional medical data.

Harnessing high-dimensional clinical data (HDCD) – health care datasets with significantly higher variables than patients – for genetic discovery and disease prediction poses a considerable challenge. HDCD analysis and processing demands immense computational resources due to its rapidly expanding data space. This further complicates interpreting models based on this data, potentially hindering clinical decisions. Traditional…

Read More

Google AI has released an AI paper, presenting FLAMe: a fundamental, large-scale auto-scoring model for trustworthy and effective evaluation of Language Model (LLM).

The evaluation of large language models (LLMs) has always been a daunting task due to the complexity and versatility of these models. However, researchers from Google DeepMind, Google, and UMass Amherst have introduced FLAMe, a new family of evaluation models developed to assess the reliability and accuracy of LLMs. FLAMe stands for Foundational Large Autorater…

Read More

Efficient Quantization-Aware Training (EfficientQAT): A New Approach to Quantification in Machine Learning for Compressing Large Language Models (LLMs).

Large Language Models (LLMs) have become increasingly important in AI and data processing tasks, but their superior size leads to substantial memory requirements and bandwidth consumption. Standard procedures such as Post-Training Quantization (PTQ) and Quantized Parameter-Efficient Fine-Tuning (Q-PEFT) can often compromise accuracy and performance, and are impractical for larger networks. To combat this, researchers have…

Read More

MUSE: An Inclusive AI Platform for Assessing Machine Forgetting in Language Models

Language models (LMs), used in applications such as autocomplete and language translation, are trained on a vast amount of text data. Yet, these models also face significant challenges in relation to privacy and copyright concerns. In some cases, the inadvertent inclusion of private and copyrighted content in training datasets can lead to legal and ethical…

Read More

Researchers from the University of Texas at Austin have launched PUTNAMBENCH, a thorough AI benchmarking tool for assessing the performance of Neural Theorem-Provers on Putnam Mathematical Problems.

Researchers at the University of Texas (UT) in Austin have introduced a new benchmark designed to evaluate the effectiveness of artificial intelligence in solving complex mathematical problems. PUTNAMBENCH is aimed at solving a key issue facing the sector as current benchmarks are not sufficiently rigorous and mainly focus on high-school level mathematics. Automating mathematical reasoning…

Read More

The Updated Open-Source Edition of DeepSeek-V2-0628 Has Been Launched: A More Advanced Version.

DeepSeek has announced the launch of its advanced open-source AI model, DeepSeek-V2-Chat-0628, on Hugging Face. The update represents a significant advancement in AI text generation and chatbot technology. This new version secures the overall ranking of #11 according to the LMSYS Chatbot Arena Leaderboard, outperforming all other existing open-source models. It is an upgrade on…

Read More

An improved and speedier method to stop AI chatbot from providing harmful responses.

AI chatbots pose unique safety risks—while they can write computer programs or provide useful summaries of articles, they can also potentially generate harmful or even illegal instructions, including how to build a bomb. To address such risks, companies typically use a process called red-teaming. Human testers aim to generate unsafe or toxic content from AI…

Read More

A new technique in artificial intelligence accurately recognizes uncertainty in health imagery.

A research team from MIT, the Broad Institute of MIT and Harvard, and Massachusetts General Hospital has developed an artificial intelligence (AI) tool, named Tyche, that presents multiple plausible interpretations of medical images, highlighting potentially important and varied insights. This tool aims to address the often complex ambiguity in medical image interpretation where different experts…

Read More

Q-Sparse: A Novel AI Method to Achieve Complete Sparsity of Activations in Large Language Models (LLMs)

Large Language Models (LLMs) are vital for tasks in natural language processing but they encounter issues when it comes to deployment. This is due to their substantial computational and memory requirements during inference. Current research studies are focused on boosting LLM efficiency by applying methods such as quantization, pruning, distillation, and improved decoding. One of…

Read More