The debate over the necessity of copyrighted materials to train top Artificial Intelligence (AI) models continues to be a hot topic within the AI industry. This discussion was fueled further when OpenAI proclaimed to the UK Parliament in 2023 that it's 'impossible' to train these models without using copyrighted content, resulting in legal disputes and…
Artificial intelligence (AI) is an industry that is developing at a rapid pace. However, there are several challenges that exist in transitioning research innovations into practical applications. It can be a difficult task to improve the quality of AI models to match the standards required for production. Even though researchers can create robust models, adapting…
In the ever-evolving fields of computer vision and artificial intelligence, traditional methodologies favor larger models for advanced visual understanding. The assumption underlying this approach is that larger models can extract more powerful representations, prompting the construction of enormous vision models. However, a recent study challenges this wisdom, with a closer look at the practice of…
Decompilation is a pivotal process in software reverse engineering facilitating the analysis and interpretation of binary executables when the source code is not directly accessible. Valuable for security analysis, bug detection, and the recovery of legacy code, the process often needs assistance in generating a human-readable and semantically accurate source code, which is a substantial…
The increasing use of facial recognition technologies is a double-edged sword, wherein it provides unprecedented convenience, but also poses a significant risk to personal privacy as facial data could unintentionally reveal private details about an individual. As such, there is an urgent need for privacy-preserving measures in these face recognition systems.
A pioneering approach to this…
Jailbreak attacks aim to identify and address security vulnerabilities in Language Models (LLMs) by bypassing their safety protocols. Despite significant advancements in LLMs, particularly in the area of natural language processing, they remain prone to such attacks. Given the increasing sophistication of new jailbreak techniques, the need for robust defense methodologies has grown. These methods,…
Microsoft researchers have introduced Garnet, a versatile and highly performant cache-store system designed to support the rapidly evolving needs of modern applications. Traditional cache-stores have struggled to keep pace with the increasing complexity and demands of interactive web applications, driving the creation of this new, open-source solution.
As opposed to its predecessor, Garnet handles not…
Data scientists and engineers often encounter difficulties when collaborating on machine learning (ML) tasks due to concerns about data reproducibility and traceability. Software code tends to be transparent about its origin and modifications, but it's often hard to ascertain the exact provenance of the data used for training ML models and the transformations conducted.
To tackle…
Researchers from IBM Research have developed a new architecture, dubbed Alignment Studio, which enables developers to mould large language models (LLMs) to fit specific societal norms, laws, values and regulations. The system is designed to mitigate ongoing challenges in the artificial intelligence (AI) sector surrounding issues such as hate speech and inappropriate language.
While efforts…
Researchers from Tsinghua University and Microsoft Corporation have unveiled a groundbreaking study known as LLMLingua-2, as part of a collaborative effort that reinforces the cruciality of interdisciplinary research. The study primarily focuses on improving the efficiency of language models, which play a pivotal role in ensuring fluent communication between humans and machines. The core challenge…
Researchers from HyperGAI have developed a ground-breaking new multimodal language learning model (LLMs) known as Hyper Pretrained Transformers (HPT) that can proficiently handle and process seamlessly, a wide array of input modalities, such as text, images, and videos. Existing LLMs, like GPT-4V and Gemini Pro, have limitations in comprehending multimodal data, which hinders progress towards…
The field of artificial intelligence (AI) has significantly advanced with the development of Large Language Models (LLMs) such as GPT-3 and GPT-4. Developed by research institutions and tech giants, LLMs have shown great promise by excelling in various reasoning tasks, from solving complex math problems to understanding natural language nuances. However, despite their notable accomplishments,…