Researchers from Alibaba Group and the Renmin University of China have developed an advanced version of MultiModal Large Language Models (MLLMs) to better understand and interpret images rich in text content. Named DocOwl 1.5, this innovative model uses Unified Structure Learning to enhance the efficiency of MLLMs across five distinct domains: document, webpage, table, chart,…
"Text mining" refers to the discovery of new patterns and insights within large amounts of textual data. Two essential activities in text mining are the creation of a taxonomy - a collection of structured, canonical labels that characterize features of a corpus - and text classification, which assigns labels to instances within the corpus according…
The capabilities of computer vision studies have been vastly expanded due to deep features, which can unlock image semantics and facilitate diverse tasks, even using minimal data. Techniques to extract features from a range of data types – for example, images, text, and audio – have been developed and underpin a number of applications in…
Large language models like GPT-4, while powerful, often struggle with basic visual perception tasks such as counting objects in an image. This can be due to the way these models process high-resolution images. Current AI systems can mainly perceive images at a fixed low resolution, leading to distortion, blurriness, and loss of detail when the…
Research in materials science is increasingly focusing on the rapid discovery and characterization of materials with specific attributes. A key aspect of this research is the comprehension of crystal structures, which are naturally complex due to their periodic and infinite nature. This complexity presents significant challenges when attempting to model and predict material properties, difficulties…
The production of realistic human facial images has been a long-standing challenge for researchers in machine learning and computer vision. Earlier techniques like Eigenfaces utilised Principal Component Analysis (PCA) to learn statistical priors from data, yet they notably struggled to capture the complexities of real-world factors such as lighting, viewpoints, and expressions beyond frontal poses.…
In the world of machine learning, large language models (LLMs) are a significant area of study. Recently, model merging or the combination of multiple LLMs into a single framework has fascinated the researcher's community because it doesn't require any additional training. This reduces the cost of creating new models considerably, sparking an interest in model…
In the ever-evolving fields of computer vision and artificial intelligence, traditional methodologies favor larger models for advanced visual understanding. The assumption underlying this approach is that larger models can extract more powerful representations, prompting the construction of enormous vision models. However, a recent study challenges this wisdom, with a closer look at the practice of…
Decompilation is a pivotal process in software reverse engineering facilitating the analysis and interpretation of binary executables when the source code is not directly accessible. Valuable for security analysis, bug detection, and the recovery of legacy code, the process often needs assistance in generating a human-readable and semantically accurate source code, which is a substantial…
The increasing use of facial recognition technologies is a double-edged sword, wherein it provides unprecedented convenience, but also poses a significant risk to personal privacy as facial data could unintentionally reveal private details about an individual. As such, there is an urgent need for privacy-preserving measures in these face recognition systems.
A pioneering approach to this…
Data scientists and engineers often encounter difficulties when collaborating on machine learning (ML) tasks due to concerns about data reproducibility and traceability. Software code tends to be transparent about its origin and modifications, but it's often hard to ascertain the exact provenance of the data used for training ML models and the transformations conducted.
To tackle…
Researchers from IBM Research have developed a new architecture, dubbed Alignment Studio, which enables developers to mould large language models (LLMs) to fit specific societal norms, laws, values and regulations. The system is designed to mitigate ongoing challenges in the artificial intelligence (AI) sector surrounding issues such as hate speech and inappropriate language.
While efforts…