Transformer-based neural networks have demonstrated proficiency in a variety of tasks, such as text generation, editing, and question-answering. Perplexity and end task accuracy measurements consistently show models with more parameters perform better, leading industries to develop larger models. However, in some cases, larger models do not guarantee superior performance. The 2 billion parameter model, MiniCPM,…
Large Language Models (LLMs) like ChatGPT are becoming increasingly significant due to their capability to execute a broad spectrum of tasks including language processing, knowledge extraction, reasoning, planning, coding, and tool use. This has catalyzed research into more refined AI models, hinting at the potential for Artificial General Intelligence (AGI).
LLMs are built on Transformer…
Large Language Models (LLMs) like ChatGPT have received significant interest due to their ability to perform varied AI tasks from language processing to tool use. These capabilities have pushed research toward creating more sophisticated AI models, opening possibilities for Artificial General Intelligence (AGI).
LLMs are built on the Transformer neural network architecture, using autoregressive learning to…
In 2019, Haofeng (Hector) Xu, a PhD student at MIT's Department of Aeronautics and Astronautics, began learning to fly helicopters. His experiences with various risks in the cockpit led to his motivation to make helicopter flight safer. Subsequently, in 2021, he established Rotor Technologies, Inc. This autonomous helicopter company aims to retrofit existing helicopters with…
Research from MIT and other institutions has developed a method, called StreamingLLM, that enables AI chatbots to maintain continuous dialogue without crashing or slowing down. The technique tweaks the key-value cache or conversation memory at the core of large language models. Failure often occurs when this cache needs to store more information than it can…
The Coca-Cola Company has launched a new "limited edition" flavor, Coca-Cola Y3000 Zero Sugar, the result of an AI-generated experiment. Developed as a combination of human ingenuity and artificial intelligence (AI), it captures consumers' future expectations and ideas using their emotions, aspirations, colours, and flavours. This human-AI partnership aims to present a taste indicative of…
Recent viral videos in China have shown global pop star Taylor Swift speaking fluent Mandarin, leading some to believe that she's secretly fluent in the language. However, these clips are the result of AI-generated deepfake technology, created by Shanghai-based tech startup HeyGen. Swift does not speak Mandarin, and the company created the videos to showcase…
Researchers from several universities in China and UK have jointly developed a new method for Graph Neural Networks (GNNs), known as Edge-Node Attention-based Differentiable Pooling (ENADPool). This method uses hard clustering and incorporates attention mechanisms to compress node features and edge strengths in GNNs. They also introduced the Multi-distance GNN (MD-GNN) model to mitigate over-smoothing…
State-space models (SSMs) are an essential part of deep learning, used for sequence modeling. They observe a system where the output depends both on current and earlier inputs. This mechanism is utilized extensively in signal processing, control systems, and natural language processing. There lays a challenge with SSMs, it lies in their execution inefficiency, especially…
The paper discussed in this largely explored the effectiveness of machine-learning-based models in wireless link path loss predictions, in lieu of traditional models like Longley-Rice and free space path loss (FSPL). Traditional models suffer in accuracy in non-line-of-sight scenarios due to their inability to account for signal attenuation, or interference caused by electromagnetic interplay with…
Researchers from Columbia University and Databricks Mosaic AI have conducted a comparative study of full finetuning and Low-Rank Adaptation (LoRA), a parameter-efficient finetuning method, in large language models (LLMs). The efficient finetuning of LLMs, which can contain billions of parameters, is an ongoing challenge due to the substantial GPU memory required. This makes the process…
Recent research suggests that incorporating demonstrating examples, or in-context learning (ICL), significantly enhances large language models' (LLM's) and large multimodal models' (LMM's) performance. Studies have shown improvements in LLM performance with increased in-context examples, particularly in out-of-domain tasks. These findings are driven by newer models such as GPT-4o and Gemini 1.5 Pro, which include longer…