AI Breakdown

agibreakdown
undefined
Mar 11, 2024 • 4min

arxiv preprint - A Generative Approach for Wikipedia-Scale Visual Entity Recognition

In this episode, we discuss A Generative Approach for Wikipedia-Scale Visual Entity Recognition by Mathilde Caron, Ahmet Iscen, Alireza Fathi, Cordelia Schmid. The paper introduces a new Generative Entity Recognition (GER) framework for visual entity recognition, aimed at associating images with corresponding entities on Wikipedia, surpassing the typical dual-encoder and captioning model methods. GER functions by decoding a unique "code" linked to an entity from the image, facilitating effective identification. The authors' tests show that GER outperforms existing methods according to the OVEN benchmark, advancing the capabilities of web-scale image-based entity recognition.
undefined
Mar 8, 2024 • 4min

arxiv preprint - Self-correcting LLM-controlled Diffusion Models

In this episode, we discuss Self-correcting LLM-controlled Diffusion Models by Tsung-Han Wu, Long Lian, Joseph E. Gonzalez, Boyi Li, Trevor Darrell. The paper introduces Self-correcting LLM-controlled Diffusion (SLD), a novel approach to improve text-to-image generation by incorporating a loop where an image is generated, evaluated, and corrected iteratively based on a given text prompt using a Language Model (LLM). SLD can be applied to existing diffusion models and has shown proficiency in generating more accurate images, particularly in aspects requiring understanding of numbers, attributes, and spatial relations. The authors also highlight SLD's capability for image editing through prompt modification and announce their intention to make the code publicly available to foster further research.
undefined
Mar 8, 2024 • 4min

arxiv preprint - tinyBenchmarks: evaluating LLMs with fewer examples

In this episode, we discuss tinyBenchmarks: evaluating LLMs with fewer examples by Felipe Maia Polo, Lucas Weber, Leshem Choshen, Yuekai Sun, Gongjun Xu, Mikhail Yurochkin. The paper discusses strategies to minimize the number of evaluations required to effectively assess the performance of large language models on major benchmarks. By analyzing a popular QA benchmark called MMLU, the authors demonstrate that evaluating a language model on merely 100 well-chosen examples can yield an accurate estimate of its performance. The authors have developed and released evaluation tools and condensed versions of benchmarks including Open LLM Leaderboard, MMLU, HELM, and AlpacaEval 2.0, which have been empirically shown to reliably replicate the outcomes of the original expansive evaluations.
undefined
Mar 6, 2024 • 4min

arxiv preprint - Asymmetry in Low-Rank Adapters of Foundation Models

In this episode, we discuss Asymmetry in Low-Rank Adapters of Foundation Models by Jiacheng Zhu, Kristjan Greenewald, Kimia Nadjahi, Haitz Sáez de Ocáriz Borde, Rickard Brüel Gabrielsson, Leshem Choshen, Marzyeh Ghassemi, Mikhail Yurochkin, Justin Solomon. The paper presents an analysis of Low-Rank Adaptation (LoRA), revealing an asymmetry in the roles of the matrices (denoted B and A) involved in updating neural network parameters. It is found that fine-tuning the B matrix is more critical than fine-tuning the A matrix, to the extent that an untrained A can suffice. This insight leads to better parameter efficiency and generalization bounds when only B is trained, with experimental validation on models like RoBERTa and BART-Large, among others, with resources shared on GitHub.
undefined
Mar 5, 2024 • 4min

arxiv preprint - When Scaling Meets LLM Finetuning: The Effect of Data, Model and Finetuning Method

In this episode, we discuss When Scaling Meets LLM Finetuning: The Effect of Data, Model and Finetuning Method by Biao Zhang, Zhongtao Liu, Colin Cherry, Orhan Firat. The paper investigates how various scaling factors impact the effectiveness of finetuning large language models (LLMs), focusing on full-model tuning (FMT) and parameter-efficient tuning (PET). Through experiments with bilingual LLMs and tasks like machine translation and summarization, the authors find that finetuning follows a joint scaling law where increasing model size is more beneficial than increasing the size of the pretraining data, and that PET's additional parameters typically don't improve performance. They conclude that the best finetuning approach depends on the specific task and the amount of finetuning data available, providing insights for selecting and improving LLM finetuning methods.
undefined
Mar 4, 2024 • 4min

arxiv preprint - EMO: Emote Portrait Alive - Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions

In this episode, we discuss EMO: Emote Portrait Alive - Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions by Linrui Tian, Qi Wang, Bang Zhang, Liefeng Bo. The paper presents a new framework named EMO for generating realistic talking head videos, improving the synchronization between audio cues and facial movements. Traditional methods often miss the complexity of human expressions and individual facial characteristics, but EMO overcomes these limitations by directly converting audio to video without relying on 3D models or facial landmarks. This direct synthesis approach results in more expressive and seamlessly animated portrait videos that are better aligned with the audio.
undefined
Mar 1, 2024 • 5min

arxiv preprint - The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits

In this episode, we discuss The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits by Shuming Ma, Hongyu Wang, Lingxiao Ma, Lei Wang, Wenhui Wang, Shaohan Huang, Li Dong, Ruiping Wang, Jilong Xue, Furu Wei. The paper introduces BitNet b1.58, a new 1-bit Large Language Model with ternary parameter values that achieves the same level of accuracy as traditional full-precision models while offering substantial improvements in speed, memory usage, throughput, and energy efficiency. This model represents a breakthrough, establishing a new scaling law for cost-effective and high-performance language model training. Moreover, the development of BitNet b1.58 potentially leads to the creation of specialized hardware optimized for 1-bit language models.
undefined
Feb 29, 2024 • 3min

arxiv preprint - Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models

In this episode, we discuss Assisting in Writing Wikipedia-like Articles From Scratch with Large Language Models by Yijia Shao, Yucheng Jiang, Theodore A. Kanell, Peter Xu, Omar Khattab, Monica S. Lam. The paper examines the use of large language models for creating detailed long-form articles similar to Wikipedia entries, focusing on the preliminary phase of article writing. The authors introduce STORM, a system that uses information retrieval and simulated expert conversations to generate diverse perspectives and build article outlines, paired with a dataset called FreshWiki for evaluation. They find that STORM improves article organization and breadth and identify challenges like source bias and fact relevance for future research in generating well-grounded articles.
undefined
Feb 28, 2024 • 3min

arxiv preprint - LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning

In this episode, we discuss LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning by Hongye Jin, Xiaotian Han, Jingfeng Yang, Zhimeng Jiang, Zirui Liu, Chia-Yuan Chang, Huiyuan Chen, Xia Hu. The paper presents SelfExtend, a novel method for extending the context window of Large Language Models (LLMs) to better handle long input sequences without the need for fine-tuning. SelfExtend incorporates bi-level attention mechanisms to manage dependencies between both distant and adjacent tokens, allowing LLMs to operate beyond their original training constraints. The method has been tested comprehensively, showing its effectiveness, and the code is shared for public use, addressing the key challenge of LLMs' fixed sequence length limitations during inference.
undefined
Feb 27, 2024 • 3min

arxiv preprint - Branch-Solve-Merge Improves Large Language Model Evaluation and Generation

In this episode, we discuss Branch-Solve-Merge Improves Large Language Model Evaluation and Generation by Swarnadeep Saha, Omer Levy, Asli Celikyilmaz, Mohit Bansal, Jason Weston, Xian Li. The paper introduces a program called BRANCH-SOLVE-MERGE (BSM) designed to enhance the performance of Large Language Models (LLMs) on complex natural language tasks. BSM uses a three-module approach that breaks tasks into parallel sub-tasks, solves each independently, and then integrates the results. The implementation of BSM shows significant improvements in LLM tasks such as response evaluation and constrained text generation, increasing human-LLM agreement, reducing biases, and enhancing story coherence and constraint satisfaction.

The AI-powered Podcast Player

Save insights by tapping your headphones, chat with episodes, discover the best highlights - and more!
App store bannerPlay store banner
Get the app