Training Diffusion Models with Reinforcement Learning – Berkeley Artificial Intelligence Research Blog

Training Diffusion Models with Reinforcement Learning - Berkeley Artificial Intelligence Research Blog

Training diffusion models with reinforcement learning

Diffusion models have recently emerged as the de facto standard for generating complex, high-dimensional output. You may know them for their productivity. Stunning AI art and highly realistic synthetic imagesbut they have also found success in other applications such as Drug design And Continuous control. The key idea behind diffusion models is to iteratively transform random noise into a sample, such as an image or protein structure. This is usually motivated as a Maximum likelihood estimation problem, where the model is trained to produce patterns that match the training data as closely as possible.

However, most use cases for diffusion models are not directly related to matching training data, but instead with a downstream objective. We don't just want an image that looks like existing images, but an image that has a unique appearance. We don't just want a drug molecule that is physiologically plausible, but one that is as effective as possible. In this post, we show how diffusion models can be trained directly on these downstream targets using reinforcement learning (RL). To do this, we finetune Stable diffusion for a number of purposes, including image compression, human-perceived aesthetic quality, and instant image alignment. The last of these purposes uses feedback. A large model of vision language to improve the model's performance on anomalous signals, showing how Powerful AI models can be used to improve each other. Without a human in the loop.

Read more

Rethinking the role of PPO in RLHF – Berkeley Artificial Intelligence Research Blog

Rethinking the role of PPO in RLHF - Berkeley Artificial Intelligence Research Blog

Rethinking the role of PPO in RLHF

TL; DR: In RLHF, there is a tension between the reward learning phase, which uses human preferences in the form of comparisons, and the RL fine-tuning phase, which optimizes a single, non-comparable reward. What if we perform RL in a comparative fashion?

Figure 1:
This diagram illustrates the difference between reinforcement learning. absolute Roy and Relative Feedback By adding a new component – ​​the pairwise policy gradient, we can combine the reward modeling stage and the RL stage, enabling direct updates based on pairwise responses.

Large language models (LLMs) have powered increasingly capable virtual assistants, e.g GPT-4, Clade-2, Bard And Bing Chat. These systems can answer complex user questions, write code, and even generate poetry. The basic technique of these amazing virtual assistants is Reinforcement Learning with Human Feedback (RLHF). The purpose of RLHF is to adapt the model to human values ​​and eliminate unintended behaviors, which can often arise because the model is subjected to a large amount of low-quality data during its pre-training phase. should be brought forward.

Proximity Policy Reform (PPO), has been reported to reveal the dominant RL optimizer, in this process. Instability And Implementation complications. More importantly, there is a persistent paradox in the RLHF process: despite training the reward model using comparisons between different responses, the RL fine-tuning step operates without comparisons on individual responses. This inconsistency can exacerbate problems, especially in the domain of challenging language generation.

Given this background, an interesting question arises: Is it possible to design an RL algorithm that learns in a comparative fashion? To explore this, we optimize the pairwise proximity policy (P3O), a method that synchronizes the training process in both the reward learning stage and the RL fine-tuning stage of RLHF, provides a satisfactory solution to this problem.

Read more

Asymmetric Inference Robustness via Feature-Convex Neural Networks – Berkeley Artificial Intelligence Research Blog

Diagram illustrating the FCNN architecture

Asymmetric Verification Robustness by Feature-Convex Neural Networks

TLDR: We recommend. Asymmetric certified strength The problem, which requires proven robustness for only one class and reflects real-world adversarial scenarios. This focused setting allows us to introduce feature-convex classifiers, which produce closed-form and deterministic validation radii on the order of milliseconds.

Figure 1. Example of feature-convex classifiers and their validation for sensitive-class input. This architecture constructs a Lipschitz-continuous feature map $varphi$ with a learned convex function $g$ . Since $g$ is convex, it is globally less than its tangent plane at $varphi(x)$ , yielding certified normal balls in the characteristic space. The Lipschutzness of $varphi$ then obtains an appropriate scale certificate in the original input space.

Despite their widespread use, deep learning classifiers are severely vulnerable. Counterexamples: Small, humanly imperceptible image glitches that fool machine learning models into misclassifying modified input. This vulnerability severely compromises the reliability of security-critical processes involving machine learning. Many experimental defenses have been proposed against hostile perturbations—often only later defeated by stronger attack strategies. So we focus Certified robust classifierswhich provide a mathematical guarantee that their predictions will be consistent for the $ell_p$ -norm ball around an input.

Traditional certification robustness methods have many drawbacks, including nondeterminism, slow execution, poor scaling, and certification against the one-attack-only principle. We argue that these problems can be addressed by refining the deterministic robustness problem to be more compatible with practical adversarial settings.

Read more

Detecting Written Text Using Large Language Models – Berkeley Artificial Intelligence Research Blog

Detecting Written Text Using Large Language Models - Berkeley Artificial Intelligence Research Blog


Structure of Ghostbuster, our state-of-the-art method for AI-generated text detection.

Big language models like ChatGPT write impressively—so well, in fact, that they've become a problem. Students have started using these models to ghostwrite assignments, which is why some schools are moving in that direction. Ban ChatGPT. In addition, these models are prone to generating text with factual errors, so astute readers may want to know if generative AI tools are used to ghost news articles or other sources before trusting them. Done to write.

What can teachers and users do? Existing tools for AI-generated text detection sometimes perform poorly on data that differs from what they were trained on. In addition, if these models classify real human writing as AI-generated, they may endanger students whose real work is questioned.

Our recent paper Introducing Ghostbuster, an advanced AI-powered text detection method. Ghostbuster works by finding the probability of generating each token in a document under several weak language models, then combining functions based on those probabilities as input for the final classification. Ghostbuster does not need to know which model was used to create the document, nor the possibility of creating the document under that particular model. This property makes Ghostbuster particularly useful for detecting potentially generated text from an unknown model or black box model, such as the popular commercial models ChatGPT and Claude, for which probabilities are not available. We were particularly interested in making sure that Ghostbusters was well-publicized, so we explored a number of ways the text could be produced, including new collections of articles, news and stories from different domains. using data sets), language models, or notations.

Read more

2024 BAIR Graduate Directory – Berkeley Artificial Intelligence Research Blog

Abdul Salam Azad


Every year, the Berkeley Artificial Intelligence Research (BAIR) Lab graduates some of the most talented and innovative minds in artificial intelligence and machine learning. Our Ph.D. The graduates have each pushed the boundaries of AI research and are now ready for new adventures in academia, industry and beyond.

These amazing individuals bring with them a wealth of knowledge, fresh ideas, and a drive to continue contributing to the advancement of AI. His work at BAIR, from deep learning, robotics, and natural language processing to computer vision, security, and more, has made significant contributions to his fields and has had a transformative impact on society.

This website is dedicated to showcasing our colleagues, making it easier for academic institutions, research organizations, and industry leaders to discover and recruit the next generation of AI pioneers. Here, you'll find detailed profiles, research interests, and contact information for each of our graduates. We invite you to explore the potential collaborations and opportunities these graduates seek to apply their skills and insights in a new environment.

Join us in celebrating the achievements of BAIR's latest PhD graduates. Their journey has just begun, and the future they will help build is bright!

Read more

Modeling Very Large Images with XT – Berkeley Artificial Intelligence Research Blog

Modeling Very Large Images with XT - Berkeley Artificial Intelligence Research Blog


As computer vision researchers, we believe that every pixel can tell a story. However, a writer's block seems to be kicking in when it comes to dealing with big pictures. Big images are no longer rare—the cameras we carry in our pockets and orbit our planet capture images so large and detailed that they stretch our current best models and hardware to their breaking points. . Typically, we encounter a quadratic increase in memory usage as a function of image size.

Today, we choose one of two suboptimal choices when handling large images: downsampling or cropping. Both of these methods result in a significant loss in the amount of information and context contained in the image. We take another look at these approaches and introduce $x$T, a new framework for end-to-end modeling of large images on contemporary GPUs while maintaining global context with local details. The race will be collected effectively.

Architecture for the $x$T framework.

Read more

Visual Haystacks Benchmark! – Berkeley Artificial Intelligence Research Blog

Visual Haystacks Benchmark!  - Berkeley Artificial Intelligence Research Blog


Humans are skilled at processing vast arrays of visual information, a skill critical to achieving artificial general intelligence (AGI). Over the decades, AI researchers have developed visual question answering (VQA) systems to interpret scenes within an image and answer related questions. Although recent advances in foundational models have significantly bridged the gap between human and machine visual processing, traditional VQA has been limited to causality. alone images at a time rather than the entire collection of visual data.

This limitation creates challenges in more complex scenarios. For example, the challenges of discerning patterns in medical image collections, monitoring deforestation through satellite imagery, mapping urban changes using autonomous navigation data, analyzing thematic elements in large art collections, or understanding consumer behavior from retail surveillance footage. Each of these scenarios involves not only visual processing of hundreds or thousands of images, but also requires cross-image processing of these results. To fill this gap, this project focuses on the “multi-image question answering” (MIQA) task, which is beyond the reach of traditional VQA systems.

Visual Haystacks: The first “visual-centric” Needle-In-A-Haystack (NIAH) benchmark designed to rigorously evaluate large multimodal models (LMMs) in processing long-context visual information.

Read more