Virtual Personas for Language Models via an Anthology of Backstories
The Berkeley Artificial Intelligence Research Blog
by
4M ago
We introduce Anthology, a method for conditioning LLMs to representative, consistent, and diverse virtual personas by generating and utilizing naturalistic backstories with rich details of individual values and experience. What does it mean for large language models (LLMs) to be trained on massive text corpora, collectively produced by millions and billions of distinctive human authors? In “Language Models as Agent Models”, compelling evidence suggests that recent language models could be considered models of agents: provided with a textual context, LLMs are capable of generating conditional ..read more
Visit website
Linguistic Bias in ChatGPT: Language Models Reinforce Dialect Discrimination
The Berkeley Artificial Intelligence Research Blog
by
5M ago
Sample language model responses to different varieties of English and native speaker reactions. ChatGPT does amazingly well at communicating with people in English. But whose English? Only 15% of ChatGPT users are from the US, where Standard American English is the default. But the model is also commonly used in countries and communities where people speak other varieties of English. Over 1 billion people around the world speak varieties such as Indian English, Nigerian English, Irish English, and African-American English. Speakers of these non-“standard” varieties often face discrimination i ..read more
Visit website
How to Evaluate Jailbreak Methods: A Case Study with the StrongREJECT Benchmark
The Berkeley Artificial Intelligence Research Blog
by
6M ago
When we began studying jailbreak evaluations, we found a fascinating paper claiming that you could jailbreak frontier LLMs simply by translating forbidden prompts into obscure languages. Excited by this result, we attempted to reproduce it and found something unexpected. The paper in question claimed an impressive 43% success rate in jailbreaking GPT-4 by translating forbidden prompts into Scots Gaelic (Yong et al., 2023). To showcase their method, the authors asked GPT-4 to provide instructions for building a homemade explosive device using household materials. They translated the prompt into ..read more
Visit website
Are We Ready for Multi-Image Reasoning? Launching VHs: The Visual Haystacks Benchmark!
The Berkeley Artificial Intelligence Research Blog
by
8M ago
Humans excel at processing vast arrays of visual information, a skill that is crucial for achieving artificial general intelligence (AGI). Over the decades, AI researchers have developed Visual Question Answering (VQA) systems to interpret scenes within single images and answer related questions. While recent advancements in foundation models have significantly closed the gap between human and machine visual processing, conventional VQA has been restricted to reason about only single images at a time rather than whole collections of visual data. This limitation poses challenges in more complex ..read more
Visit website
TinyAgent: Function Calling at the Edge
The Berkeley Artificial Intelligence Research Blog
by
10M ago
The ability of LLMs to execute commands through plain language (e.g. English) has enabled agentic systems that can complete a user query by orchestrating the right set of tools (e.g. ToolFormer, Gorilla). This, along with the recent multi-modal efforts such as the GPT-4o or Gemini-1.5 model, has expanded the realm of possibilities with AI agents. While this is quite exciting, the large model size and computational requirements of these models often requires their inference to be performed on the cloud. This can create several challenges for their widespread adoption. First and foremost, uploa ..read more
Visit website
Modeling Extremely Large Images with xT
The Berkeley Artificial Intelligence Research Blog
by
1y ago
As computer vision researchers, we believe that every pixel can tell a story. However, there seems to be a writer’s block settling into the field when it comes to dealing with large images. Large images are no longer rare—the cameras we carry in our pockets and those orbiting our planet snap pictures so big and detailed that they stretch our current best models and hardware to their breaking points when handling them. Generally, we face a quadratic increase in memory usage as a function of image size. Today, we make one of two sub-optimal choices when handling large images: down-sampling or cr ..read more
Visit website
2024 BAIR Graduate Directory
The Berkeley Artificial Intelligence Research Blog
by
1y ago
Every year, the Berkeley Artificial Intelligence Research (BAIR) Lab graduates some of the most talented and innovative minds in artificial intelligence and machine learning. Our Ph.D. graduates have each expanded the frontiers of AI research and are now ready to embark on new adventures in academia, industry, and beyond. These fantastic individuals bring with them a wealth of knowledge, fresh ideas, and a drive to continue contributing to the advancement of AI. Their work at BAIR, ranging from deep learning, robotics, and natural language processing to computer vision, security, and much more ..read more
Visit website
The Shift from Models to Compound AI Systems
The Berkeley Artificial Intelligence Research Blog
by
1y ago
AI caught everyone’s attention in 2023 with Large Language Models (LLMs) that can be instructed to perform general tasks, such as translation or coding, just by prompting. This naturally led to an intense focus on models as the primary ingredient in AI application development, with everyone wondering what capabilities new LLMs will bring. As more developers begin to build using LLMs, however, we believe that this focus is rapidly changing: state-of-the-art AI results are increasingly obtained by compound systems with multiple components, not just monolithic models. For example, Google’s AlphaC ..read more
Visit website
Ghostbuster: Detecting Text Ghostwritten by Large Language Models
The Berkeley Artificial Intelligence Research Blog
by
1y ago
The structure of Ghostbuster, our new state-of-the-art method for detecting AI-generated text. Large language models like ChatGPT write impressively well—so well, in fact, that they’ve become a problem. Students have begun using these models to ghostwrite assignments, leading some schools to ban ChatGPT. In addition, these models are also prone to producing text with factual errors, so wary readers may want to know if generative AI tools have been used to ghostwrite news articles or other sources before trusting them. What can teachers and consumers do? Existing tools to detect AI-generated t ..read more
Visit website
Asymmetric Certified Robustness via Feature-Convex Neural Networks
The Berkeley Artificial Intelligence Research Blog
by
1y ago
Asymmetric Certified Robustness via Feature-Convex Neural Networks TLDR: We propose the asymmetric certified robustness problem, which requires certified robustness for only one class and reflects real-world adversarial scenarios. This focused setting allows us to introduce feature-convex classifiers, which produce closed-form and deterministic certified radii on the order of milliseconds. Figure 1. Illustration of feature-convex classifiers and their certification for sensitive-class inputs. This architecture composes a Lipschitz-continuous feature map $\varphi$ with a learned convex functio ..read more
Visit website

Follow The Berkeley Artificial Intelligence Research Blog on FeedSpot

Continue with Google
Continue with Apple
OR