LinkBERT: Improving Language Model Training with Document Link
The Stanford AI Lab Blog
by
1y ago
Language Model Pretraining Language models (LMs), like BERT 1 and the GPT series 2, achieve remarkable performance on many natural language processing (NLP) tasks. They are now the foundation of today’s NLP systems. 3 These models serve important roles in products and tools that we use every day, such as search engines like Google 4 and personal assistants like Alexa 5. These LMs are powerful because they can be pretrained via self-supervised learning on massive amounts of text data on the web without the need for labels, after which the pretrained models can be quickly adapted to a wide range ..read more
Visit website
Stanford AI Lab Papers and Talks at ACL 2022
The Stanford AI Lab Blog
by
1y ago
The 60th Annual Meeting of the Association for Computational Linguistics (ACL) 2022 is taking place May 22nd - May 27th. We’re excited to share all the work from SAIL that’s being presented, and you’ll find links to papers, videos and blogs below. Feel free to reach out to the contact authors directly to learn more about the work that’s happening at Stanford! List of Accepted Papers LinkBERT: Pretraining Language Models with Document Links Authors: Michihiro Yasunaga, Jure Leskovec*, Percy Liang* Contact: myasu@cs.stanford.edu Links: Paper | Website Keywords: language model, pretraining, know ..read more
Visit website
Stanford AI Lab Papers and Talks at ICLR 2022
The Stanford AI Lab Blog
by
1y ago
The International Conference on Learning Representations (ICLR) 2022 is being hosted virtually from April 25th - April 29th. We’re excited to share all the work from SAIL that’s being presented, and you’ll find links to papers, videos and blogs below. Feel free to reach out to the contact authors directly to learn more about the work that’s happening at Stanford! List of Accepted Papers Autonomous Reinforcement Learning: Formalism and Benchmarking Authors: Archit Sharma*, Kelvin Xu*, Nikhil Sardana, Abhishek Gupta, Karol Hausman, Sergey Levine, Chelsea Finn Contact: architsh@stanford.edu Link ..read more
Visit website
Discovering the systematic errors made by machine learning models
The Stanford AI Lab Blog
by
1y ago
Discovering systematic errors with cross-modal embeddings In this blog post, we introduce Domino, a new approach for discovering systematic errors made by machine learning models. We also discuss a framework for quantitatively evaluating methods like Domino. Links: ? Paper (ICLR 2022) ? Longer Walkthrough ? GitHub ? Docs ? Google Colab Machine learning models that achieve high overall accuracy often make systematic errors on coherent slices of validation data. What is a slice? A slice is a set of data samples that share a common characteristic. As an example, in large image datasets, photos ..read more
Visit website
Grading Complex Interactive Coding Programs with Reinforcement Learning
The Stanford AI Lab Blog
by
1y ago
[Summary] tl;dr: A tremendous amount of effort has been poured into training AI algorithms to competitively play games that computers have traditionally had trouble with, such as the retro games published by Atari, Go, DotA, and StarCraft II. The practical machine learning knowledge accumulated in developing these algorithms has paved the way for people to now routinely train game-playing AI agents for many games. Following this line of work, we focus on a specific category of games – those developed by students as part of a programming assignment. Can the same algorithms that master Atari gam ..read more
Visit website
Understanding Deep Learning Algorithms that Leverage Unlabeled Data, Part 1: Self-training
The Stanford AI Lab Blog
by
1y ago
Deep models require a lot of training examples, but labeled data is difficult to obtain. This motivates an important line of research on leveraging unlabeled data, which is often more readily available. For example, large quantities of unlabeled image data can be obtained by crawling the web, whereas labeled datasets such as ImageNet require expensive labeling procedures. In recent empirical developments, models trained with unlabeled data have begun to approach fully-supervised performance (e.g., Chen et al., 2020, Sohn et al., 2020). This series of blog posts will discuss our theoretical wor ..read more
Visit website
Stanford AI Lab Papers and Talks at AAAI 2022
The Stanford AI Lab Blog
by
1y ago
The 36th AAAI Conference on Artificial Intelligence (AAAI 2022) is being hosted virtually from February 22th - March 1st. We’re excited to share all the work from SAIL that’s being presented, and you’ll find links to papers, videos and blogs below. Feel free to reach out to the contact authors directly to learn more about the work that’s happening at Stanford. List of Accepted Papers Partner-Aware Algorithms in Decentralized Cooperative Bandit Teams Authors: Erdem Bıyık, Anusha Lalitha, Rajarshi Saha, Andrea Goldsmith, Dorsa Sadigh Contact: ebiyik@stanford.edu Links: Paper | Video | 2nd Video ..read more
Visit website
How to Improve User Experience (and Behavior): Three Papers from Stanford's Alexa Prize Team
The Stanford AI Lab Blog
by
1y ago
Introduction In 2019, Stanford entered the Alexa Prize Socialbot Grand Challenge 3 for the first time, with its bot Chirpy Cardinal, which went on to win 2nd place in the competition. In our previous post, we discussed the technical structure of our socialbot and how developers can use our open-source code to develop their own. In this post we share further research conducted while developing Chirpy Cardinal to discover common pain points that users encounter when interacting with socialbots, and strategies for addressing them. The Alexa Prize is a unique research setting, as it allows researc ..read more
Visit website
Reward Isn't Free: Supervising Robot Learning with Language and Video from the Web
The Stanford AI Lab Blog
by
1y ago
This work was conducted as part of SAIL and CRFM. Deep learning has enabled improvements in the capabilities of robots on a range of problems such as grasping 1 and locomotion 2 in recent years. However, building the quintessential home robot that can perform a range of interactive tasks, from cooking to cleaning, in novel environments has remained elusive. While a number of hardware and software challenges remain, a necessary component is robots that can generalize their prior knowledge to new environments, tasks, and objects in a zero or few shot manner. For example, a home robot tasked with ..read more
Visit website
BanditPAM: Almost Linear-Time k-medoids Clustering via Multi-Armed Bandits
The Stanford AI Lab Blog
by
1y ago
TL;DR Want something better than \(k\)-means? Our state-of-the-art \(k\)-medoids algorithm from NeurIPS, BanditPAM, is now publicly available! \(\texttt{pip install banditpam}\) and you're good to go! Like the \(k\)-means problem, the \(k\)-medoids problem is a clustering problem in which our objective is to partition a dataset into disjoint subsets. In \(k\)-medoids, however, we require that the cluster centers must be actual datapoints, which permits greater interpretability of the cluster centers. \(k\)-medoids also works better with arbitrary distance metrics, so your clustering can be mo ..read more
Visit website

Follow The Stanford AI Lab Blog on FeedSpot

Continue with Google
Continue with Apple
OR