Core AI
Large Language Models
Deep dives into GPT, Claude, Gemini, Llama and the transformers powering modern AI
Skills in this topic
5 skills — Sign in to track your progress
LLM Foundations
beginner
Explain how transformers generate text
Prompt Craft
beginner
Write zero-shot and few-shot prompts
LLM Engineering
intermediate
Call LLM APIs with function/tool use
Fine-tuning LLMs
advanced
Prepare fine-tuning datasets
Multimodal LLMs
advanced
Use GPT-4V / Claude Vision for image understanding
Showing 5,770 reads from curated sources
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Jukebox
We’re introducing Jukebox, a neural net that generates music, including rudimentary singing, as raw audio in a variety of genres and artist styles. We’re releas
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Improving verifiability in AI development
We’ve contributed to a multi-stakeholder report by 58 co-authors at 30 organizations, including the Centre for the Future of Intelligence, Mila, Schwartz Reisma
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
OpenAI standardizes on PyTorch
We are standardizing OpenAI’s deep learning framework on PyTorch.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Procgen Benchmark
We’re releasing Procgen Benchmark, 16 simple-to-use procedurally-generated environments which provide a direct measure of how quickly a reinforcement learning a
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Safety Gym
We’re releasing Safety Gym, a suite of environments and tools for measuring progress towards reinforcement learning agents that respect safety constraints while
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
6y ago
Self-Supervised Representation Learning
[Updated on 2020-01-09: add a new section on Contrastive Predictive Coding ]. [Updated on 2020-04-13: add a “Momentum Contra
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
GPT-2: 1.5B release
As the final model release of GPT-2’s staged release, we’re releasing the largest version (1.5B parameters) of GPT-2 along with code and model weights to facili
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Solving Rubik’s Cube with a robot hand
We’ve trained a pair of neural networks to solve the Rubik’s Cube with a human-like robot hand. The neural networks are trained entirely in simulation, using th
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
OpenAI Scholars 2020: Applications open
We are now accepting applications for our third class of OpenAI Scholars.
Distill.pub
🧠 Large Language Models
📄 Paper
⚡ AI Lesson
6y ago
The Paths Perspective on Value Learning
A closer look at how Temporal Difference Learning merges paths of experience for greater statistical efficiency
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Fine-tuning GPT-2 from human preferences
We’ve fine-tuned the 774M parameter GPT-2 language model using human feedback for various tasks, successfully matching the preferences of the external human lab
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
6y ago
Evolution Strategies
Stochastic gradient descent is a universal choice for optimizing deep learning models. However, it is not the only option. With bl
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Testing robustness against unforeseen adversaries
We’ve developed a method to assess whether a neural network classifier can reliably defend against adversarial attacks not seen during training. Our method yiel
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
GPT-2: 6-month follow-up
We’re releasing the 774 million parameter GPT-2 language model after the release of our small 124M model in February, staged release of our medium 355M model in
Distill.pub
🧠 Large Language Models
📄 Paper
⚡ AI Lesson
6y ago
A Discussion of 'Adversarial Examples Are Not Bugs, They Are Features': Learning from Incorrectly Labeled Data
Section 3.2 of Ilyas et al. (2019) shows that training a model on only adversarial errors leads to non-trivial generalization on the original test set. We show
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Microsoft invests in and partners with OpenAI to support us building beneficial AGI
Microsoft is investing $1 billion in OpenAI to support us building artificial general intelligence (AGI) with widely distributed economic benefits. We’re partne
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
Why responsible AI development needs cooperation on safety
We’ve written a policy research paper identifying four strategies that can be used today to improve the likelihood of long-term industry cooperation on safety n
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
6y ago
Meta Reinforcement Learning
In my earlier post on meta-learning , the problem is mainly defined in the context of few-shot classificati
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
OpenAI Robotics Symposium 2019
We hosted the first OpenAI Robotics Symposium on April 27, 2019.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
OpenAI Scholars 2019: Final projects
Our second class of OpenAI Scholars has concluded, with all eight scholars producing an exciting final project showcased at Scholars Demo Day at OpenAI.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
6y ago
OpenAI Fellows Fall 2018: Final projects
Our second class of OpenAI Fellows has wrapped up, with each Fellow going from a machine learning beginner to core OpenAI contributor in the course of a 6-month
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
6y ago
Domain Randomization for Sim2Real Transfer
In Robotics, one of the hardest problems is how to make your model transfer to the real world. Due to the sample inefficiency of deep RL algorithms and the cost
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
MuseNet
We’ve created MuseNet, a deep neural network that can generate 4-minute musical compositions with 10 different instruments, and can combine styles from country
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Generative modeling with sparse transformers
We’ve developed the Sparse Transformer, a deep neural network which sets new records at predicting what comes next in a sequence—whether text, images, or sound.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Five defeats Dota 2 world champions
OpenAI Five is the first AI to beat the world champions in an esports game, having won two back-to-back games versus the world champion Dota 2 team, OG, at Fina
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Five Finals
We’ll be holding our final live event for OpenAI Five at 11:30am PT on April 13.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Implicit generation and generalization methods for energy-based models
We’ve made progress towards stable and scalable training of energy-based models (EBMs) resulting in better sample quality and generalization ability than existi
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Scholars 2019: Meet our Scholars
Our class of eight scholars (out of 550 applicants) brings together collective expertise in literature, philosophy, cell biology, statistics, economics, quantum
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Introducing Activation Atlases
We’ve created activation atlases (in collaboration with Google researchers), a new technique for visualizing what interactions between neurons can represent. As
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Neural MMO: A massively multiagent game environment
We’re releasing a Neural MMO, a massively multiagent game environment for reinforcement learning agents. Our platform supports a large, variable number of agent
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Spinning Up in Deep RL: Workshop review
On February 2, we held our first Spinning Up Workshop as part of our new education initiative at OpenAI.
Distill.pub
🧠 Large Language Models
📄 Paper
⚡ AI Lesson
7y ago
AI Safety Needs Social Scientists
If we want to train AI to do what humans want, we need to study humans.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
AI safety needs social scientists
We’ve written a paper arguing that long-term AI safety research needs social scientists to ensure AI alignment algorithms succeed when actual humans are involve
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Better language models and their implications
We’ve trained a large-scale unsupervised language model which generates coherent paragraphs of text, achieves state-of-the-art performance on many language mode
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
7y ago
Generalized Language Models
[Updated on 2019-02-14: add ULMFiT and GPT-2 .] [Updated on 2020-02-29: add ALBERT .] <span class="updat
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Fellows Summer 2018: Final projects
Our first cohort of OpenAI Fellows has concluded, with each Fellow going from a machine learning beginner to core OpenAI contributor in the course of a 6-month
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
How AI training scales
We’ve discovered that the gradient noise scale, a simple statistical metric, predicts the parallelizability of neural network training on a wide range of tasks.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Quantifying generalization in reinforcement learning
We’re releasing CoinRun, a training environment which provides a metric for an agent’s ability to transfer its experience to novel situations and has already he
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
7y ago
Meta-Learning: Learning to Learn Fast
[Updated on 2019-10-01: thanks to Tianhao, we have
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Learning concepts with energy functions
We’ve developed an energy-based model that can quickly learn to identify and generate instances of concepts, such as near, above, between, closest, and furthest
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Reinforcement learning with prediction-based rewards
We’ve developed Random Network Distillation (RND), a prediction-based method for encouraging reinforcement learning agents to explore their environments through
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
Learning complex goals with iterated amplification
We’re proposing an AI safety technique called iterated amplification that lets us specify complicated behaviors and goals that are beyond human scale, by demons
Lilian Weng's Blog
🧠 Large Language Models
⚡ AI Lesson
7y ago
Flow-based Deep Generative Models
So far, I’ve written about two types of generative models, GAN and VAE . Neither of them explicitly learns the probability density function of real da
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Scholars 2019: Applications open
We are now accepting applications for our second cohort of OpenAI Scholars, a program where we provide 6–10 stipends and mentorship to individuals from underrep
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Fellows Winter 2019 & Interns Summer 2019
We are now accepting applications for OpenAI Fellows and Interns for 2019.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Scholars 2018: Final projects
Our first cohort of OpenAI Scholars has now completed the program.
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
The International 2018: Results
OpenAI Five lost two games against top Dota 2 players at The International in Vancouver this week, maintaining a good chance of winning for the first 20–35 minu
OpenAI News
🧠 Large Language Models
⚡ AI Lesson
7y ago
OpenAI Five Benchmark: Results
Yesterday, OpenAI Five won a best-of-three against a team of 99.95th percentile Dota players: Blitz, Cap, Fogged, Merlini, and MoonMeander—four of whom have pla
DeepCamp AI