Reinforcement Learning

Reinforcement Learning

Reinforcement learning, self-supervised learning, agents

10 articles5 sources

Top 10 Articles

AlphaGenome Deciphers Non-Coding DNA for Gene Regulation
IEEE Spectrum AI
IEEE Spectrum AI

AlphaGenome Deciphers Non-Coding DNA for Gene Regulation

When AlphaFold solved the protein-folding problem in 2020, it showed that artificial intelligence could crack one of biology’s deepest mysteries: how a string of amino acids folds itself into a working molecular machine. The team at Google DeepMind behind that Nobel Prize-winning platform then turn

#AI#Google#TPU
View Original
AI Hunts for the Next Big Thing in Physics
IEEE Spectrum AI
IEEE Spectrum AI

AI Hunts for the Next Big Thing in Physics

In 1930, a young physicist named Carl D. Anderson was tasked by his mentor with measuring the energies of cosmic rays—particles arriving at high speed from outer space. Anderson built an improved version of a cloud chamber, a device that visually records the trajectories of particles. In 1932, he sa

#AI#TPU#AGI
View Original
Don’t Regulate AI Models. Regulate AI Use
IEEE Spectrum AI
IEEE Spectrum AI

Don’t Regulate AI Models. Regulate AI Use

At times, it can seem like efforts to regulate and rein in AI are everything, everywhere, all at once. China issued the first AI-specific regulations in 2021. The focus is squarely on providers and content governance, enforced through platform control and recordkeeping requirements. In Europe, the

#AI#TPU#RAG
View Original
RL without TD learning
BAIR Blog
BAIR Blog

RL without TD learning

In this post, I’ll introduce a reinforcement learning (RL) algorithm based on an “alternative” paradigm: divide and conquer. Unlike traditional methods, this algorithm is not based on temporal difference (TD) learning (which has scalability challenges), and scales well to long-horizon tasks. We can

#AI#LLM#AGI
View Original
What exactly does word2vec learn?
BAIR Blog
BAIR Blog

What exactly does word2vec learn?

What exactly does word2vec learn, and how? Answering this question amounts to understanding representation learning in a minimal yet interesting language modeling task. Despite the fact that word2vec is a well-known precursor to modern language models, for many years, researchers lacked a quantitati

#AI#LLM
View Original
Whole-Body Conditioned Egocentric Video Prediction
BAIR Blog
BAIR Blog

Whole-Body Conditioned Egocentric Video Prediction

.modal { display: none; position: fixed; z-index: 9999; padding-top: 50px; left: 0; top: 0; width: 100%; height: 100%; overflow: auto; background-color: rgba(0,0,0,0.9); } .modal-content { margin: auto; display: block; max-width: 90%; max-height: 90%; } .close { posit

#AI#Agent#RAG
View Original

Browse Other Categories