RoPE Rotary Position Embedding to 100K context length

  Рет қаралды 2,444

code_your_own_AI

code_your_own_AI

Ай бұрын

ROPE - Rotary Position Embedding explained in simple terms for calculating the self attention in Transformers with a relative position encoding for extended Context lengths of LLMs.
All rights w/ authors:
ROFORMER: ENHANCED TRANSFORMER WITH ROTARY POSITION EMBEDDING (RoPE)
arxiv.org/pdf/2104.09864
#airesearch
#aiexplained

Пікірлер: 8
@LamontCranston-qh2rv
@LamontCranston-qh2rv Ай бұрын
Thank you SO MUCH for providing such high quality conten! Very much enjoying all your many videos! If you have a chance, I'd love to see you discuss the recent work in giving AI spatial reasoning. I.e. artificial "imagination". (In it's natural form, very much a core feature of human thought.) Perhaps one might think about the creation of a "right brain" to go along with the "left brain" language models we have now? (Please forgive the over-simplification of human neuroscience.) Thanks again! All the best to you sincerely!
@desmur36
@desmur36 Ай бұрын
Amazing content! The explanations are SO clear! Thank you!
@paratracker
@paratracker Ай бұрын
Maybe it's obvious to YOU that the solution is that complex exponential, but I wish you hadn't assumed that WE would all see that as self-evident as you do.
@code4AI
@code4AI Ай бұрын
I see what you mean. You know, I spend some days to find simple explanations for the not so self explanatory RoPE algo, especially I will build on this in my second video, and then we examine more complex, more recent ideas about RoPE. I decided for a way, that will enable my audience to understand the main ideas and methods, and go from there. I recorded 90 min for the second part, and currently I cut it to max 60 min, striking a balance of providing insights for all my viewers. I'll try harder ....
@AYUSHSINGH-db6ev
@AYUSHSINGH-db6ev Ай бұрын
Hi Sir! Really love your videos! How can we access your presentation slides?
@mshonle
@mshonle Ай бұрын
If one rotation is good, how about going into three dimensional rotations and using quaternions? Is there any work using that?
@hangjianyu
@hangjianyu 26 күн бұрын
there is a mistake, smaller dimensions change more quickly,and large dimensions change more slowly
5 Easy Ways to help LLMs to Reason
50:37
code_your_own_AI
Рет қаралды 3,9 М.
LongRoPE & Theta Scaling to 1 Mio Token (2/2)
58:30
code_your_own_AI
Рет қаралды 1,1 М.
LOVE LETTER - POPPY PLAYTIME CHAPTER 3 | GH'S ANIMATION
00:15
Did you believe it was real? #tiktok
00:25
Анастасия Тарасова
Рет қаралды 56 МЛН
Positional encodings in transformers (NLP817 11.5)
19:29
Herman Kamper
Рет қаралды 1,9 М.
Rotary Positional Embeddings: Combining Absolute and Relative
11:17
Efficient NLP
Рет қаралды 27 М.
Q* explained: Complex Multi-Step AI Reasoning
55:11
code_your_own_AI
Рет қаралды 7 М.
This is What Limits Current LLMs
7:05
Edan Meyer
Рет қаралды 93 М.
LLM - Reasoning SOLVED (new research)
47:51
code_your_own_AI
Рет қаралды 15 М.
Self-Attention Using Scaled Dot-Product Approach
16:09
Machine Learning Studio
Рет қаралды 13 М.
Собери ПК и Получи 10,000₽
1:00
build monsters
Рет қаралды 2,7 МЛН
Мой инст: denkiselef. Как забрать телефон через экран.
0:54
⚡️Супер БЫСТРАЯ Зарядка | Проверка
1:00