File Edit Format We define multi head self attention like below: Y (X) = Concat [H1, HH]W(O) Hh oh Sof tmax [QhKT√hDkh] Vh XW(q) h Kh = XW(k)h XWCKDh Vh XWCVDh It includes some redundancy in consecutive multiplications of matrix w(v) corresponding to every head and also output matrix W(o). Removing this redundancy enables us to multi head self attention as sum of the effect of every head. Now prove we can write multihead self attention formula as below: Y (X) = ΣHh=1 Softmax [QhKTVhDkh] xw(h) (Hint: w(h) equals to wh (v)wh (o) if we devide matrix W(o) in horizontal direction as the number of heads then wh(o) is for the hth head) write

Database System Concepts
7th Edition
ISBN:9780078022159
Author:Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher:Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Chapter1: Introduction
Section: Chapter Questions
Problem 1PE
icon
Related questions
Question

Alert dont submit AI generated answer.

File Edit Format
We define multi head self attention like below:
Y (X) = Concat [H1,
HH]W(O)
Hh
oh
Sof tmax [QhKT√hDkh] Vh
XW(q) h Kh = XW(k)h
XWCKDh
Vh
XWCVDh
It includes some redundancy in consecutive multiplications of matrix w(v)
corresponding
to every head and also output matrix W(o). Removing this redundancy enables us to
multi head self attention as sum of the effect of every head. Now prove we can write
multihead self attention formula as below:
Y (X) = ΣHh=1 Softmax [QhKTVhDkh] xw(h)
(Hint: w(h) equals to wh (v)wh (o) if we devide matrix W(o) in horizontal direction
as the number of heads then wh(o) is for the hth head)
write
Transcribed Image Text:File Edit Format We define multi head self attention like below: Y (X) = Concat [H1, HH]W(O) Hh oh Sof tmax [QhKT√hDkh] Vh XW(q) h Kh = XW(k)h XWCKDh Vh XWCVDh It includes some redundancy in consecutive multiplications of matrix w(v) corresponding to every head and also output matrix W(o). Removing this redundancy enables us to multi head self attention as sum of the effect of every head. Now prove we can write multihead self attention formula as below: Y (X) = ΣHh=1 Softmax [QhKTVhDkh] xw(h) (Hint: w(h) equals to wh (v)wh (o) if we devide matrix W(o) in horizontal direction as the number of heads then wh(o) is for the hth head) write
AI-Generated Solution
AI-generated content may present inaccurate or offensive content that does not represent bartleby’s views.
steps

Unlock instant AI solutions

Tap the button
to generate a solution

Recommended textbooks for you
Database System Concepts
Database System Concepts
Computer Science
ISBN:
9780078022159
Author:
Abraham Silberschatz Professor, Henry F. Korth, S. Sudarshan
Publisher:
McGraw-Hill Education
Starting Out with Python (4th Edition)
Starting Out with Python (4th Edition)
Computer Science
ISBN:
9780134444321
Author:
Tony Gaddis
Publisher:
PEARSON
Digital Fundamentals (11th Edition)
Digital Fundamentals (11th Edition)
Computer Science
ISBN:
9780132737968
Author:
Thomas L. Floyd
Publisher:
PEARSON
C How to Program (8th Edition)
C How to Program (8th Edition)
Computer Science
ISBN:
9780133976892
Author:
Paul J. Deitel, Harvey Deitel
Publisher:
PEARSON
Database Systems: Design, Implementation, & Manag…
Database Systems: Design, Implementation, & Manag…
Computer Science
ISBN:
9781337627900
Author:
Carlos Coronel, Steven Morris
Publisher:
Cengage Learning
Programmable Logic Controllers
Programmable Logic Controllers
Computer Science
ISBN:
9780073373843
Author:
Frank D. Petruzella
Publisher:
McGraw-Hill Education