NLP-Transformers Last  Quiz

NLP-Transformers Last Quiz

University

10 Qs

quiz-placeholder

Similar activities

Day #01 | Quiz

Day #01 | Quiz

University

11 Qs

ITE Elective 3 Finals Quiz Reviewer

ITE Elective 3 Finals Quiz Reviewer

University

15 Qs

AI Quiz

AI Quiz

University

9 Qs

Tri2135 : Quiz 2

Tri2135 : Quiz 2

University

15 Qs

Six Pillars of Character

Six Pillars of Character

KG - University

7 Qs

OCR A Level Computer Science (H446) - Component 1.1.3 - Input, Output and Storage

OCR A Level Computer Science (H446) - Component 1.1.3 - Input, Output and Storage

9th Grade - University

13 Qs

Hadoop

Hadoop

University

11 Qs

NCAS_FDP_05_August_2023

NCAS_FDP_05_August_2023

University

8 Qs

NLP-Transformers Last  Quiz

NLP-Transformers Last Quiz

Assessment

Quiz

Computers

University

Medium

Created by

Hazem Abdelazim

Used 12+ times

FREE Resource

10 questions

Show all answers

1.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the main purpose of multi-head self-attention in the Transformer model?

To ignore certain aspects of the input

To process inputs sequentially

To process inputs in parallel

To learn multiple contextual relationships at once

2.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the dimension of each weight matrix in the multi-head self-attention layer if the embedding dimension is 512 and there are 8 heads?

512 x 64

512 x 128

256 x 128

256 x 64

3.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the purpose of using sinusoids in the original Transformer model?

To process inputs sequentially

To reduce the dimension of the input

To add non-linearity to the system

To inject position information back into the model

4.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the purpose of the pointwise feed forward neural network in the Transformer model?

To ignore certain aspects of the input

To process inputs sequentially

To introduce non-linearity to learn complex relationships

To reduce the dimension of the input

5.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the main training task used in the original BERT model?

Sentiment Analysis

Text Generation

Masked Language Modeling (MLM)

Next Sentence Prediction (NSP)

6.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the purpose of fine-tuning a pre-trained language model?

To adapt the model for a specific task with less data and time

To train the model from scratch

To reduce the dimension of the input

To ignore certain aspects of the input

7.

MULTIPLE CHOICE QUESTION

30 sec • 1 pt

What is the main difference between BERT-base and BERT-large models?

More layers in the decoder

Larger vocabulary size

More layers in the encoder

More blocks, more attention heads, and larger embedding dimensions

Create a free account and access millions of resources

Create resources
Host any resource
Get auto-graded reports
or continue with
Microsoft
Apple
Others
By signing up, you agree to our Terms of Service & Privacy Policy
Already have an account?