0
0
TensorFlowml~5 mins

Sequence-to-sequence basics in TensorFlow - Cheat Sheet & Quick Revision

Choose your learning style9 modes available
Recall & Review
beginner
What is a sequence-to-sequence (seq2seq) model?
A seq2seq model is a type of neural network that takes a sequence as input and produces another sequence as output. It is often used for tasks like language translation or text summarization.
Click to reveal answer
beginner
What are the two main parts of a seq2seq model?
The two main parts are the encoder and the decoder. The encoder reads the input sequence and creates a summary, and the decoder uses this summary to generate the output sequence.
Click to reveal answer
intermediate
Why do seq2seq models use recurrent neural networks (RNNs) or similar layers?
Because RNNs can process sequences of varying lengths by remembering information from previous steps, making them suitable for handling input and output sequences in seq2seq models.
Click to reveal answer
intermediate
What is the role of the 'context vector' in a basic seq2seq model?
The context vector is the fixed-size summary output by the encoder that contains information about the entire input sequence. The decoder uses it to generate the output sequence.
Click to reveal answer
advanced
What is teacher forcing in seq2seq training?
Teacher forcing is a training technique where the true output from the training data is fed as the next input to the decoder instead of the decoder's own previous prediction. This helps the model learn faster.
Click to reveal answer
What does the encoder in a seq2seq model do?
AGenerates the output sequence
BProcesses the input sequence and creates a summary
CCalculates the loss function
DFeeds data into the training loop
Which layer type is commonly used in seq2seq models to handle sequences?
AConvolutional layers
BPooling layers
CDense layers only
DRecurrent layers (RNN, LSTM, GRU)
What is the purpose of the decoder in a seq2seq model?
ATo generate the output sequence from the context vector
BTo encode the input sequence
CTo preprocess the data
DTo calculate accuracy
What does teacher forcing help with during training?
ARemoves the need for a decoder
BSlows down training
CImproves model learning by using true previous outputs
DIncreases model size
In a basic seq2seq model, what is the 'context vector'?
AA fixed-size summary of the input sequence
BThe output sequence
CThe input sequence
DThe loss value
Explain the roles of the encoder and decoder in a sequence-to-sequence model.
Think about how information flows from input to output.
You got /4 concepts.
    Describe why recurrent layers like LSTM or GRU are important in seq2seq models.
    Consider how the model keeps track of previous words or steps.
    You got /4 concepts.