DAT255: Deep Learning Lecture 7 - Sequences and Time Series

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/15

flashcard set

Earn XP

Description and Tags

Flashcards based on lecture notes about sequences, time series, and recurrent neural networks.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

16 Terms

1
New cards

What are Sequences?

Ordered series of data, like natural language, time series, audio, video, or DNA.

<p>Ordered series of data, like natural language, time series, audio, video, or DNA.</p>
2
New cards

Give examples of Sequence Classification Tasks.

Speech recognition, fraud detection, medical diagnostics, sentiment analysis, topic classification.

3
New cards

Give examples of Forecasting tasks.

Predicting weather, energy prices, or stock prices.

4
New cards

Give examples of Sequence-to-sequence learning tasks

Language translation, image captioning, or text summarization.

5
New cards

How do sequence classification methods work with 1D sequences?

Instead of needing patterns between neighboring pixels in 2D, you look for patterns between neighboring elements in 1D.

6
New cards

What is cross-correlation?

The operation is similar to convolution but reverses one of the functions.

7
New cards

In forecasting, what assumption is typically made about recent data?

Recent data is more informative than old data.

8
New cards

What is a limitation of standard neural networks?

Networks without state, unable to 'remember' previous inputs.

9
New cards

How is state introduced in recurrent neural networks?

Each node stores its previous output.

10
New cards

What is the simplest possible autoregressive forecast model?

The value tomorrow is the same as the value today. yi = yt-1

11
New cards

What is a more advanced autoregressive forecast?

Accounts for the value tomorrow being a weighted sum of previous time steps plus a noise term.

12
New cards

What are two common approaches to mitigate vanishing/exploding gradients in RNNs?

LSTMs and GRUs.

13
New cards

When stacking recurrent layers in a deep RNN, what is important?

Intermediate layers should return the entire sequence.

14
New cards

What are some tricks to efficiently train RNNs?

Use saturating activation functions (tanh, sigmoid), layer normalization, and recurrent dropout.

15
New cards

What is one bonus trick for training RNNs using CNN processing?

Extract small-scale patterns with convolutional layers first, then apply recurrent layers.

16
New cards

What do bidirectional RNNs do?

Process sequences both forwards and in reverse.