RAR-files |
linkedin.learning.tensorflow.working.with.nlp-xqzt.rar |
50,000,000 |
E90CFC14 |
linkedin.learning.tensorflow.working.with.nlp-xqzt.r00 |
50,000,000 |
6E7389C4 |
linkedin.learning.tensorflow.working.with.nlp-xqzt.r01 |
18,079,197 |
0F33517E |
|
Total size: |
118,079,197 |
|
|
Archived
files |
01.01-why_tensorflow.mkv
[88876f658c3806ce]
|
2,812,135 |
81969966 |
01.02-what_you_should_know.mkv
[90e830dd2c6fe58d]
|
1,010,420 |
A958F216 |
01.03-what_is_tensorflow.mkv
[13a0861847127346]
|
5,160,643 |
03A074FD |
02.01-what_is_nlp.mkv
[cc784cd93626ecdc]
|
4,981,125 |
8CFF29E6 |
02.02-transformers_for_nlp.mkv
[572dec885d8a7fe2]
|
3,879,288 |
085D047A |
02.03-transformers_their_use_and_history.mkv
[707ce83b76fb089f]
|
12,319,331 |
DAA9C3F1 |
02.04-challenge_nlp_model_size.mkv
[3d2443fb97b91b19]
|
2,020,360 |
B0B99FFD |
02.05-solution_nlp_model_size.mkv
[7b07f67844f2f2c2]
|
10,910,057 |
36CA787F |
03.01-bias_in_bert_and_gpt.mkv
[50777b9d26c49063]
|
14,905,806 |
2B307F28 |
03.02-how_was_bert_trained.mkv
[166d004961cc9627]
|
3,660,213 |
A75ECA23 |
03.03-transfer_learning.mkv
[f84a541facd6964]
|
5,050,901 |
782EAD6A |
04.01-transformer_architecture_overview.mkv
[8833811e345574a8]
|
2,879,703 |
00F52287 |
04.02-bert_model_and_tokenization.mkv
[af5e91611292ce8]
|
6,782,766 |
FECC19B1 |
04.03-tokenizers.mkv
[a3858f3cbdcfc2ef]
|
7,935,338 |
3296F801 |
04.04-self-attention.mkv
[98afb18402b9dc30]
|
2,907,662 |
5C0390A6 |
04.05-multi-head_attention_and_feedforward_network.mkv
[a7229d75d48be986]
|
1,544,122 |
126F50DC |
04.06-fine-tuning_bert.mkv
[eca548510c22cb03]
|
27,635,954 |
5F79DCC5 |
05.01-next_steps.mkv
[8e548d3ea2413257]
|
1,681,902 |
C23BBB30 |
|
Total size: |
118,077,726 |
|
|