* Cantinho Satkeys

Refresh History
  • okapa:
    24 de Dezembro de 2025, 19:01
  • sacana10: A todos um feliz natal
    24 de Dezembro de 2025, 17:57
  • cereal killa: dgtgtr passei por ca para vos desejar feliz natal e familias  :smiles_natal:
    24 de Dezembro de 2025, 15:46
  • bruno mirandela: deso a todos um feliz natal
    24 de Dezembro de 2025, 14:31
  • FELISCUNHA: ghyt74   :34rbzg9:  e bom natal  :13arvoresnatalmagiagifs:
    24 de Dezembro de 2025, 10:15
  • tgh12: mikrotik
    24 de Dezembro de 2025, 07:49
  • tgh12: Spanish for Beginners: Spanish from 0 to Conversational
    24 de Dezembro de 2025, 04:57
  • JPratas: try65hytr Pessoal  4tj97u<z
    24 de Dezembro de 2025, 03:03
  • m1957: Para toda a equipa e membros deste fórum, desejo um Natal feliz e que o novo ano de 2026, seja muito próspero a todos os níveis.
    24 de Dezembro de 2025, 00:47
  • FELISCUNHA: Bom dia pessoal   :34rbzg9:
    22 de Dezembro de 2025, 10:35
  • j.s.: :13arvoresnatalmagiagifs:
    21 de Dezembro de 2025, 19:01
  • j.s.: try65hytr a todos  :smiles_natal: :smiles_natal:
    21 de Dezembro de 2025, 19:01
  • FELISCUNHA: ghyt74  49E09B4F  e bom fim de semana  4tj97u<z
    20 de Dezembro de 2025, 11:20
  • JPratas: try65hytr Pessoal  2dgh8i k7y8j0 classic dgf64y
    19 de Dezembro de 2025, 05:26
  • cereal killa: ghyt74 e boa semana de chuva e frio  RGG45wj erfb57j
    15 de Dezembro de 2025, 11:26
  • FELISCUNHA: Votos de um santo domingo para todo o auditório  4tj97u<z
    14 de Dezembro de 2025, 09:28
  • j.s.: tenham um excelente fim de semana com muitas comprinhas  :13arvoresnatalmagiagifs: sdfgsdg
    13 de Dezembro de 2025, 14:58
  • j.s.: dgtgtr a todos  :smiles_natal:
    13 de Dezembro de 2025, 14:57
  • FELISCUNHA: dgtgtr   49E09B4F  e bom fim de semana   :34rbzg9:
    13 de Dezembro de 2025, 12:29
  • JPratas: try65hytr Pessoal  4tj97u<z 2dgh8i classic bve567o+
    12 de Dezembro de 2025, 05:34

Autor Tópico: Deep Learning for NLP - Part 3  (Lida 130 vezes)

0 Membros e 1 Visitante estão a ver este tópico.

Online mitsumi

  • Sub-Administrador
  • ****
  • Mensagens: 129146
  • Karma: +0/-0
Deep Learning for NLP - Part 3
« em: 13 de Agosto de 2021, 14:29 »
MP4 | Video: h264, 1280x720 | Audio: AAC, 44.1 KHz
Language: English | Size: 1.58 GB | Duration: 3h 26m

What you'll learn
Deep Learning for Natural Language Processing
Sentence Embeddings: Bag of words, Doc2Vec, SkipThought, InferSent, DSSM, USE, MTDNN, SentenceBERT
Generative Transformer Models: UniLM, Transformer-XL and XLNet, MASS, BART, CTRL, T5, ProphetNet
DL for NLP
Requirements
Basics of machine learning
Recurrent Models: RNNs, LSTMs, GRUs and variants
Basic understanding of Transformer based models and word embeddings
Description
This course is a part of "Deep Learning for NLP" Series. In this course, I will introduce concepts like Sentence embeddings and Generative Transformer Models. These concepts form the base for good understanding of advanced deep learning models for modern Natural Language Generation.

The course consists of two main sections as follows.

In the first section, I will talk about sentence embeddings. We will start with basic bag of words methods where sentence embedddings are obtained using an aggregation over word embeddings of constituent words. We will talk about averaged bag of words, word mover's distance, SIF and Power means method. Then we will discuss two unsupervised methods: Doc2Vec and SkipThought. Further, we will discuss about supervised sentence embedding methods like recursive neural networks, deep averaging networks and InferSent. CNNs can also be used for computing semantic similarity between two text strings; we will talk about DSSMs for the same. We will also discuss 3 multi-task learning methods including Universal Sentence Encodings and MT-DNN. Lastly, I will talk about SentenceBERT.

In the second section, I will talk about multiple Generative Transformer Models. We will start with UniLM. Then we will talk about segment recurrence and relative position embeddings in Transformer-XL. Then get to XLNets which use Transformer-XL along with permutation language modeling. Next we will understand span masking in MASS and also discuss various noising methods on BART. We will then discuss about controlled natural language generation using CTRL. We will discuss how T5 models every learning task as a text-to-text task. Finally, we will discuss how ProphetNet extends 2-stream attention modeling from XLNet to n-stream attention modeling, thereby enabling n-gram predictions.

Who this course is for:
Beginners in deep learning
Python developers interested in data science concepts
Masters or PhD students who wish to learn deep learning concepts quickly

Screenshots


Download link:
Só visivel para registados e com resposta ao tópico.

Only visible to registered and with a reply to the topic.

Links are Interchangeable - No Password - Single Extraction