* Cantinho Satkeys

Refresh History
  • j.s.: tenham um excelente fim de semana  49E09B4F
    08 de Novembro de 2025, 16:19
  • j.s.: dgtgtr a todos  4tj97u<z
    08 de Novembro de 2025, 16:18
  • FELISCUNHA: ghyt74   49E09B4F  e bom fim de semana  4tj97u<z
    07 de Novembro de 2025, 12:04
  • JPratas: try65hytr Pessoal  2dgh8i classic k7y8j0 yu7gh8
    07 de Novembro de 2025, 03:38
  • j.s.: try65hytr a todos
    06 de Novembro de 2025, 19:11
  • FELISCUNHA: Votos de um santo domingo para todo o auditório  101041
    02 de Novembro de 2025, 11:58
  • j.s.: tenham um excelente domingo  49E09B4F
    02 de Novembro de 2025, 11:27
  • j.s.: ghyt74 a todos  4tj97u<z
    02 de Novembro de 2025, 11:26
  • FELISCUNHA: ghyt74   49E09B4F  e bom fim de semana  4tj97u<z
    01 de Novembro de 2025, 11:04
  • JPratas: try65hytr Pessoal  2dgh8i classic k7y8j0 yu7gh8
    31 de Outubro de 2025, 04:19
  • j.s.: try65hytr a todos  4tj97u<z
    30 de Outubro de 2025, 18:51
  • FELISCUNHA: ghyt74  pessoal  49E09B4F
    30 de Outubro de 2025, 11:38
  • haruri: Delta
    29 de Outubro de 2025, 07:54
  • FELISCUNHA: ghyt74   49E09B4F  e bom fim de semana  4tj97u<z
    25 de Outubro de 2025, 12:03
  • JPratas: try65hytr Pessoal  2dgh8i k7y8j0 yu7gh8
    24 de Outubro de 2025, 03:28
  • FELISCUNHA: Votos de um santo domingo para todo o auditório  4tj97u<z
    19 de Outubro de 2025, 11:16
  • j.s.: tenham um excelente domingo  43e5r6 49E09B4F
    19 de Outubro de 2025, 10:32
  • j.s.: ghyt74 a todos  4tj97u<z
    19 de Outubro de 2025, 10:32
  • FELISCUNHA: dgtgtr   49E09B4F  e bom fim de semana  4tj97u<z
    17 de Outubro de 2025, 12:08
  • JPratas: try65hytr Pessoal  4tj97u<z htg6454y k7y8j0
    17 de Outubro de 2025, 03:34

Autor Tópico: Natural Language Processing From First Principles  (Lida 117 vezes)

0 Membros e 1 Visitante estão a ver este tópico.

Online mitsumi

  • Sub-Administrador
  • ****
  • Mensagens: 126356
  • Karma: +0/-0
Natural Language Processing From First Principles
« em: 25 de Junho de 2021, 09:24 »

Genre: eLearning | MP4 | Video: h264, 1280x720 | Audio: AAC, 48.0 KHz
Language: English | Size: 1.29 GB | Duration: 3h 13m
A Beginner Friendly Introduction to Deep Learning and Artificial Intelligence with Python: Word Embedding from Scratch

What you'll learn
How to code word embeddings from scratch
How to perform stochastic gradient descent
How to augment data for natural language processing
How to perform negative sampling
How to perform sub-sampling

Description
In this course motivated beginners will learn the fundamentals of natural language processing and deep learning. Students will code their own word embedding vectors from scratch, using just Numpy and a little bit of calculus. For students who don't have the required background, a crash course in the required mathematics is included. We'll cover the fundamentals of differential calculus and linear algebra in a succinct overview, so students can easily follow all mathematical derivations.

Rather than simply be presented with results, each step of the mathematical derivations is included. This is to help students foster a deeper understanding of natural language processing and artificial intelligence in general.

Far from being a course where students are simply spoon fed the instructors' interpretation, students will learn to gather information directly from the source. I will show you a repeatable and easy to remember framework to read, understand, and implement deep learning research papers. You will get insight into how the verbiage in research papers maps to real world code. This is an essential skill set for all practitioners of artificial intelligence and data science, and will help you stand out from the crowd.

Throughout the course, good coding practices will be stressed. Students will learn the fundamentals of writing pythonic and extensible code from the very beginning, so that they can easily transition into writing more complex code for production.

By the end of the course, students will be able to answer the following questions:

What is the difference between the skip-gram and continuous bag of words models?

What is distributional semantics?

How can we use vectors to teach computers about language?

How do we derive the word2vec gradients?

Why is the softmax function so slow in natural language processing?

How can we deal with small datasets for natural language processing?

How can we improve word embedding using negative sampling?

What is the best way to to deal with proper nouns in natural language processing?

What were some of the historical approaches to natural language processing?

What can word Descriptions teach us about how computers understand language?

There is zero fluff in this course. It is taught at a brisk pace, and is intended for motivated beginners who want deeper insights into natural language processing. Those that complete this course will learn how to implement research papers on there own; you'll never have to rely on Medium blog posts again.

Screenshots


Download link:
Só visivel para registados e com resposta ao tópico.

Only visible to registered and with a reply to the topic.

Links are Interchangeable - No Password - Single Extraction