最新消息:请大家多多支持

Natural Language Processing: NLP With Transformers in Python

其他教程 dsgsd 164浏览 0评论

MP4 | Video: h264, 1280×720 | Audio: AAC, 44.1 KHz
Language: English | Size: 3.61 GB | Duration: 11h 27m

What you’ll learn
Industry standard NLP using transformer models
Build full-stack question-answering transformer models
Perform sentiment analysis with transformers models in PyTorch and TensorFlow
Advanced search technologies like Elasticsearch and Facebook AI Similarity Search (FAISS)
Create fine-tuned transformers models for specialized use-cases
Measure performance of language models using advanced metrics like ROUGE
Vector building techniques like BM25 or dense passage retrievers (DPR)
An overview of recent developments in NLP
Understand attention and other key components of transformers
Learn about key transformers models such as BERT
Preprocess text data for NLP
Named entity recognition (NER) using spaCy and transformers
Fine-tune language classification models

Requirements
Knowledge of Python
Experience in data science a plus
Experience in NLP a plus

Description
Transformer models are the de-facto standard in modern NLP. They have proven themselves as the most expressive, powerful models for language by a large margin, beating all major language-based benchmarks time and time again.

In this course, we cover everything you need to get started with building cutting-edge performance NLP applications using transformer models like Google AI’s BERT, or Facebook AI’s DPR.

We cover several key NLP frameworks including:

HuggingFace’s Transformers

TensorFlow 2

PyTorch

spaCy

NLTK

Flair

And learn how to apply transformers to some of the most popular NLP use-cases:

Language classification/sentiment analysis

Named entity recognition (NER)

Question and Answering

Similarity/comparative learning

Throughout each of these use-cases we work through a variety of examples to ensure that what, how, and why transformers are so important. Alongside these sections we also work through two full-size NLP projects, one for sentiment analysis of financial Reddit data, and another covering a fully-fledged open domain question-answering application.

All of this is supported by several other sections that encourage us to learn how to better design, implement, and measure the performance of our models, such as:

History of NLP and where transformers come from

Common preprocessing techniques for NLP

The theory behind transformers

How to fine-tune transformers

We cover all this and more, I look forward to seeing you in the course!

Who this course is for:
Aspiring data scientists and ML engineers interested in NLP
Practitioners looking to upgrade their skills
Developers looking to implement NLP solutions
Data scientist
Machine Learning Engineer
Python Developers


Password/解压密码www.tbtos.com

Download rapidgator
https://rapidgator.net/file/564c4de9f8ba19facb8ca0560cdc6c99/0913_21.z01.html
https://rapidgator.net/file/86c80d21993750f383528624da7a01a3/0913_21.z02.html
https://rapidgator.net/file/f18be07c2867a0284676f29cec3c49c6/0913_21.z03.html
https://rapidgator.net/file/f4910b072753d50829330d9f61488a14/0913_21.zip.html

Download nitroflare
https://nitro.download/view/71E97503D873FCB/0913_21.z01
https://nitro.download/view/AE1B5F2D3AC4F3B/0913_21.z02
https://nitro.download/view/F2B9ED9C8FDEB90/0913_21.z03
https://nitro.download/view/B59055B118F0671/0913_21.zip

资源下载此资源仅限VIP下载,请先

转载请注明:0daytown » Natural Language Processing: NLP With Transformers in Python

发表我的评论
取消评论
表情

Hi,您需要填写昵称和邮箱!

  • 昵称 (必填)
  • 邮箱 (必填)
  • 网址