ODBIERZ TWÓJ BONUS :: »

Building Transformer Models with PyTorch 2.0 Prem Timsina

Język publikacji: 1
Autor:
Prem Timsina
Ocena:
Bądź pierwszym, który oceni tę książkę
Stron:
310
Dostępne formaty:
     ePub
     Mobi
Czytaj fragment

Ebook 76,49 zł najniższa cena z 30 dni

89,90 zł (-25%)
67,43 zł

Dodaj do koszyka lub Kup na prezent Kup 1-kliknięciem

76,49 zł najniższa cena z 30 dni

Poleć tę książkę znajomemu Poleć tę książkę znajomemu!!

Przenieś na półkę

Do przechowalni

Prezent last minute w ebookpoint.pl
Zostało Ci na świąteczne zamówienie opcje wysyłki »
Your key to transformer based NLP, vision, speech, and multimodalities

Key Features
Transformer architecture for different modalities and multimodalities.
Practical guidelines to build and fine-tune transformer models.
Comprehensive code samples with detailed documentation.

Description
This book covers transformer architecture for various applications including NLP, computer vision, speech processing, and predictive modeling with tabular data. It is a valuable resource for anyone looking to harness the power of transformer architecture in their machine learning projects.

The book provides a step-by-step guide to building transformer models from scratch and fine-tuning pre-trained open-source models. It explores foundational model architecture, including GPT, VIT, Whisper, TabTransformer, Stable Diffusion, and the core principles for solving various problems with transformers. The book also covers transfer learning, model training, and fine-tuning, and discusses how to utilize recent models from Hugging Face. Additionally, the book explores advanced topics such as model benchmarking, multimodal learning, reinforcement learning, and deploying and serving transformer models.

In conclusion, this book offers a comprehensive and thorough guide to transformer models and their various applications.

What you will learn
Understand the core architecture of various foundational models, including single and multimodalities.
Step-by-step approach to developing transformer-based Machine Learning models.
Utilize various open-source models to solve your business problems.
Train and fine-tune various open-source models using PyTorch 2.0 and the Hugging Face ecosystem.
Deploy and serve transformer models.
Best practices and guidelines for building transformer-based models.

Who this book is for
This book caters to data scientists, Machine Learning engineers, developers, and software architects interested in the world of generative AI.

Table of Contents
1. Transformer Architecture
2. Hugging Face Ecosystem
3. Transformer Model in PyTorch
4. Transfer Learning with PyTorch and Hugging Face
5. Large Language Models: BERT, GPT-3, and BART
6. NLP Tasks with Transformers
7. CV Model Anatomy: ViT, DETR, and DeiT
8. Computer Vision Tasks with Transformers
9. Speech Processing Model Anatomy: Whisper, SpeechT5, and Wav2Vec
10. Speech Tasks with Transformers
11. Transformer Architecture for Tabular Data Processing
12. Transformers for Tabular Data Regression and Classification
13. Multimodal Transformers, Architectures and Applications
14. Explore Reinforcement Learning for Transformer
15. Model Export, Serving, and Deployment
16. Transformer Model Interpretability, and Experimental Visualization
17. PyTorch Models: Best Practices and Debugging

Wybrane bestsellery

BPB Publications - inne książki

Zamknij

Przenieś na półkę
Dodano produkt na półkę
Usunięto produkt z półki
Przeniesiono produkt do archiwum
Przeniesiono produkt do biblioteki

Zamknij

Wybierz metodę płatności

Ebook
67,43 zł
Dodaj do koszyka
Sposób płatności
Zabrania się wykorzystania treści strony do celów eksploracji tekstu i danych (TDM), w tym eksploracji w celu szkolenia technologii AI i innych systemów uczenia maszynowego. It is forbidden to use the content of the site for text and data mining (TDM), including mining for training AI technologies and other machine learning systems.