Deep learning with PyTorch :: a practical approach to building neural network models using PyTorch /
Build neural network models in text, vision and advanced analytics using PyTorch About This Book Learn PyTorch for implementing cutting-edge deep learning algorithms. Train your neural networks for higher speed and flexibility and learn how to implement them in various scenarios; Cover various advan...
Gespeichert in:
1. Verfasser: | |
---|---|
Format: | Elektronisch E-Book |
Sprache: | English |
Veröffentlicht: |
Birmingham, UK :
Packt Publishing,
2018.
|
Schlagworte: | |
Online-Zugang: | Volltext |
Zusammenfassung: | Build neural network models in text, vision and advanced analytics using PyTorch About This Book Learn PyTorch for implementing cutting-edge deep learning algorithms. Train your neural networks for higher speed and flexibility and learn how to implement them in various scenarios; Cover various advanced neural network architecture such as ResNet, Inception, DenseNet and more with practical examples; Who This Book Is For This book is for machine learning engineers, data analysts, data scientists interested in deep learning and are looking to explore implementing advanced algorithms in PyTorch. Some knowledge of machine learning is helpful but not a mandatory need. Working knowledge of Python programming is expected. What You Will Learn Use PyTorch for GPU-accelerated tensor computations Build custom datasets and data loaders for images and test the models using torchvision and torchtext Build an image classifier by implementing CNN architectures using PyTorch Build systems that do text classification and language modeling using RNN, LSTM, and GRU Learn advanced CNN architectures such as ResNet, Inception, Densenet, and learn how to use them for transfer learning Learn how to mix multiple models for a powerful ensemble model Generate new images using GAN's and generate artistic images using style transfer In Detail Deep learning powers the most intelligent systems in the world, such as Google Voice, Siri, and Alexa. Advancements in powerful hardware, such as GPUs, software frameworks such as PyTorch, Keras, Tensorflow, and CNTK along with the availability of big data have made it easier to implement solutions to problems in the areas of text, vision, and advanced analytics. This book will get you up and running with one of the most cutting-edge deep learning libraries--PyTorch. PyTorch is grabbing the attention of deep learning researchers and data science professionals due to its accessibility, efficiency and being more native to Python way of development. You'll start off by installing PyTorch, then quickly move on to learn various fundamental blocks that power modern deep learning. You will also learn how to use CNN, RNN, LSTM and other networks to solve real-world problems. This book explains the concepts of various state-of-the-art deep learning architectures, such as ResNet, DenseNet, Inception, and Seq2Seq, without diving deep into the math behind them. You will also learn about GPU computing during the course of the book. You will see how ... |
Beschreibung: | 1 online resource (1 volume) : illustrations |
ISBN: | 9781788626071 1788626079 1788624335 9781788624336 |
Internformat
MARC
LEADER | 00000cam a2200000 i 4500 | ||
---|---|---|---|
001 | ZDB-4-EBA-on1028641678 | ||
003 | OCoLC | ||
005 | 20241004212047.0 | ||
006 | m o d | ||
007 | cr unu|||||||| | ||
008 | 180314s2018 enka o 000 0 eng d | ||
040 | |a UMI |b eng |e rda |e pn |c UMI |d STF |d NLE |d TOH |d OCLCF |d YDX |d OCLCQ |d N$T |d CEF |d KSU |d DEBBG |d TEFOD |d G3B |d S9I |d ZCU |d UAB |d OCLCQ |d OCLCO |d NZAUC |d OCLCQ |d YDX |d OCLCO |d OCLCL |d DXU | ||
019 | |a 1028619675 |a 1028670181 | ||
020 | |a 9781788626071 |q electronic book | ||
020 | |a 1788626079 |q electronic book | ||
020 | |a 1788624335 | ||
020 | |a 9781788624336 | ||
020 | |z 9781788624336 | ||
035 | |a (OCoLC)1028641678 |z (OCoLC)1028619675 |z (OCoLC)1028670181 | ||
037 | |a CL0500000947 |b Safari Books Online | ||
037 | |a F5A8F7D3-52AB-42CC-B809-28F1DA7F100E |b OverDrive, Inc. |n http://www.overdrive.com | ||
050 | 4 | |a QA76.87 |b .S83 2018 | |
072 | 7 | |a COM |x 000000 |2 bisacsh | |
082 | 7 | |a 006.32 |2 23/eng/20230522 | |
049 | |a MAIN | ||
100 | 1 | |a Subramanian, Vishnu, |e author. | |
245 | 1 | 0 | |a Deep learning with PyTorch : |b a practical approach to building neural network models using PyTorch / |c Vishnu Subramanian. |
264 | 1 | |a Birmingham, UK : |b Packt Publishing, |c 2018. | |
300 | |a 1 online resource (1 volume) : |b illustrations | ||
336 | |a text |b txt |2 rdacontent | ||
337 | |a computer |b c |2 rdamedia | ||
338 | |a online resource |b cr |2 rdacarrier | ||
347 | |a data file | ||
588 | 0 | |a Online resource; title from title page (Safari, viewed March 13, 2018). | |
505 | 0 | |a Cover -- Copyright and Credits -- Dedication -- Packt Upsell -- Foreword -- Contributors -- Table of Contents -- Preface -- Chapter 1: Getting Started with Deep Learning Using PyTorch -- Artificial intelligence -- The history of AI -- Machine learning -- Examples of machine learning in real life -- Deep learning -- Applications of deep learning -- Hype associated with deep learning -- The history of deep learning -- Why now? -- Hardware availability -- Data and algorithms -- Deep learning frameworks -- PyTorch -- Summary -- Chapter 2: Building Blocks of Neural Networks -- Installing PyTorch -- Our first neural network -- Data preparation -- Scalar (0-D tensors) -- Vectors (1-D tensors) -- Matrix (2-D tensors) -- 3-D tensors -- Slicing tensors -- 4-D tensors -- 5-D tensors -- Tensors on GPU -- Variables -- Creating data for our neural network -- Creating learnable parameters -- Neural network model -- Network implementation -- Loss function -- Optimize the neural network -- Loading data -- Dataset class -- DataLoader class -- Summary -- Chapter 3: Diving Deep into Neural Networks -- Deep dive into the building blocks of neural networks -- Layers -- fundamental blocks of neural networks -- Non-linear activations -- Sigmoid -- Tanh -- ReLU -- Leaky ReLU -- PyTorch non-linear activations -- The PyTorch way of building deep learning algorithms -- Model architecture for different machine learning problems -- Loss functions -- Optimizing network architecture -- Image classification using deep learning -- Loading data into PyTorch tensors -- Loading PyTorch tensors as batches -- Building the network architecture -- Training the model -- Summary -- Chapter 4: Fundamentals of Machine Learning -- Three kinds of machine learning problems -- Supervised learning -- Unsupervised learning -- Reinforcement learning -- Machine learning glossary. | |
505 | 8 | |a Evaluating machine learning models -- Training, validation, and test split -- Simple holdout validation -- K-fold validation -- K-fold validation with shuffling -- Data representativeness -- Time sensitivity -- Data redundancy -- Data preprocessing and feature engineering -- Vectorization -- Value normalization -- Handling missing values -- Feature engineering -- Overfitting and underfitting -- Getting more data -- Reducing the size of the network -- Applying weight regularization -- Dropout -- Underfitting -- Workflow of a machine learning project -- Problem definition and dataset creation -- Measure of success -- Evaluation protocol -- Prepare your data -- Baseline model -- Large model enough to overfit -- Applying regularization -- Learning rate picking strategies -- Summary -- Chapter 5: Deep Learning for Computer Vision -- Introduction to neural networks -- MNIST -- getting data -- Building a CNN model from scratch -- Conv2d -- Pooling -- Nonlinear activation -- ReLU -- View -- Linear layer -- Training the model -- Classifying dogs and cats -- CNN from scratch -- Classifying dogs and cats using transfer learning -- Creating and exploring a VGG16 model -- Freezing the layers -- Fine-tuning VGG16 -- Training the VGG16 model -- Calculating pre-convoluted features -- Understanding what a CNN model learns -- Visualizing outputs from intermediate layers -- Visualizing weights of the CNN layer -- Summary -- Chapter 6: Deep Learning with Sequence Data and Text -- Working with text data -- Tokenization -- Converting text into characters -- Converting text into words -- N-gram representation -- Vectorization -- One-hot encoding -- Word embedding -- Training word embedding by building a sentiment classifier -- Downloading IMDB data and performing text tokenization -- torchtext.data -- torchtext.datasets -- Building vocabulary -- Generate batches of vectors. | |
505 | 8 | |a Creating a network model with embedding -- Training the model -- Using pretrained word embeddings -- Downloading the embeddings -- Loading the embeddings in the model -- Freeze the embedding layer weights -- Recursive neural networks -- Understanding how RNN works with an example -- LSTM -- Long-term dependency -- LSTM networks -- Preparing the data -- Creating batches -- Creating the network -- Training the model -- Convolutional network on sequence data -- Understanding one-dimensional convolution for sequence data -- Creating the network -- Training the model -- Summary -- Chapter 7: Generative Networks -- Neural style transfer -- Loading the data -- Creating the VGG model -- Content loss -- Style loss -- Extracting the losses -- Creating loss function for each layers -- Creating the optimizer -- Training -- Generative adversarial networks -- Deep convolutional GAN -- Defining the generator network -- Transposed convolutions -- Batch normalization -- Generator -- Defining the discriminator network -- Defining loss and optimizer -- Training the discriminator -- Training the discriminator with real images -- Training the discriminator with fake images -- Training the generator network -- Training the complete network -- Inspecting the generated images -- Language modeling -- Preparing the data -- Generating the batches -- Batches -- Backpropagation through time -- Defining a model based on LSTM -- Defining the train and evaluate functions -- Training the model -- Summary -- Chapter 8: Modern Network Architectures -- Modern network architectures -- ResNet -- Creating PyTorch datasets -- Creating loaders for training and validation -- Creating a ResNet model -- Extracting convolutional features -- Creating a custom PyTorch dataset class for the pre-convoluted features and loader -- Creating a simple linear model -- Training and validating the model. | |
505 | 8 | |a Inception -- Creating an Inception model -- Extracting convolutional features using register_forward_hook -- Creating a new dataset for the convoluted features -- Creating a fully connected model -- Training and validating the model -- Densely connected convolutional networks -- DenseNet -- DenseBlock -- DenseLayer -- Creating a DenseNet model -- Extracting DenseNet features -- Creating a dataset and loaders -- Creating a fully connected model and train -- Model ensembling -- Creating models -- Extracting the image features -- Creating a custom dataset along with data loaders -- Creating an ensembling model -- Training and validating the model -- Encoder-decoder architecture -- Encoder -- Decoder -- Summary -- Chapter 9: What Next? -- What next? -- Overview -- Interesting ideas to explore -- Object detection -- Image segmentation -- OpenNMT in PyTorch -- Alien NLP -- fast.ai -- making neural nets uncool again -- Open Neural Network Exchange -- How to keep yourself updated -- Summary -- Other Books You May Enjoy -- Index. | |
520 | |a Build neural network models in text, vision and advanced analytics using PyTorch About This Book Learn PyTorch for implementing cutting-edge deep learning algorithms. Train your neural networks for higher speed and flexibility and learn how to implement them in various scenarios; Cover various advanced neural network architecture such as ResNet, Inception, DenseNet and more with practical examples; Who This Book Is For This book is for machine learning engineers, data analysts, data scientists interested in deep learning and are looking to explore implementing advanced algorithms in PyTorch. Some knowledge of machine learning is helpful but not a mandatory need. Working knowledge of Python programming is expected. What You Will Learn Use PyTorch for GPU-accelerated tensor computations Build custom datasets and data loaders for images and test the models using torchvision and torchtext Build an image classifier by implementing CNN architectures using PyTorch Build systems that do text classification and language modeling using RNN, LSTM, and GRU Learn advanced CNN architectures such as ResNet, Inception, Densenet, and learn how to use them for transfer learning Learn how to mix multiple models for a powerful ensemble model Generate new images using GAN's and generate artistic images using style transfer In Detail Deep learning powers the most intelligent systems in the world, such as Google Voice, Siri, and Alexa. Advancements in powerful hardware, such as GPUs, software frameworks such as PyTorch, Keras, Tensorflow, and CNTK along with the availability of big data have made it easier to implement solutions to problems in the areas of text, vision, and advanced analytics. This book will get you up and running with one of the most cutting-edge deep learning libraries--PyTorch. PyTorch is grabbing the attention of deep learning researchers and data science professionals due to its accessibility, efficiency and being more native to Python way of development. You'll start off by installing PyTorch, then quickly move on to learn various fundamental blocks that power modern deep learning. You will also learn how to use CNN, RNN, LSTM and other networks to solve real-world problems. This book explains the concepts of various state-of-the-art deep learning architectures, such as ResNet, DenseNet, Inception, and Seq2Seq, without diving deep into the math behind them. You will also learn about GPU computing during the course of the book. You will see how ... | ||
650 | 0 | |a Neural networks (Computer science) |0 http://id.loc.gov/authorities/subjects/sh90001937 | |
650 | 0 | |a Machine learning. |0 http://id.loc.gov/authorities/subjects/sh85079324 | |
650 | 0 | |a Python (Computer program language) |0 http://id.loc.gov/authorities/subjects/sh96008834 | |
650 | 0 | |a Artificial intelligence. |0 http://id.loc.gov/authorities/subjects/sh85008180 | |
650 | 2 | |a Neural Networks, Computer |0 https://id.nlm.nih.gov/mesh/D016571 | |
650 | 2 | |a Artificial Intelligence |0 https://id.nlm.nih.gov/mesh/D001185 | |
650 | 2 | |a Machine Learning |0 https://id.nlm.nih.gov/mesh/D000069550 | |
650 | 6 | |a Réseaux neuronaux (Informatique) | |
650 | 6 | |a Apprentissage automatique. | |
650 | 6 | |a Python (Langage de programmation) | |
650 | 6 | |a Intelligence artificielle. | |
650 | 7 | |a artificial intelligence. |2 aat | |
650 | 7 | |a Artificial intelligence. |2 bicssc | |
650 | 7 | |a Neural networks & fuzzy systems. |2 bicssc | |
650 | 7 | |a Data capture & analysis. |2 bicssc | |
650 | 7 | |a Computers |x Intelligence (AI) & Semantics. |2 bisacsh | |
650 | 7 | |a Computers |x Neural Networks. |2 bisacsh | |
650 | 7 | |a Computers |x Data Processing. |2 bisacsh | |
650 | 7 | |a COMPUTERS |x General. |2 bisacsh | |
650 | 7 | |a Artificial intelligence |2 fast | |
650 | 7 | |a Machine learning |2 fast | |
650 | 7 | |a Neural networks (Computer science) |2 fast | |
650 | 7 | |a Python (Computer program language) |2 fast | |
758 | |i has work: |a Deep learning with PyTorch (Text) |1 https://id.oclc.org/worldcat/entity/E39PCFxfWFG3CMcXGDwjxVVBRq |4 https://id.oclc.org/worldcat/ontology/hasWork | ||
856 | 4 | 0 | |l FWS01 |p ZDB-4-EBA |q FWS_PDA_EBA |u https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1728031 |3 Volltext |
938 | |a EBSCOhost |b EBSC |n 1728031 | ||
938 | |a YBP Library Services |b YANK |n 15211188 | ||
994 | |a 92 |b GEBAY | ||
912 | |a ZDB-4-EBA | ||
049 | |a DE-863 |
Datensatz im Suchindex
DE-BY-FWS_katkey | ZDB-4-EBA-on1028641678 |
---|---|
_version_ | 1816882415730163712 |
adam_text | |
any_adam_object | |
author | Subramanian, Vishnu |
author_facet | Subramanian, Vishnu |
author_role | aut |
author_sort | Subramanian, Vishnu |
author_variant | v s vs |
building | Verbundindex |
bvnumber | localFWS |
callnumber-first | Q - Science |
callnumber-label | QA76 |
callnumber-raw | QA76.87 .S83 2018 |
callnumber-search | QA76.87 .S83 2018 |
callnumber-sort | QA 276.87 S83 42018 |
callnumber-subject | QA - Mathematics |
collection | ZDB-4-EBA |
contents | Cover -- Copyright and Credits -- Dedication -- Packt Upsell -- Foreword -- Contributors -- Table of Contents -- Preface -- Chapter 1: Getting Started with Deep Learning Using PyTorch -- Artificial intelligence -- The history of AI -- Machine learning -- Examples of machine learning in real life -- Deep learning -- Applications of deep learning -- Hype associated with deep learning -- The history of deep learning -- Why now? -- Hardware availability -- Data and algorithms -- Deep learning frameworks -- PyTorch -- Summary -- Chapter 2: Building Blocks of Neural Networks -- Installing PyTorch -- Our first neural network -- Data preparation -- Scalar (0-D tensors) -- Vectors (1-D tensors) -- Matrix (2-D tensors) -- 3-D tensors -- Slicing tensors -- 4-D tensors -- 5-D tensors -- Tensors on GPU -- Variables -- Creating data for our neural network -- Creating learnable parameters -- Neural network model -- Network implementation -- Loss function -- Optimize the neural network -- Loading data -- Dataset class -- DataLoader class -- Summary -- Chapter 3: Diving Deep into Neural Networks -- Deep dive into the building blocks of neural networks -- Layers -- fundamental blocks of neural networks -- Non-linear activations -- Sigmoid -- Tanh -- ReLU -- Leaky ReLU -- PyTorch non-linear activations -- The PyTorch way of building deep learning algorithms -- Model architecture for different machine learning problems -- Loss functions -- Optimizing network architecture -- Image classification using deep learning -- Loading data into PyTorch tensors -- Loading PyTorch tensors as batches -- Building the network architecture -- Training the model -- Summary -- Chapter 4: Fundamentals of Machine Learning -- Three kinds of machine learning problems -- Supervised learning -- Unsupervised learning -- Reinforcement learning -- Machine learning glossary. Evaluating machine learning models -- Training, validation, and test split -- Simple holdout validation -- K-fold validation -- K-fold validation with shuffling -- Data representativeness -- Time sensitivity -- Data redundancy -- Data preprocessing and feature engineering -- Vectorization -- Value normalization -- Handling missing values -- Feature engineering -- Overfitting and underfitting -- Getting more data -- Reducing the size of the network -- Applying weight regularization -- Dropout -- Underfitting -- Workflow of a machine learning project -- Problem definition and dataset creation -- Measure of success -- Evaluation protocol -- Prepare your data -- Baseline model -- Large model enough to overfit -- Applying regularization -- Learning rate picking strategies -- Summary -- Chapter 5: Deep Learning for Computer Vision -- Introduction to neural networks -- MNIST -- getting data -- Building a CNN model from scratch -- Conv2d -- Pooling -- Nonlinear activation -- ReLU -- View -- Linear layer -- Training the model -- Classifying dogs and cats -- CNN from scratch -- Classifying dogs and cats using transfer learning -- Creating and exploring a VGG16 model -- Freezing the layers -- Fine-tuning VGG16 -- Training the VGG16 model -- Calculating pre-convoluted features -- Understanding what a CNN model learns -- Visualizing outputs from intermediate layers -- Visualizing weights of the CNN layer -- Summary -- Chapter 6: Deep Learning with Sequence Data and Text -- Working with text data -- Tokenization -- Converting text into characters -- Converting text into words -- N-gram representation -- Vectorization -- One-hot encoding -- Word embedding -- Training word embedding by building a sentiment classifier -- Downloading IMDB data and performing text tokenization -- torchtext.data -- torchtext.datasets -- Building vocabulary -- Generate batches of vectors. Creating a network model with embedding -- Training the model -- Using pretrained word embeddings -- Downloading the embeddings -- Loading the embeddings in the model -- Freeze the embedding layer weights -- Recursive neural networks -- Understanding how RNN works with an example -- LSTM -- Long-term dependency -- LSTM networks -- Preparing the data -- Creating batches -- Creating the network -- Training the model -- Convolutional network on sequence data -- Understanding one-dimensional convolution for sequence data -- Creating the network -- Training the model -- Summary -- Chapter 7: Generative Networks -- Neural style transfer -- Loading the data -- Creating the VGG model -- Content loss -- Style loss -- Extracting the losses -- Creating loss function for each layers -- Creating the optimizer -- Training -- Generative adversarial networks -- Deep convolutional GAN -- Defining the generator network -- Transposed convolutions -- Batch normalization -- Generator -- Defining the discriminator network -- Defining loss and optimizer -- Training the discriminator -- Training the discriminator with real images -- Training the discriminator with fake images -- Training the generator network -- Training the complete network -- Inspecting the generated images -- Language modeling -- Preparing the data -- Generating the batches -- Batches -- Backpropagation through time -- Defining a model based on LSTM -- Defining the train and evaluate functions -- Training the model -- Summary -- Chapter 8: Modern Network Architectures -- Modern network architectures -- ResNet -- Creating PyTorch datasets -- Creating loaders for training and validation -- Creating a ResNet model -- Extracting convolutional features -- Creating a custom PyTorch dataset class for the pre-convoluted features and loader -- Creating a simple linear model -- Training and validating the model. Inception -- Creating an Inception model -- Extracting convolutional features using register_forward_hook -- Creating a new dataset for the convoluted features -- Creating a fully connected model -- Training and validating the model -- Densely connected convolutional networks -- DenseNet -- DenseBlock -- DenseLayer -- Creating a DenseNet model -- Extracting DenseNet features -- Creating a dataset and loaders -- Creating a fully connected model and train -- Model ensembling -- Creating models -- Extracting the image features -- Creating a custom dataset along with data loaders -- Creating an ensembling model -- Training and validating the model -- Encoder-decoder architecture -- Encoder -- Decoder -- Summary -- Chapter 9: What Next? -- What next? -- Overview -- Interesting ideas to explore -- Object detection -- Image segmentation -- OpenNMT in PyTorch -- Alien NLP -- fast.ai -- making neural nets uncool again -- Open Neural Network Exchange -- How to keep yourself updated -- Summary -- Other Books You May Enjoy -- Index. |
ctrlnum | (OCoLC)1028641678 |
dewey-full | 006.32 |
dewey-hundreds | 000 - Computer science, information, general works |
dewey-ones | 006 - Special computer methods |
dewey-raw | 006.32 |
dewey-search | 006.32 |
dewey-sort | 16.32 |
dewey-tens | 000 - Computer science, information, general works |
discipline | Informatik |
format | Electronic eBook |
fullrecord | <?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>12534cam a2200769 i 4500</leader><controlfield tag="001">ZDB-4-EBA-on1028641678</controlfield><controlfield tag="003">OCoLC</controlfield><controlfield tag="005">20241004212047.0</controlfield><controlfield tag="006">m o d </controlfield><controlfield tag="007">cr unu||||||||</controlfield><controlfield tag="008">180314s2018 enka o 000 0 eng d</controlfield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">UMI</subfield><subfield code="b">eng</subfield><subfield code="e">rda</subfield><subfield code="e">pn</subfield><subfield code="c">UMI</subfield><subfield code="d">STF</subfield><subfield code="d">NLE</subfield><subfield code="d">TOH</subfield><subfield code="d">OCLCF</subfield><subfield code="d">YDX</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">N$T</subfield><subfield code="d">CEF</subfield><subfield code="d">KSU</subfield><subfield code="d">DEBBG</subfield><subfield code="d">TEFOD</subfield><subfield code="d">G3B</subfield><subfield code="d">S9I</subfield><subfield code="d">ZCU</subfield><subfield code="d">UAB</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">OCLCO</subfield><subfield code="d">NZAUC</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">YDX</subfield><subfield code="d">OCLCO</subfield><subfield code="d">OCLCL</subfield><subfield code="d">DXU</subfield></datafield><datafield tag="019" ind1=" " ind2=" "><subfield code="a">1028619675</subfield><subfield code="a">1028670181</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">9781788626071</subfield><subfield code="q">electronic book</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">1788626079</subfield><subfield code="q">electronic book</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">1788624335</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">9781788624336</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="z">9781788624336</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(OCoLC)1028641678</subfield><subfield code="z">(OCoLC)1028619675</subfield><subfield code="z">(OCoLC)1028670181</subfield></datafield><datafield tag="037" ind1=" " ind2=" "><subfield code="a">CL0500000947</subfield><subfield code="b">Safari Books Online</subfield></datafield><datafield tag="037" ind1=" " ind2=" "><subfield code="a">F5A8F7D3-52AB-42CC-B809-28F1DA7F100E</subfield><subfield code="b">OverDrive, Inc.</subfield><subfield code="n">http://www.overdrive.com</subfield></datafield><datafield tag="050" ind1=" " ind2="4"><subfield code="a">QA76.87</subfield><subfield code="b">.S83 2018</subfield></datafield><datafield tag="072" ind1=" " ind2="7"><subfield code="a">COM</subfield><subfield code="x">000000</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="082" ind1="7" ind2=" "><subfield code="a">006.32</subfield><subfield code="2">23/eng/20230522</subfield></datafield><datafield tag="049" ind1=" " ind2=" "><subfield code="a">MAIN</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Subramanian, Vishnu,</subfield><subfield code="e">author.</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Deep learning with PyTorch :</subfield><subfield code="b">a practical approach to building neural network models using PyTorch /</subfield><subfield code="c">Vishnu Subramanian.</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="a">Birmingham, UK :</subfield><subfield code="b">Packt Publishing,</subfield><subfield code="c">2018.</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">1 online resource (1 volume) :</subfield><subfield code="b">illustrations</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">computer</subfield><subfield code="b">c</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">online resource</subfield><subfield code="b">cr</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="347" ind1=" " ind2=" "><subfield code="a">data file</subfield></datafield><datafield tag="588" ind1="0" ind2=" "><subfield code="a">Online resource; title from title page (Safari, viewed March 13, 2018).</subfield></datafield><datafield tag="505" ind1="0" ind2=" "><subfield code="a">Cover -- Copyright and Credits -- Dedication -- Packt Upsell -- Foreword -- Contributors -- Table of Contents -- Preface -- Chapter 1: Getting Started with Deep Learning Using PyTorch -- Artificial intelligence -- The history of AI -- Machine learning -- Examples of machine learning in real life -- Deep learning -- Applications of deep learning -- Hype associated with deep learning -- The history of deep learning -- Why now? -- Hardware availability -- Data and algorithms -- Deep learning frameworks -- PyTorch -- Summary -- Chapter 2: Building Blocks of Neural Networks -- Installing PyTorch -- Our first neural network -- Data preparation -- Scalar (0-D tensors) -- Vectors (1-D tensors) -- Matrix (2-D tensors) -- 3-D tensors -- Slicing tensors -- 4-D tensors -- 5-D tensors -- Tensors on GPU -- Variables -- Creating data for our neural network -- Creating learnable parameters -- Neural network model -- Network implementation -- Loss function -- Optimize the neural network -- Loading data -- Dataset class -- DataLoader class -- Summary -- Chapter 3: Diving Deep into Neural Networks -- Deep dive into the building blocks of neural networks -- Layers -- fundamental blocks of neural networks -- Non-linear activations -- Sigmoid -- Tanh -- ReLU -- Leaky ReLU -- PyTorch non-linear activations -- The PyTorch way of building deep learning algorithms -- Model architecture for different machine learning problems -- Loss functions -- Optimizing network architecture -- Image classification using deep learning -- Loading data into PyTorch tensors -- Loading PyTorch tensors as batches -- Building the network architecture -- Training the model -- Summary -- Chapter 4: Fundamentals of Machine Learning -- Three kinds of machine learning problems -- Supervised learning -- Unsupervised learning -- Reinforcement learning -- Machine learning glossary.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">Evaluating machine learning models -- Training, validation, and test split -- Simple holdout validation -- K-fold validation -- K-fold validation with shuffling -- Data representativeness -- Time sensitivity -- Data redundancy -- Data preprocessing and feature engineering -- Vectorization -- Value normalization -- Handling missing values -- Feature engineering -- Overfitting and underfitting -- Getting more data -- Reducing the size of the network -- Applying weight regularization -- Dropout -- Underfitting -- Workflow of a machine learning project -- Problem definition and dataset creation -- Measure of success -- Evaluation protocol -- Prepare your data -- Baseline model -- Large model enough to overfit -- Applying regularization -- Learning rate picking strategies -- Summary -- Chapter 5: Deep Learning for Computer Vision -- Introduction to neural networks -- MNIST -- getting data -- Building a CNN model from scratch -- Conv2d -- Pooling -- Nonlinear activation -- ReLU -- View -- Linear layer -- Training the model -- Classifying dogs and cats -- CNN from scratch -- Classifying dogs and cats using transfer learning -- Creating and exploring a VGG16 model -- Freezing the layers -- Fine-tuning VGG16 -- Training the VGG16 model -- Calculating pre-convoluted features -- Understanding what a CNN model learns -- Visualizing outputs from intermediate layers -- Visualizing weights of the CNN layer -- Summary -- Chapter 6: Deep Learning with Sequence Data and Text -- Working with text data -- Tokenization -- Converting text into characters -- Converting text into words -- N-gram representation -- Vectorization -- One-hot encoding -- Word embedding -- Training word embedding by building a sentiment classifier -- Downloading IMDB data and performing text tokenization -- torchtext.data -- torchtext.datasets -- Building vocabulary -- Generate batches of vectors.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">Creating a network model with embedding -- Training the model -- Using pretrained word embeddings -- Downloading the embeddings -- Loading the embeddings in the model -- Freeze the embedding layer weights -- Recursive neural networks -- Understanding how RNN works with an example -- LSTM -- Long-term dependency -- LSTM networks -- Preparing the data -- Creating batches -- Creating the network -- Training the model -- Convolutional network on sequence data -- Understanding one-dimensional convolution for sequence data -- Creating the network -- Training the model -- Summary -- Chapter 7: Generative Networks -- Neural style transfer -- Loading the data -- Creating the VGG model -- Content loss -- Style loss -- Extracting the losses -- Creating loss function for each layers -- Creating the optimizer -- Training -- Generative adversarial networks -- Deep convolutional GAN -- Defining the generator network -- Transposed convolutions -- Batch normalization -- Generator -- Defining the discriminator network -- Defining loss and optimizer -- Training the discriminator -- Training the discriminator with real images -- Training the discriminator with fake images -- Training the generator network -- Training the complete network -- Inspecting the generated images -- Language modeling -- Preparing the data -- Generating the batches -- Batches -- Backpropagation through time -- Defining a model based on LSTM -- Defining the train and evaluate functions -- Training the model -- Summary -- Chapter 8: Modern Network Architectures -- Modern network architectures -- ResNet -- Creating PyTorch datasets -- Creating loaders for training and validation -- Creating a ResNet model -- Extracting convolutional features -- Creating a custom PyTorch dataset class for the pre-convoluted features and loader -- Creating a simple linear model -- Training and validating the model.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">Inception -- Creating an Inception model -- Extracting convolutional features using register_forward_hook -- Creating a new dataset for the convoluted features -- Creating a fully connected model -- Training and validating the model -- Densely connected convolutional networks -- DenseNet -- DenseBlock -- DenseLayer -- Creating a DenseNet model -- Extracting DenseNet features -- Creating a dataset and loaders -- Creating a fully connected model and train -- Model ensembling -- Creating models -- Extracting the image features -- Creating a custom dataset along with data loaders -- Creating an ensembling model -- Training and validating the model -- Encoder-decoder architecture -- Encoder -- Decoder -- Summary -- Chapter 9: What Next? -- What next? -- Overview -- Interesting ideas to explore -- Object detection -- Image segmentation -- OpenNMT in PyTorch -- Alien NLP -- fast.ai -- making neural nets uncool again -- Open Neural Network Exchange -- How to keep yourself updated -- Summary -- Other Books You May Enjoy -- Index.</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">Build neural network models in text, vision and advanced analytics using PyTorch About This Book Learn PyTorch for implementing cutting-edge deep learning algorithms. Train your neural networks for higher speed and flexibility and learn how to implement them in various scenarios; Cover various advanced neural network architecture such as ResNet, Inception, DenseNet and more with practical examples; Who This Book Is For This book is for machine learning engineers, data analysts, data scientists interested in deep learning and are looking to explore implementing advanced algorithms in PyTorch. Some knowledge of machine learning is helpful but not a mandatory need. Working knowledge of Python programming is expected. What You Will Learn Use PyTorch for GPU-accelerated tensor computations Build custom datasets and data loaders for images and test the models using torchvision and torchtext Build an image classifier by implementing CNN architectures using PyTorch Build systems that do text classification and language modeling using RNN, LSTM, and GRU Learn advanced CNN architectures such as ResNet, Inception, Densenet, and learn how to use them for transfer learning Learn how to mix multiple models for a powerful ensemble model Generate new images using GAN's and generate artistic images using style transfer In Detail Deep learning powers the most intelligent systems in the world, such as Google Voice, Siri, and Alexa. Advancements in powerful hardware, such as GPUs, software frameworks such as PyTorch, Keras, Tensorflow, and CNTK along with the availability of big data have made it easier to implement solutions to problems in the areas of text, vision, and advanced analytics. This book will get you up and running with one of the most cutting-edge deep learning libraries--PyTorch. PyTorch is grabbing the attention of deep learning researchers and data science professionals due to its accessibility, efficiency and being more native to Python way of development. You'll start off by installing PyTorch, then quickly move on to learn various fundamental blocks that power modern deep learning. You will also learn how to use CNN, RNN, LSTM and other networks to solve real-world problems. This book explains the concepts of various state-of-the-art deep learning architectures, such as ResNet, DenseNet, Inception, and Seq2Seq, without diving deep into the math behind them. You will also learn about GPU computing during the course of the book. You will see how ...</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Neural networks (Computer science)</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh90001937</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Machine learning.</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh85079324</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Python (Computer program language)</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh96008834</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Artificial intelligence.</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh85008180</subfield></datafield><datafield tag="650" ind1=" " ind2="2"><subfield code="a">Neural Networks, Computer</subfield><subfield code="0">https://id.nlm.nih.gov/mesh/D016571</subfield></datafield><datafield tag="650" ind1=" " ind2="2"><subfield code="a">Artificial Intelligence</subfield><subfield code="0">https://id.nlm.nih.gov/mesh/D001185</subfield></datafield><datafield tag="650" ind1=" " ind2="2"><subfield code="a">Machine Learning</subfield><subfield code="0">https://id.nlm.nih.gov/mesh/D000069550</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Réseaux neuronaux (Informatique)</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Apprentissage automatique.</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Python (Langage de programmation)</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Intelligence artificielle.</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">artificial intelligence.</subfield><subfield code="2">aat</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Artificial intelligence.</subfield><subfield code="2">bicssc</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Neural networks & fuzzy systems.</subfield><subfield code="2">bicssc</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Data capture & analysis.</subfield><subfield code="2">bicssc</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Computers</subfield><subfield code="x">Intelligence (AI) & Semantics.</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Computers</subfield><subfield code="x">Neural Networks.</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Computers</subfield><subfield code="x">Data Processing.</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">COMPUTERS</subfield><subfield code="x">General.</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Artificial intelligence</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Machine learning</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Neural networks (Computer science)</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Python (Computer program language)</subfield><subfield code="2">fast</subfield></datafield><datafield tag="758" ind1=" " ind2=" "><subfield code="i">has work:</subfield><subfield code="a">Deep learning with PyTorch (Text)</subfield><subfield code="1">https://id.oclc.org/worldcat/entity/E39PCFxfWFG3CMcXGDwjxVVBRq</subfield><subfield code="4">https://id.oclc.org/worldcat/ontology/hasWork</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="l">FWS01</subfield><subfield code="p">ZDB-4-EBA</subfield><subfield code="q">FWS_PDA_EBA</subfield><subfield code="u">https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1728031</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="938" ind1=" " ind2=" "><subfield code="a">EBSCOhost</subfield><subfield code="b">EBSC</subfield><subfield code="n">1728031</subfield></datafield><datafield tag="938" ind1=" " ind2=" "><subfield code="a">YBP Library Services</subfield><subfield code="b">YANK</subfield><subfield code="n">15211188</subfield></datafield><datafield tag="994" ind1=" " ind2=" "><subfield code="a">92</subfield><subfield code="b">GEBAY</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">ZDB-4-EBA</subfield></datafield><datafield tag="049" ind1=" " ind2=" "><subfield code="a">DE-863</subfield></datafield></record></collection> |
id | ZDB-4-EBA-on1028641678 |
illustrated | Illustrated |
indexdate | 2024-11-27T13:28:15Z |
institution | BVB |
isbn | 9781788626071 1788626079 1788624335 9781788624336 |
language | English |
oclc_num | 1028641678 |
open_access_boolean | |
owner | MAIN DE-863 DE-BY-FWS |
owner_facet | MAIN DE-863 DE-BY-FWS |
physical | 1 online resource (1 volume) : illustrations |
psigel | ZDB-4-EBA |
publishDate | 2018 |
publishDateSearch | 2018 |
publishDateSort | 2018 |
publisher | Packt Publishing, |
record_format | marc |
spelling | Subramanian, Vishnu, author. Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / Vishnu Subramanian. Birmingham, UK : Packt Publishing, 2018. 1 online resource (1 volume) : illustrations text txt rdacontent computer c rdamedia online resource cr rdacarrier data file Online resource; title from title page (Safari, viewed March 13, 2018). Cover -- Copyright and Credits -- Dedication -- Packt Upsell -- Foreword -- Contributors -- Table of Contents -- Preface -- Chapter 1: Getting Started with Deep Learning Using PyTorch -- Artificial intelligence -- The history of AI -- Machine learning -- Examples of machine learning in real life -- Deep learning -- Applications of deep learning -- Hype associated with deep learning -- The history of deep learning -- Why now? -- Hardware availability -- Data and algorithms -- Deep learning frameworks -- PyTorch -- Summary -- Chapter 2: Building Blocks of Neural Networks -- Installing PyTorch -- Our first neural network -- Data preparation -- Scalar (0-D tensors) -- Vectors (1-D tensors) -- Matrix (2-D tensors) -- 3-D tensors -- Slicing tensors -- 4-D tensors -- 5-D tensors -- Tensors on GPU -- Variables -- Creating data for our neural network -- Creating learnable parameters -- Neural network model -- Network implementation -- Loss function -- Optimize the neural network -- Loading data -- Dataset class -- DataLoader class -- Summary -- Chapter 3: Diving Deep into Neural Networks -- Deep dive into the building blocks of neural networks -- Layers -- fundamental blocks of neural networks -- Non-linear activations -- Sigmoid -- Tanh -- ReLU -- Leaky ReLU -- PyTorch non-linear activations -- The PyTorch way of building deep learning algorithms -- Model architecture for different machine learning problems -- Loss functions -- Optimizing network architecture -- Image classification using deep learning -- Loading data into PyTorch tensors -- Loading PyTorch tensors as batches -- Building the network architecture -- Training the model -- Summary -- Chapter 4: Fundamentals of Machine Learning -- Three kinds of machine learning problems -- Supervised learning -- Unsupervised learning -- Reinforcement learning -- Machine learning glossary. Evaluating machine learning models -- Training, validation, and test split -- Simple holdout validation -- K-fold validation -- K-fold validation with shuffling -- Data representativeness -- Time sensitivity -- Data redundancy -- Data preprocessing and feature engineering -- Vectorization -- Value normalization -- Handling missing values -- Feature engineering -- Overfitting and underfitting -- Getting more data -- Reducing the size of the network -- Applying weight regularization -- Dropout -- Underfitting -- Workflow of a machine learning project -- Problem definition and dataset creation -- Measure of success -- Evaluation protocol -- Prepare your data -- Baseline model -- Large model enough to overfit -- Applying regularization -- Learning rate picking strategies -- Summary -- Chapter 5: Deep Learning for Computer Vision -- Introduction to neural networks -- MNIST -- getting data -- Building a CNN model from scratch -- Conv2d -- Pooling -- Nonlinear activation -- ReLU -- View -- Linear layer -- Training the model -- Classifying dogs and cats -- CNN from scratch -- Classifying dogs and cats using transfer learning -- Creating and exploring a VGG16 model -- Freezing the layers -- Fine-tuning VGG16 -- Training the VGG16 model -- Calculating pre-convoluted features -- Understanding what a CNN model learns -- Visualizing outputs from intermediate layers -- Visualizing weights of the CNN layer -- Summary -- Chapter 6: Deep Learning with Sequence Data and Text -- Working with text data -- Tokenization -- Converting text into characters -- Converting text into words -- N-gram representation -- Vectorization -- One-hot encoding -- Word embedding -- Training word embedding by building a sentiment classifier -- Downloading IMDB data and performing text tokenization -- torchtext.data -- torchtext.datasets -- Building vocabulary -- Generate batches of vectors. Creating a network model with embedding -- Training the model -- Using pretrained word embeddings -- Downloading the embeddings -- Loading the embeddings in the model -- Freeze the embedding layer weights -- Recursive neural networks -- Understanding how RNN works with an example -- LSTM -- Long-term dependency -- LSTM networks -- Preparing the data -- Creating batches -- Creating the network -- Training the model -- Convolutional network on sequence data -- Understanding one-dimensional convolution for sequence data -- Creating the network -- Training the model -- Summary -- Chapter 7: Generative Networks -- Neural style transfer -- Loading the data -- Creating the VGG model -- Content loss -- Style loss -- Extracting the losses -- Creating loss function for each layers -- Creating the optimizer -- Training -- Generative adversarial networks -- Deep convolutional GAN -- Defining the generator network -- Transposed convolutions -- Batch normalization -- Generator -- Defining the discriminator network -- Defining loss and optimizer -- Training the discriminator -- Training the discriminator with real images -- Training the discriminator with fake images -- Training the generator network -- Training the complete network -- Inspecting the generated images -- Language modeling -- Preparing the data -- Generating the batches -- Batches -- Backpropagation through time -- Defining a model based on LSTM -- Defining the train and evaluate functions -- Training the model -- Summary -- Chapter 8: Modern Network Architectures -- Modern network architectures -- ResNet -- Creating PyTorch datasets -- Creating loaders for training and validation -- Creating a ResNet model -- Extracting convolutional features -- Creating a custom PyTorch dataset class for the pre-convoluted features and loader -- Creating a simple linear model -- Training and validating the model. Inception -- Creating an Inception model -- Extracting convolutional features using register_forward_hook -- Creating a new dataset for the convoluted features -- Creating a fully connected model -- Training and validating the model -- Densely connected convolutional networks -- DenseNet -- DenseBlock -- DenseLayer -- Creating a DenseNet model -- Extracting DenseNet features -- Creating a dataset and loaders -- Creating a fully connected model and train -- Model ensembling -- Creating models -- Extracting the image features -- Creating a custom dataset along with data loaders -- Creating an ensembling model -- Training and validating the model -- Encoder-decoder architecture -- Encoder -- Decoder -- Summary -- Chapter 9: What Next? -- What next? -- Overview -- Interesting ideas to explore -- Object detection -- Image segmentation -- OpenNMT in PyTorch -- Alien NLP -- fast.ai -- making neural nets uncool again -- Open Neural Network Exchange -- How to keep yourself updated -- Summary -- Other Books You May Enjoy -- Index. Build neural network models in text, vision and advanced analytics using PyTorch About This Book Learn PyTorch for implementing cutting-edge deep learning algorithms. Train your neural networks for higher speed and flexibility and learn how to implement them in various scenarios; Cover various advanced neural network architecture such as ResNet, Inception, DenseNet and more with practical examples; Who This Book Is For This book is for machine learning engineers, data analysts, data scientists interested in deep learning and are looking to explore implementing advanced algorithms in PyTorch. Some knowledge of machine learning is helpful but not a mandatory need. Working knowledge of Python programming is expected. What You Will Learn Use PyTorch for GPU-accelerated tensor computations Build custom datasets and data loaders for images and test the models using torchvision and torchtext Build an image classifier by implementing CNN architectures using PyTorch Build systems that do text classification and language modeling using RNN, LSTM, and GRU Learn advanced CNN architectures such as ResNet, Inception, Densenet, and learn how to use them for transfer learning Learn how to mix multiple models for a powerful ensemble model Generate new images using GAN's and generate artistic images using style transfer In Detail Deep learning powers the most intelligent systems in the world, such as Google Voice, Siri, and Alexa. Advancements in powerful hardware, such as GPUs, software frameworks such as PyTorch, Keras, Tensorflow, and CNTK along with the availability of big data have made it easier to implement solutions to problems in the areas of text, vision, and advanced analytics. This book will get you up and running with one of the most cutting-edge deep learning libraries--PyTorch. PyTorch is grabbing the attention of deep learning researchers and data science professionals due to its accessibility, efficiency and being more native to Python way of development. You'll start off by installing PyTorch, then quickly move on to learn various fundamental blocks that power modern deep learning. You will also learn how to use CNN, RNN, LSTM and other networks to solve real-world problems. This book explains the concepts of various state-of-the-art deep learning architectures, such as ResNet, DenseNet, Inception, and Seq2Seq, without diving deep into the math behind them. You will also learn about GPU computing during the course of the book. You will see how ... Neural networks (Computer science) http://id.loc.gov/authorities/subjects/sh90001937 Machine learning. http://id.loc.gov/authorities/subjects/sh85079324 Python (Computer program language) http://id.loc.gov/authorities/subjects/sh96008834 Artificial intelligence. http://id.loc.gov/authorities/subjects/sh85008180 Neural Networks, Computer https://id.nlm.nih.gov/mesh/D016571 Artificial Intelligence https://id.nlm.nih.gov/mesh/D001185 Machine Learning https://id.nlm.nih.gov/mesh/D000069550 Réseaux neuronaux (Informatique) Apprentissage automatique. Python (Langage de programmation) Intelligence artificielle. artificial intelligence. aat Artificial intelligence. bicssc Neural networks & fuzzy systems. bicssc Data capture & analysis. bicssc Computers Intelligence (AI) & Semantics. bisacsh Computers Neural Networks. bisacsh Computers Data Processing. bisacsh COMPUTERS General. bisacsh Artificial intelligence fast Machine learning fast Neural networks (Computer science) fast Python (Computer program language) fast has work: Deep learning with PyTorch (Text) https://id.oclc.org/worldcat/entity/E39PCFxfWFG3CMcXGDwjxVVBRq https://id.oclc.org/worldcat/ontology/hasWork FWS01 ZDB-4-EBA FWS_PDA_EBA https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1728031 Volltext |
spellingShingle | Subramanian, Vishnu Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / Cover -- Copyright and Credits -- Dedication -- Packt Upsell -- Foreword -- Contributors -- Table of Contents -- Preface -- Chapter 1: Getting Started with Deep Learning Using PyTorch -- Artificial intelligence -- The history of AI -- Machine learning -- Examples of machine learning in real life -- Deep learning -- Applications of deep learning -- Hype associated with deep learning -- The history of deep learning -- Why now? -- Hardware availability -- Data and algorithms -- Deep learning frameworks -- PyTorch -- Summary -- Chapter 2: Building Blocks of Neural Networks -- Installing PyTorch -- Our first neural network -- Data preparation -- Scalar (0-D tensors) -- Vectors (1-D tensors) -- Matrix (2-D tensors) -- 3-D tensors -- Slicing tensors -- 4-D tensors -- 5-D tensors -- Tensors on GPU -- Variables -- Creating data for our neural network -- Creating learnable parameters -- Neural network model -- Network implementation -- Loss function -- Optimize the neural network -- Loading data -- Dataset class -- DataLoader class -- Summary -- Chapter 3: Diving Deep into Neural Networks -- Deep dive into the building blocks of neural networks -- Layers -- fundamental blocks of neural networks -- Non-linear activations -- Sigmoid -- Tanh -- ReLU -- Leaky ReLU -- PyTorch non-linear activations -- The PyTorch way of building deep learning algorithms -- Model architecture for different machine learning problems -- Loss functions -- Optimizing network architecture -- Image classification using deep learning -- Loading data into PyTorch tensors -- Loading PyTorch tensors as batches -- Building the network architecture -- Training the model -- Summary -- Chapter 4: Fundamentals of Machine Learning -- Three kinds of machine learning problems -- Supervised learning -- Unsupervised learning -- Reinforcement learning -- Machine learning glossary. Evaluating machine learning models -- Training, validation, and test split -- Simple holdout validation -- K-fold validation -- K-fold validation with shuffling -- Data representativeness -- Time sensitivity -- Data redundancy -- Data preprocessing and feature engineering -- Vectorization -- Value normalization -- Handling missing values -- Feature engineering -- Overfitting and underfitting -- Getting more data -- Reducing the size of the network -- Applying weight regularization -- Dropout -- Underfitting -- Workflow of a machine learning project -- Problem definition and dataset creation -- Measure of success -- Evaluation protocol -- Prepare your data -- Baseline model -- Large model enough to overfit -- Applying regularization -- Learning rate picking strategies -- Summary -- Chapter 5: Deep Learning for Computer Vision -- Introduction to neural networks -- MNIST -- getting data -- Building a CNN model from scratch -- Conv2d -- Pooling -- Nonlinear activation -- ReLU -- View -- Linear layer -- Training the model -- Classifying dogs and cats -- CNN from scratch -- Classifying dogs and cats using transfer learning -- Creating and exploring a VGG16 model -- Freezing the layers -- Fine-tuning VGG16 -- Training the VGG16 model -- Calculating pre-convoluted features -- Understanding what a CNN model learns -- Visualizing outputs from intermediate layers -- Visualizing weights of the CNN layer -- Summary -- Chapter 6: Deep Learning with Sequence Data and Text -- Working with text data -- Tokenization -- Converting text into characters -- Converting text into words -- N-gram representation -- Vectorization -- One-hot encoding -- Word embedding -- Training word embedding by building a sentiment classifier -- Downloading IMDB data and performing text tokenization -- torchtext.data -- torchtext.datasets -- Building vocabulary -- Generate batches of vectors. Creating a network model with embedding -- Training the model -- Using pretrained word embeddings -- Downloading the embeddings -- Loading the embeddings in the model -- Freeze the embedding layer weights -- Recursive neural networks -- Understanding how RNN works with an example -- LSTM -- Long-term dependency -- LSTM networks -- Preparing the data -- Creating batches -- Creating the network -- Training the model -- Convolutional network on sequence data -- Understanding one-dimensional convolution for sequence data -- Creating the network -- Training the model -- Summary -- Chapter 7: Generative Networks -- Neural style transfer -- Loading the data -- Creating the VGG model -- Content loss -- Style loss -- Extracting the losses -- Creating loss function for each layers -- Creating the optimizer -- Training -- Generative adversarial networks -- Deep convolutional GAN -- Defining the generator network -- Transposed convolutions -- Batch normalization -- Generator -- Defining the discriminator network -- Defining loss and optimizer -- Training the discriminator -- Training the discriminator with real images -- Training the discriminator with fake images -- Training the generator network -- Training the complete network -- Inspecting the generated images -- Language modeling -- Preparing the data -- Generating the batches -- Batches -- Backpropagation through time -- Defining a model based on LSTM -- Defining the train and evaluate functions -- Training the model -- Summary -- Chapter 8: Modern Network Architectures -- Modern network architectures -- ResNet -- Creating PyTorch datasets -- Creating loaders for training and validation -- Creating a ResNet model -- Extracting convolutional features -- Creating a custom PyTorch dataset class for the pre-convoluted features and loader -- Creating a simple linear model -- Training and validating the model. Inception -- Creating an Inception model -- Extracting convolutional features using register_forward_hook -- Creating a new dataset for the convoluted features -- Creating a fully connected model -- Training and validating the model -- Densely connected convolutional networks -- DenseNet -- DenseBlock -- DenseLayer -- Creating a DenseNet model -- Extracting DenseNet features -- Creating a dataset and loaders -- Creating a fully connected model and train -- Model ensembling -- Creating models -- Extracting the image features -- Creating a custom dataset along with data loaders -- Creating an ensembling model -- Training and validating the model -- Encoder-decoder architecture -- Encoder -- Decoder -- Summary -- Chapter 9: What Next? -- What next? -- Overview -- Interesting ideas to explore -- Object detection -- Image segmentation -- OpenNMT in PyTorch -- Alien NLP -- fast.ai -- making neural nets uncool again -- Open Neural Network Exchange -- How to keep yourself updated -- Summary -- Other Books You May Enjoy -- Index. Neural networks (Computer science) http://id.loc.gov/authorities/subjects/sh90001937 Machine learning. http://id.loc.gov/authorities/subjects/sh85079324 Python (Computer program language) http://id.loc.gov/authorities/subjects/sh96008834 Artificial intelligence. http://id.loc.gov/authorities/subjects/sh85008180 Neural Networks, Computer https://id.nlm.nih.gov/mesh/D016571 Artificial Intelligence https://id.nlm.nih.gov/mesh/D001185 Machine Learning https://id.nlm.nih.gov/mesh/D000069550 Réseaux neuronaux (Informatique) Apprentissage automatique. Python (Langage de programmation) Intelligence artificielle. artificial intelligence. aat Artificial intelligence. bicssc Neural networks & fuzzy systems. bicssc Data capture & analysis. bicssc Computers Intelligence (AI) & Semantics. bisacsh Computers Neural Networks. bisacsh Computers Data Processing. bisacsh COMPUTERS General. bisacsh Artificial intelligence fast Machine learning fast Neural networks (Computer science) fast Python (Computer program language) fast |
subject_GND | http://id.loc.gov/authorities/subjects/sh90001937 http://id.loc.gov/authorities/subjects/sh85079324 http://id.loc.gov/authorities/subjects/sh96008834 http://id.loc.gov/authorities/subjects/sh85008180 https://id.nlm.nih.gov/mesh/D016571 https://id.nlm.nih.gov/mesh/D001185 https://id.nlm.nih.gov/mesh/D000069550 |
title | Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / |
title_auth | Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / |
title_exact_search | Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / |
title_full | Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / Vishnu Subramanian. |
title_fullStr | Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / Vishnu Subramanian. |
title_full_unstemmed | Deep learning with PyTorch : a practical approach to building neural network models using PyTorch / Vishnu Subramanian. |
title_short | Deep learning with PyTorch : |
title_sort | deep learning with pytorch a practical approach to building neural network models using pytorch |
title_sub | a practical approach to building neural network models using PyTorch / |
topic | Neural networks (Computer science) http://id.loc.gov/authorities/subjects/sh90001937 Machine learning. http://id.loc.gov/authorities/subjects/sh85079324 Python (Computer program language) http://id.loc.gov/authorities/subjects/sh96008834 Artificial intelligence. http://id.loc.gov/authorities/subjects/sh85008180 Neural Networks, Computer https://id.nlm.nih.gov/mesh/D016571 Artificial Intelligence https://id.nlm.nih.gov/mesh/D001185 Machine Learning https://id.nlm.nih.gov/mesh/D000069550 Réseaux neuronaux (Informatique) Apprentissage automatique. Python (Langage de programmation) Intelligence artificielle. artificial intelligence. aat Artificial intelligence. bicssc Neural networks & fuzzy systems. bicssc Data capture & analysis. bicssc Computers Intelligence (AI) & Semantics. bisacsh Computers Neural Networks. bisacsh Computers Data Processing. bisacsh COMPUTERS General. bisacsh Artificial intelligence fast Machine learning fast Neural networks (Computer science) fast Python (Computer program language) fast |
topic_facet | Neural networks (Computer science) Machine learning. Python (Computer program language) Artificial intelligence. Neural Networks, Computer Artificial Intelligence Machine Learning Réseaux neuronaux (Informatique) Apprentissage automatique. Python (Langage de programmation) Intelligence artificielle. artificial intelligence. Neural networks & fuzzy systems. Data capture & analysis. Computers Intelligence (AI) & Semantics. Computers Neural Networks. Computers Data Processing. COMPUTERS General. Artificial intelligence Machine learning |
url | https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1728031 |
work_keys_str_mv | AT subramanianvishnu deeplearningwithpytorchapracticalapproachtobuildingneuralnetworkmodelsusingpytorch |