Natural language processing with TensorFlow :: teach language to machines using Python's deep learning library /
TensorFlow is the leading framework for deep learning algorithms critical to artificial intelligence, and natural language processing (NLP) makes much of the data used by deep learning applications accessible to them. This book brings the two together and teaches deep learning developers how to work...
Gespeichert in:
1. Verfasser: | |
---|---|
Format: | Elektronisch E-Book |
Sprache: | English |
Veröffentlicht: |
Birmingham, UK :
Packt,
[2018]
|
Schlagworte: | |
Online-Zugang: | Volltext |
Zusammenfassung: | TensorFlow is the leading framework for deep learning algorithms critical to artificial intelligence, and natural language processing (NLP) makes much of the data used by deep learning applications accessible to them. This book brings the two together and teaches deep learning developers how to work with today's vast amount of unstructured data. |
Beschreibung: | Implementing subsampling. |
Beschreibung: | 1 online resource (472 pages) |
Bibliographie: | Includes bibliographical references and index. |
ISBN: | 9781788477758 1788477758 |
Internformat
MARC
LEADER | 00000cam a2200000 i 4500 | ||
---|---|---|---|
001 | ZDB-4-EBA-on1039700926 | ||
003 | OCoLC | ||
005 | 20241004212047.0 | ||
006 | m o d | ||
007 | cr cnu---unuuu | ||
008 | 180609s2018 enk ob 001 0 eng d | ||
040 | |a EBLCP |b eng |e rda |e pn |c EBLCP |d N$T |d MERUC |d OCLCF |d IDB |d NLE |d TEFOD |d OCLCQ |d UKMGB |d LVT |d UKAHL |d OCLCQ |d UX1 |d K6U |d OCLCQ |d OCLCO |d NZAUC |d OCLCQ |d OCLCO |d OCLCL |d SXB | ||
015 | |a GBB8A8385 |2 bnb | ||
016 | 7 | |a 018897090 |2 Uk | |
019 | |a 1175633515 | ||
020 | |a 9781788477758 |q (electronic bk.) | ||
020 | |a 1788477758 |q (electronic bk.) | ||
020 | |z 9781788478311 | ||
035 | |a (OCoLC)1039700926 |z (OCoLC)1175633515 | ||
037 | |a 6C52D1D4-6E23-433E-B2ED-09E5671203F8 |b OverDrive, Inc. |n http://www.overdrive.com | ||
050 | 4 | |a Q325.5 | |
072 | 7 | |a COM |x 000000 |2 bisacsh | |
082 | 7 | |a 006.31 |2 23 | |
049 | |a MAIN | ||
100 | 1 | |a Ganegedara, Thushan, |e author. | |
245 | 1 | 0 | |a Natural language processing with TensorFlow : |b teach language to machines using Python's deep learning library / |c Thushan Ganegedara. |
264 | 1 | |a Birmingham, UK : |b Packt, |c [2018] | |
264 | 4 | |c ©2018 | |
300 | |a 1 online resource (472 pages) | ||
336 | |a text |b txt |2 rdacontent | ||
337 | |a computer |b c |2 rdamedia | ||
338 | |a online resource |b cr |2 rdacarrier | ||
504 | |a Includes bibliographical references and index. | ||
588 | 0 | |a Print version record. | |
505 | 0 | |a Cover; Copyright; Packt Upsell; Contributors; Table of Contents; Preface; Chapter 1: Introduction to Natural Language Processing; What is Natural Language Processing?; Tasks of Natural Language Processing; The traditional approach to Natural Language Processing; Understanding the traditional approach; Example -- generating football game summaries; Drawbacks of the traditional approach; The deep learning approach to Natural Language Processing; History of deep learning; The current state of deep learning and NLP; Understanding a simple deep model -- a Fully Connected Neural Network. | |
505 | 8 | |a The roadmap -- beyond this chapterIntroduction to the technical tools; Description of the tools; Installing Python and scikit-learn; Installing Jupyter Notebook; Installing TensorFlow; Summary; Chapter 2: Understanding TensorFlow; What is TensorFlow?; Getting started with TensorFlow; TensorFlow client in detail; TensorFlow architecture -- what happens when you execute the client?; Cafe Le TensorFlow -- understanding TensorFlow with an analogy; Inputs, variables, outputs, and operations; Defining inputs in TensorFlow; Feeding data with Python code; Preloading and storing data as tensors. | |
505 | 8 | |a Building an input pipelineDefining variables in TensorFlow; Defining TensorFlow outputs; Defining TensorFlow operations; Comparison operations; Mathematical operations; Scatter and gather operations; Neural network-related operations; Reusing variables with scoping; Implementing our first neural network; Preparing the data; Defining the TensorFlow graph; Running the neural network; Summary; Chapter 3: Word2vec -- Learning Word Embeddings; What is a word representation or meaning?; Classical approaches to learning word representation. | |
505 | 8 | |a WordNet -- using an external lexical knowledge base for learning word representationsTour of WordNet; Problems with WordNet; One-hot encoded representation; The TF-IDF method; Co-occurrence matrix; Word2vec -- a neural network-based approach to learning word representation; Exercise: is queen = king -- he + she?; Designing a loss function for learning word embeddings; The skip-gram algorithm; From raw text to structured data; Learning the word embeddings with a neural network; Formulating a practical loss function; Efficiently approximating the loss function. | |
505 | 8 | |a Implementing skip-gram with TensorFlowThe Continuous Bag-of-Words algorithm; Implementing CBOW in TensorFlow; Summary; Chapter 4: Advanced Word2vec; The original skip-gram algorithm; Implementing the original skip-gram algorithm; Comparing the original skip-gram with the improved skip-gram; Comparing skip-gram with CBOW; Performance comparison; Which is the winner, skip-gram or CBOW?; Extensions to the word embeddings algorithms; Using the unigram distribution for negative sampling; Implementing unigram-based negative sampling; Subsampling -- probabilistically ignoring the common words. | |
500 | |a Implementing subsampling. | ||
520 | |a TensorFlow is the leading framework for deep learning algorithms critical to artificial intelligence, and natural language processing (NLP) makes much of the data used by deep learning applications accessible to them. This book brings the two together and teaches deep learning developers how to work with today's vast amount of unstructured data. | ||
650 | 0 | |a Machine learning. |0 http://id.loc.gov/authorities/subjects/sh85079324 | |
650 | 0 | |a Artificial intelligence. |0 http://id.loc.gov/authorities/subjects/sh85008180 | |
650 | 0 | |a Python (Computer program language) |0 http://id.loc.gov/authorities/subjects/sh96008834 | |
650 | 6 | |a Apprentissage automatique. | |
650 | 6 | |a Intelligence artificielle. | |
650 | 6 | |a Python (Langage de programmation) | |
650 | 7 | |a artificial intelligence. |2 aat | |
650 | 7 | |a Programming & scripting languages: general. |2 bicssc | |
650 | 7 | |a Neural networks & fuzzy systems. |2 bicssc | |
650 | 7 | |a Artificial intelligence. |2 bicssc | |
650 | 7 | |a COMPUTERS |x General. |2 bisacsh | |
650 | 7 | |a Artificial intelligence |2 fast | |
650 | 7 | |a Machine learning |2 fast | |
650 | 7 | |a Python (Computer program language) |2 fast | |
655 | 4 | |a Electronic book. | |
758 | |i has work: |a Natural language processing with TensorFlow (Text) |1 https://id.oclc.org/worldcat/entity/E39PCGPFcfyqXFQwPxgyHWGQ9C |4 https://id.oclc.org/worldcat/ontology/hasWork | ||
776 | 0 | 8 | |i Print version: |a Ganegedara, Thushan. |t Natural Language Processing with TensorFlow : Teach language to machines using Python's deep learning library. |d Birmingham : Packt Publishing, ©2018 |
856 | 4 | 0 | |l FWS01 |p ZDB-4-EBA |q FWS_PDA_EBA |u https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1823678 |3 Volltext |
938 | |a Askews and Holts Library Services |b ASKH |n BDZ0036924764 | ||
938 | |a EBL - Ebook Library |b EBLB |n EBL5405681 | ||
938 | |a EBSCOhost |b EBSC |n 1823678 | ||
994 | |a 92 |b GEBAY | ||
912 | |a ZDB-4-EBA | ||
049 | |a DE-863 |
Datensatz im Suchindex
DE-BY-FWS_katkey | ZDB-4-EBA-on1039700926 |
---|---|
_version_ | 1816882462161108992 |
adam_text | |
any_adam_object | |
author | Ganegedara, Thushan |
author_facet | Ganegedara, Thushan |
author_role | aut |
author_sort | Ganegedara, Thushan |
author_variant | t g tg |
building | Verbundindex |
bvnumber | localFWS |
callnumber-first | Q - Science |
callnumber-label | Q325 |
callnumber-raw | Q325.5 |
callnumber-search | Q325.5 |
callnumber-sort | Q 3325.5 |
callnumber-subject | Q - General Science |
collection | ZDB-4-EBA |
contents | Cover; Copyright; Packt Upsell; Contributors; Table of Contents; Preface; Chapter 1: Introduction to Natural Language Processing; What is Natural Language Processing?; Tasks of Natural Language Processing; The traditional approach to Natural Language Processing; Understanding the traditional approach; Example -- generating football game summaries; Drawbacks of the traditional approach; The deep learning approach to Natural Language Processing; History of deep learning; The current state of deep learning and NLP; Understanding a simple deep model -- a Fully Connected Neural Network. The roadmap -- beyond this chapterIntroduction to the technical tools; Description of the tools; Installing Python and scikit-learn; Installing Jupyter Notebook; Installing TensorFlow; Summary; Chapter 2: Understanding TensorFlow; What is TensorFlow?; Getting started with TensorFlow; TensorFlow client in detail; TensorFlow architecture -- what happens when you execute the client?; Cafe Le TensorFlow -- understanding TensorFlow with an analogy; Inputs, variables, outputs, and operations; Defining inputs in TensorFlow; Feeding data with Python code; Preloading and storing data as tensors. Building an input pipelineDefining variables in TensorFlow; Defining TensorFlow outputs; Defining TensorFlow operations; Comparison operations; Mathematical operations; Scatter and gather operations; Neural network-related operations; Reusing variables with scoping; Implementing our first neural network; Preparing the data; Defining the TensorFlow graph; Running the neural network; Summary; Chapter 3: Word2vec -- Learning Word Embeddings; What is a word representation or meaning?; Classical approaches to learning word representation. WordNet -- using an external lexical knowledge base for learning word representationsTour of WordNet; Problems with WordNet; One-hot encoded representation; The TF-IDF method; Co-occurrence matrix; Word2vec -- a neural network-based approach to learning word representation; Exercise: is queen = king -- he + she?; Designing a loss function for learning word embeddings; The skip-gram algorithm; From raw text to structured data; Learning the word embeddings with a neural network; Formulating a practical loss function; Efficiently approximating the loss function. Implementing skip-gram with TensorFlowThe Continuous Bag-of-Words algorithm; Implementing CBOW in TensorFlow; Summary; Chapter 4: Advanced Word2vec; The original skip-gram algorithm; Implementing the original skip-gram algorithm; Comparing the original skip-gram with the improved skip-gram; Comparing skip-gram with CBOW; Performance comparison; Which is the winner, skip-gram or CBOW?; Extensions to the word embeddings algorithms; Using the unigram distribution for negative sampling; Implementing unigram-based negative sampling; Subsampling -- probabilistically ignoring the common words. |
ctrlnum | (OCoLC)1039700926 |
dewey-full | 006.31 |
dewey-hundreds | 000 - Computer science, information, general works |
dewey-ones | 006 - Special computer methods |
dewey-raw | 006.31 |
dewey-search | 006.31 |
dewey-sort | 16.31 |
dewey-tens | 000 - Computer science, information, general works |
discipline | Informatik |
format | Electronic eBook |
fullrecord | <?xml version="1.0" encoding="UTF-8"?><collection xmlns="http://www.loc.gov/MARC21/slim"><record><leader>06293cam a2200721 i 4500</leader><controlfield tag="001">ZDB-4-EBA-on1039700926</controlfield><controlfield tag="003">OCoLC</controlfield><controlfield tag="005">20241004212047.0</controlfield><controlfield tag="006">m o d </controlfield><controlfield tag="007">cr cnu---unuuu</controlfield><controlfield tag="008">180609s2018 enk ob 001 0 eng d</controlfield><datafield tag="040" ind1=" " ind2=" "><subfield code="a">EBLCP</subfield><subfield code="b">eng</subfield><subfield code="e">rda</subfield><subfield code="e">pn</subfield><subfield code="c">EBLCP</subfield><subfield code="d">N$T</subfield><subfield code="d">MERUC</subfield><subfield code="d">OCLCF</subfield><subfield code="d">IDB</subfield><subfield code="d">NLE</subfield><subfield code="d">TEFOD</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">UKMGB</subfield><subfield code="d">LVT</subfield><subfield code="d">UKAHL</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">UX1</subfield><subfield code="d">K6U</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">OCLCO</subfield><subfield code="d">NZAUC</subfield><subfield code="d">OCLCQ</subfield><subfield code="d">OCLCO</subfield><subfield code="d">OCLCL</subfield><subfield code="d">SXB</subfield></datafield><datafield tag="015" ind1=" " ind2=" "><subfield code="a">GBB8A8385</subfield><subfield code="2">bnb</subfield></datafield><datafield tag="016" ind1="7" ind2=" "><subfield code="a">018897090</subfield><subfield code="2">Uk</subfield></datafield><datafield tag="019" ind1=" " ind2=" "><subfield code="a">1175633515</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">9781788477758</subfield><subfield code="q">(electronic bk.)</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="a">1788477758</subfield><subfield code="q">(electronic bk.)</subfield></datafield><datafield tag="020" ind1=" " ind2=" "><subfield code="z">9781788478311</subfield></datafield><datafield tag="035" ind1=" " ind2=" "><subfield code="a">(OCoLC)1039700926</subfield><subfield code="z">(OCoLC)1175633515</subfield></datafield><datafield tag="037" ind1=" " ind2=" "><subfield code="a">6C52D1D4-6E23-433E-B2ED-09E5671203F8</subfield><subfield code="b">OverDrive, Inc.</subfield><subfield code="n">http://www.overdrive.com</subfield></datafield><datafield tag="050" ind1=" " ind2="4"><subfield code="a">Q325.5</subfield></datafield><datafield tag="072" ind1=" " ind2="7"><subfield code="a">COM</subfield><subfield code="x">000000</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="082" ind1="7" ind2=" "><subfield code="a">006.31</subfield><subfield code="2">23</subfield></datafield><datafield tag="049" ind1=" " ind2=" "><subfield code="a">MAIN</subfield></datafield><datafield tag="100" ind1="1" ind2=" "><subfield code="a">Ganegedara, Thushan,</subfield><subfield code="e">author.</subfield></datafield><datafield tag="245" ind1="1" ind2="0"><subfield code="a">Natural language processing with TensorFlow :</subfield><subfield code="b">teach language to machines using Python's deep learning library /</subfield><subfield code="c">Thushan Ganegedara.</subfield></datafield><datafield tag="264" ind1=" " ind2="1"><subfield code="a">Birmingham, UK :</subfield><subfield code="b">Packt,</subfield><subfield code="c">[2018]</subfield></datafield><datafield tag="264" ind1=" " ind2="4"><subfield code="c">©2018</subfield></datafield><datafield tag="300" ind1=" " ind2=" "><subfield code="a">1 online resource (472 pages)</subfield></datafield><datafield tag="336" ind1=" " ind2=" "><subfield code="a">text</subfield><subfield code="b">txt</subfield><subfield code="2">rdacontent</subfield></datafield><datafield tag="337" ind1=" " ind2=" "><subfield code="a">computer</subfield><subfield code="b">c</subfield><subfield code="2">rdamedia</subfield></datafield><datafield tag="338" ind1=" " ind2=" "><subfield code="a">online resource</subfield><subfield code="b">cr</subfield><subfield code="2">rdacarrier</subfield></datafield><datafield tag="504" ind1=" " ind2=" "><subfield code="a">Includes bibliographical references and index.</subfield></datafield><datafield tag="588" ind1="0" ind2=" "><subfield code="a">Print version record.</subfield></datafield><datafield tag="505" ind1="0" ind2=" "><subfield code="a">Cover; Copyright; Packt Upsell; Contributors; Table of Contents; Preface; Chapter 1: Introduction to Natural Language Processing; What is Natural Language Processing?; Tasks of Natural Language Processing; The traditional approach to Natural Language Processing; Understanding the traditional approach; Example -- generating football game summaries; Drawbacks of the traditional approach; The deep learning approach to Natural Language Processing; History of deep learning; The current state of deep learning and NLP; Understanding a simple deep model -- a Fully Connected Neural Network.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">The roadmap -- beyond this chapterIntroduction to the technical tools; Description of the tools; Installing Python and scikit-learn; Installing Jupyter Notebook; Installing TensorFlow; Summary; Chapter 2: Understanding TensorFlow; What is TensorFlow?; Getting started with TensorFlow; TensorFlow client in detail; TensorFlow architecture -- what happens when you execute the client?; Cafe Le TensorFlow -- understanding TensorFlow with an analogy; Inputs, variables, outputs, and operations; Defining inputs in TensorFlow; Feeding data with Python code; Preloading and storing data as tensors.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">Building an input pipelineDefining variables in TensorFlow; Defining TensorFlow outputs; Defining TensorFlow operations; Comparison operations; Mathematical operations; Scatter and gather operations; Neural network-related operations; Reusing variables with scoping; Implementing our first neural network; Preparing the data; Defining the TensorFlow graph; Running the neural network; Summary; Chapter 3: Word2vec -- Learning Word Embeddings; What is a word representation or meaning?; Classical approaches to learning word representation.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">WordNet -- using an external lexical knowledge base for learning word representationsTour of WordNet; Problems with WordNet; One-hot encoded representation; The TF-IDF method; Co-occurrence matrix; Word2vec -- a neural network-based approach to learning word representation; Exercise: is queen = king -- he + she?; Designing a loss function for learning word embeddings; The skip-gram algorithm; From raw text to structured data; Learning the word embeddings with a neural network; Formulating a practical loss function; Efficiently approximating the loss function.</subfield></datafield><datafield tag="505" ind1="8" ind2=" "><subfield code="a">Implementing skip-gram with TensorFlowThe Continuous Bag-of-Words algorithm; Implementing CBOW in TensorFlow; Summary; Chapter 4: Advanced Word2vec; The original skip-gram algorithm; Implementing the original skip-gram algorithm; Comparing the original skip-gram with the improved skip-gram; Comparing skip-gram with CBOW; Performance comparison; Which is the winner, skip-gram or CBOW?; Extensions to the word embeddings algorithms; Using the unigram distribution for negative sampling; Implementing unigram-based negative sampling; Subsampling -- probabilistically ignoring the common words.</subfield></datafield><datafield tag="500" ind1=" " ind2=" "><subfield code="a">Implementing subsampling.</subfield></datafield><datafield tag="520" ind1=" " ind2=" "><subfield code="a">TensorFlow is the leading framework for deep learning algorithms critical to artificial intelligence, and natural language processing (NLP) makes much of the data used by deep learning applications accessible to them. This book brings the two together and teaches deep learning developers how to work with today's vast amount of unstructured data.</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Machine learning.</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh85079324</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Artificial intelligence.</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh85008180</subfield></datafield><datafield tag="650" ind1=" " ind2="0"><subfield code="a">Python (Computer program language)</subfield><subfield code="0">http://id.loc.gov/authorities/subjects/sh96008834</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Apprentissage automatique.</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Intelligence artificielle.</subfield></datafield><datafield tag="650" ind1=" " ind2="6"><subfield code="a">Python (Langage de programmation)</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">artificial intelligence.</subfield><subfield code="2">aat</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Programming & scripting languages: general.</subfield><subfield code="2">bicssc</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Neural networks & fuzzy systems.</subfield><subfield code="2">bicssc</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Artificial intelligence.</subfield><subfield code="2">bicssc</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">COMPUTERS</subfield><subfield code="x">General.</subfield><subfield code="2">bisacsh</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Artificial intelligence</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Machine learning</subfield><subfield code="2">fast</subfield></datafield><datafield tag="650" ind1=" " ind2="7"><subfield code="a">Python (Computer program language)</subfield><subfield code="2">fast</subfield></datafield><datafield tag="655" ind1=" " ind2="4"><subfield code="a">Electronic book.</subfield></datafield><datafield tag="758" ind1=" " ind2=" "><subfield code="i">has work:</subfield><subfield code="a">Natural language processing with TensorFlow (Text)</subfield><subfield code="1">https://id.oclc.org/worldcat/entity/E39PCGPFcfyqXFQwPxgyHWGQ9C</subfield><subfield code="4">https://id.oclc.org/worldcat/ontology/hasWork</subfield></datafield><datafield tag="776" ind1="0" ind2="8"><subfield code="i">Print version:</subfield><subfield code="a">Ganegedara, Thushan.</subfield><subfield code="t">Natural Language Processing with TensorFlow : Teach language to machines using Python's deep learning library.</subfield><subfield code="d">Birmingham : Packt Publishing, ©2018</subfield></datafield><datafield tag="856" ind1="4" ind2="0"><subfield code="l">FWS01</subfield><subfield code="p">ZDB-4-EBA</subfield><subfield code="q">FWS_PDA_EBA</subfield><subfield code="u">https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1823678</subfield><subfield code="3">Volltext</subfield></datafield><datafield tag="938" ind1=" " ind2=" "><subfield code="a">Askews and Holts Library Services</subfield><subfield code="b">ASKH</subfield><subfield code="n">BDZ0036924764</subfield></datafield><datafield tag="938" ind1=" " ind2=" "><subfield code="a">EBL - Ebook Library</subfield><subfield code="b">EBLB</subfield><subfield code="n">EBL5405681</subfield></datafield><datafield tag="938" ind1=" " ind2=" "><subfield code="a">EBSCOhost</subfield><subfield code="b">EBSC</subfield><subfield code="n">1823678</subfield></datafield><datafield tag="994" ind1=" " ind2=" "><subfield code="a">92</subfield><subfield code="b">GEBAY</subfield></datafield><datafield tag="912" ind1=" " ind2=" "><subfield code="a">ZDB-4-EBA</subfield></datafield><datafield tag="049" ind1=" " ind2=" "><subfield code="a">DE-863</subfield></datafield></record></collection> |
genre | Electronic book. |
genre_facet | Electronic book. |
id | ZDB-4-EBA-on1039700926 |
illustrated | Not Illustrated |
indexdate | 2024-11-27T13:29:00Z |
institution | BVB |
isbn | 9781788477758 1788477758 |
language | English |
oclc_num | 1039700926 |
open_access_boolean | |
owner | MAIN DE-863 DE-BY-FWS |
owner_facet | MAIN DE-863 DE-BY-FWS |
physical | 1 online resource (472 pages) |
psigel | ZDB-4-EBA |
publishDate | 2018 |
publishDateSearch | 2018 |
publishDateSort | 2018 |
publisher | Packt, |
record_format | marc |
spelling | Ganegedara, Thushan, author. Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / Thushan Ganegedara. Birmingham, UK : Packt, [2018] ©2018 1 online resource (472 pages) text txt rdacontent computer c rdamedia online resource cr rdacarrier Includes bibliographical references and index. Print version record. Cover; Copyright; Packt Upsell; Contributors; Table of Contents; Preface; Chapter 1: Introduction to Natural Language Processing; What is Natural Language Processing?; Tasks of Natural Language Processing; The traditional approach to Natural Language Processing; Understanding the traditional approach; Example -- generating football game summaries; Drawbacks of the traditional approach; The deep learning approach to Natural Language Processing; History of deep learning; The current state of deep learning and NLP; Understanding a simple deep model -- a Fully Connected Neural Network. The roadmap -- beyond this chapterIntroduction to the technical tools; Description of the tools; Installing Python and scikit-learn; Installing Jupyter Notebook; Installing TensorFlow; Summary; Chapter 2: Understanding TensorFlow; What is TensorFlow?; Getting started with TensorFlow; TensorFlow client in detail; TensorFlow architecture -- what happens when you execute the client?; Cafe Le TensorFlow -- understanding TensorFlow with an analogy; Inputs, variables, outputs, and operations; Defining inputs in TensorFlow; Feeding data with Python code; Preloading and storing data as tensors. Building an input pipelineDefining variables in TensorFlow; Defining TensorFlow outputs; Defining TensorFlow operations; Comparison operations; Mathematical operations; Scatter and gather operations; Neural network-related operations; Reusing variables with scoping; Implementing our first neural network; Preparing the data; Defining the TensorFlow graph; Running the neural network; Summary; Chapter 3: Word2vec -- Learning Word Embeddings; What is a word representation or meaning?; Classical approaches to learning word representation. WordNet -- using an external lexical knowledge base for learning word representationsTour of WordNet; Problems with WordNet; One-hot encoded representation; The TF-IDF method; Co-occurrence matrix; Word2vec -- a neural network-based approach to learning word representation; Exercise: is queen = king -- he + she?; Designing a loss function for learning word embeddings; The skip-gram algorithm; From raw text to structured data; Learning the word embeddings with a neural network; Formulating a practical loss function; Efficiently approximating the loss function. Implementing skip-gram with TensorFlowThe Continuous Bag-of-Words algorithm; Implementing CBOW in TensorFlow; Summary; Chapter 4: Advanced Word2vec; The original skip-gram algorithm; Implementing the original skip-gram algorithm; Comparing the original skip-gram with the improved skip-gram; Comparing skip-gram with CBOW; Performance comparison; Which is the winner, skip-gram or CBOW?; Extensions to the word embeddings algorithms; Using the unigram distribution for negative sampling; Implementing unigram-based negative sampling; Subsampling -- probabilistically ignoring the common words. Implementing subsampling. TensorFlow is the leading framework for deep learning algorithms critical to artificial intelligence, and natural language processing (NLP) makes much of the data used by deep learning applications accessible to them. This book brings the two together and teaches deep learning developers how to work with today's vast amount of unstructured data. Machine learning. http://id.loc.gov/authorities/subjects/sh85079324 Artificial intelligence. http://id.loc.gov/authorities/subjects/sh85008180 Python (Computer program language) http://id.loc.gov/authorities/subjects/sh96008834 Apprentissage automatique. Intelligence artificielle. Python (Langage de programmation) artificial intelligence. aat Programming & scripting languages: general. bicssc Neural networks & fuzzy systems. bicssc Artificial intelligence. bicssc COMPUTERS General. bisacsh Artificial intelligence fast Machine learning fast Python (Computer program language) fast Electronic book. has work: Natural language processing with TensorFlow (Text) https://id.oclc.org/worldcat/entity/E39PCGPFcfyqXFQwPxgyHWGQ9C https://id.oclc.org/worldcat/ontology/hasWork Print version: Ganegedara, Thushan. Natural Language Processing with TensorFlow : Teach language to machines using Python's deep learning library. Birmingham : Packt Publishing, ©2018 FWS01 ZDB-4-EBA FWS_PDA_EBA https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1823678 Volltext |
spellingShingle | Ganegedara, Thushan Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / Cover; Copyright; Packt Upsell; Contributors; Table of Contents; Preface; Chapter 1: Introduction to Natural Language Processing; What is Natural Language Processing?; Tasks of Natural Language Processing; The traditional approach to Natural Language Processing; Understanding the traditional approach; Example -- generating football game summaries; Drawbacks of the traditional approach; The deep learning approach to Natural Language Processing; History of deep learning; The current state of deep learning and NLP; Understanding a simple deep model -- a Fully Connected Neural Network. The roadmap -- beyond this chapterIntroduction to the technical tools; Description of the tools; Installing Python and scikit-learn; Installing Jupyter Notebook; Installing TensorFlow; Summary; Chapter 2: Understanding TensorFlow; What is TensorFlow?; Getting started with TensorFlow; TensorFlow client in detail; TensorFlow architecture -- what happens when you execute the client?; Cafe Le TensorFlow -- understanding TensorFlow with an analogy; Inputs, variables, outputs, and operations; Defining inputs in TensorFlow; Feeding data with Python code; Preloading and storing data as tensors. Building an input pipelineDefining variables in TensorFlow; Defining TensorFlow outputs; Defining TensorFlow operations; Comparison operations; Mathematical operations; Scatter and gather operations; Neural network-related operations; Reusing variables with scoping; Implementing our first neural network; Preparing the data; Defining the TensorFlow graph; Running the neural network; Summary; Chapter 3: Word2vec -- Learning Word Embeddings; What is a word representation or meaning?; Classical approaches to learning word representation. WordNet -- using an external lexical knowledge base for learning word representationsTour of WordNet; Problems with WordNet; One-hot encoded representation; The TF-IDF method; Co-occurrence matrix; Word2vec -- a neural network-based approach to learning word representation; Exercise: is queen = king -- he + she?; Designing a loss function for learning word embeddings; The skip-gram algorithm; From raw text to structured data; Learning the word embeddings with a neural network; Formulating a practical loss function; Efficiently approximating the loss function. Implementing skip-gram with TensorFlowThe Continuous Bag-of-Words algorithm; Implementing CBOW in TensorFlow; Summary; Chapter 4: Advanced Word2vec; The original skip-gram algorithm; Implementing the original skip-gram algorithm; Comparing the original skip-gram with the improved skip-gram; Comparing skip-gram with CBOW; Performance comparison; Which is the winner, skip-gram or CBOW?; Extensions to the word embeddings algorithms; Using the unigram distribution for negative sampling; Implementing unigram-based negative sampling; Subsampling -- probabilistically ignoring the common words. Machine learning. http://id.loc.gov/authorities/subjects/sh85079324 Artificial intelligence. http://id.loc.gov/authorities/subjects/sh85008180 Python (Computer program language) http://id.loc.gov/authorities/subjects/sh96008834 Apprentissage automatique. Intelligence artificielle. Python (Langage de programmation) artificial intelligence. aat Programming & scripting languages: general. bicssc Neural networks & fuzzy systems. bicssc Artificial intelligence. bicssc COMPUTERS General. bisacsh Artificial intelligence fast Machine learning fast Python (Computer program language) fast |
subject_GND | http://id.loc.gov/authorities/subjects/sh85079324 http://id.loc.gov/authorities/subjects/sh85008180 http://id.loc.gov/authorities/subjects/sh96008834 |
title | Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / |
title_auth | Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / |
title_exact_search | Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / |
title_full | Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / Thushan Ganegedara. |
title_fullStr | Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / Thushan Ganegedara. |
title_full_unstemmed | Natural language processing with TensorFlow : teach language to machines using Python's deep learning library / Thushan Ganegedara. |
title_short | Natural language processing with TensorFlow : |
title_sort | natural language processing with tensorflow teach language to machines using python s deep learning library |
title_sub | teach language to machines using Python's deep learning library / |
topic | Machine learning. http://id.loc.gov/authorities/subjects/sh85079324 Artificial intelligence. http://id.loc.gov/authorities/subjects/sh85008180 Python (Computer program language) http://id.loc.gov/authorities/subjects/sh96008834 Apprentissage automatique. Intelligence artificielle. Python (Langage de programmation) artificial intelligence. aat Programming & scripting languages: general. bicssc Neural networks & fuzzy systems. bicssc Artificial intelligence. bicssc COMPUTERS General. bisacsh Artificial intelligence fast Machine learning fast Python (Computer program language) fast |
topic_facet | Machine learning. Artificial intelligence. Python (Computer program language) Apprentissage automatique. Intelligence artificielle. Python (Langage de programmation) artificial intelligence. Programming & scripting languages: general. Neural networks & fuzzy systems. COMPUTERS General. Artificial intelligence Machine learning Electronic book. |
url | https://search.ebscohost.com/login.aspx?direct=true&scope=site&db=nlebk&AN=1823678 |
work_keys_str_mv | AT ganegedarathushan naturallanguageprocessingwithtensorflowteachlanguagetomachinesusingpythonsdeeplearninglibrary |