Große Auswahl an günstigen Büchern
Schnelle Lieferung per Post und DHL

Embeddings in Natural Language Processing

Über Embeddings in Natural Language Processing

Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents. This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP. Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.

Mehr anzeigen
  • Sprache:
  • Englisch
  • ISBN:
  • 9783031010491
  • Einband:
  • Taschenbuch
  • Seitenzahl:
  • 176
  • Veröffentlicht:
  • 13 November 2020
  • Abmessungen:
  • 191x10x235 mm.
  • Gewicht:
  • 341 g.
  Versandkostenfrei
  Versandfertig in 1-2 Wochen.

Beschreibung von Embeddings in Natural Language Processing

Embeddings have undoubtedly been one of the most influential research areas in Natural Language Processing (NLP). Encoding information into a low-dimensional vector representation, which is easily integrable in modern machine learning models, has played a central role in the development of NLP. Embedding techniques initially focused on words, but the attention soon started to shift to other forms: from graph structures, such as knowledge bases, to other types of textual content, such as sentences and documents. This book provides a high-level synthesis of the main embedding techniques in NLP, in the broad sense. The book starts by explaining conventional word vector space models and word embeddings (e.g., Word2Vec and GloVe) and then moves to other types of embeddings, such as word sense, sentence and document, and graph embeddings. The book also provides an overview of recent developments in contextualized representations (e.g., ELMo and BERT) and explains their potential in NLP. Throughout the book, the reader can find both essential information for understanding a certain topic from scratch and a broad overview of the most successful techniques developed in the literature.

Kund*innenbewertungen von Embeddings in Natural Language Processing



Ähnliche Bücher finden
Das Buch Embeddings in Natural Language Processing ist in den folgenden Kategorien erhältlich:

Willkommen bei den Tales Buchfreunden und -freundinnen

Jetzt zum Newsletter anmelden und tolle Angebote und Anregungen für Ihre nächste Lektüre erhalten.