This book introduces basic supervised learning algorithms applicable to natural language processing (NLP) and shows how the performance of these algorithms can often be improved by exploiting the marginal distribution of large amounts of unlabeled data. One reason for that is data sparsity, i.e., the limited amounts of data we have available in NLP. However, in most real-world NLP applications our labeled data is also heavily biased. This book introduces extensions of supervised learning algorithms to cope with data sparsity and different kinds of sampling bias. This book is intended to be both readable by first-year students and interesting to the expert audience. My intention was to introduce what is necessary to appreciate the major challenges we face in contemporary NLP related to data sparsity and sampling bias, without wasting too much time on details about supervised learning algorithms or particular NLP applications. I use text classification, part-of-speech tagging, and dependency parsing as running examples, and limit myself to a small set of cardinal learning algorithms. I have worried less about theoretical guarantees ("this algorithm never does too badly") than about useful rules of thumb ("in this case this algorithm may perform really well"). In NLP, data is so noisy, biased, and non-stationary that few theoretical guarantees can be established and we are typically left with our gut feelings and a catalogue of crazy ideas. I hope this book will provide its readers with both. Throughout the book we include snippets of Python code and empirical evaluations, when relevant.
Die Inhaltsangabe kann sich auf eine andere Ausgabe dieses Titels beziehen.
Anders Søgaard is a father of three and a published poet, as well as a Full Professor in Computer Science the University of Copenhagen. He is currently funded by the Novo Nordisk Foundation, the Lundbeck Foundation, and the Innovation Fund Denmark; before that, he held an ERC Starting Grant and a Google Focused Research Award. He has won best paper awards at NAACL, EACL, CoNLL, etc. He previously wrote Semi-Supervised Learning and Domain Adaptation in NLP (Morgan & Claypool, 2013) and Cross-Lingual Word Embeddings (Morgan & Claypool, 2019), the latter with co-authors Ivan Vulic, Sebastian Ruder, and Manaal Faruqui.
„Über diesen Titel“ kann sich auf eine andere Ausgabe dieses Titels beziehen.
EUR 2,25 für den Versand innerhalb von/der USA
Versandziele, Kosten & DauerEUR 3,41 für den Versand innerhalb von/der USA
Versandziele, Kosten & DauerAnbieter: Lucky's Textbooks, Dallas, TX, USA
Zustand: New. Bestandsnummer des Verkäufers ABLIING23Mar3113020035051
Anzahl: Mehr als 20 verfügbar
Anbieter: GreatBookPrices, Columbia, MD, USA
Zustand: New. Bestandsnummer des Verkäufers 44545671-n
Anzahl: Mehr als 20 verfügbar
Anbieter: Best Price, Torrance, CA, USA
Zustand: New. SUPER FAST SHIPPING. Bestandsnummer des Verkäufers 9783031010217
Anzahl: 1 verfügbar
Anbieter: California Books, Miami, FL, USA
Zustand: New. Bestandsnummer des Verkäufers I-9783031010217
Anzahl: Mehr als 20 verfügbar
Anbieter: GreatBookPrices, Columbia, MD, USA
Zustand: As New. Unread book in perfect condition. Bestandsnummer des Verkäufers 44545671
Anzahl: Mehr als 20 verfügbar
Anbieter: Books Puddle, New York, NY, USA
Zustand: New. 1st edition NO-PA16APR2015-KAP. Bestandsnummer des Verkäufers 26394683698
Anzahl: 4 verfügbar
Anbieter: Ria Christie Collections, Uxbridge, Vereinigtes Königreich
Zustand: New. In. Bestandsnummer des Verkäufers ria9783031010217_new
Anzahl: Mehr als 20 verfügbar
Anbieter: Majestic Books, Hounslow, Vereinigtes Königreich
Zustand: New. Print on Demand. Bestandsnummer des Verkäufers 401726189
Anzahl: 4 verfügbar
Anbieter: GreatBookPricesUK, Woodford Green, Vereinigtes Königreich
Zustand: New. Bestandsnummer des Verkäufers 44545671-n
Anzahl: Mehr als 20 verfügbar
Anbieter: BuchWeltWeit Ludwig Meier e.K., Bergisch Gladbach, Deutschland
Taschenbuch. Zustand: Neu. This item is printed on demand - it takes 3-4 days longer - Neuware -This book introduces basic supervised learning algorithms applicable to natural language processing (NLP) and shows how the performance of these algorithms can often be improved by exploiting the marginal distribution of large amounts of unlabeled data. One reason for that is data sparsity, i.e., the limited amounts of data we have available in NLP. However, in most real-world NLP applications our labeled data is also heavily biased. This book introduces extensions of supervised learning algorithms to cope with data sparsity and different kinds of sampling bias. This book is intended to be both readable by first-year students and interesting to the expert audience. My intention was to introduce what is necessary to appreciate the major challenges we face in contemporary NLP related to data sparsity and sampling bias, without wasting too much time on details about supervised learning algorithms or particular NLP applications. I use text classification, part-of-speech tagging, and dependency parsing as running examples, and limit myself to a small set of cardinal learning algorithms. I have worried less about theoretical guarantees ('this algorithm never does too badly') than about useful rules of thumb ('in this case this algorithm may perform really well'). In NLP, data is so noisy, biased, and non-stationary that few theoretical guarantees can be established and we are typically left with our gut feelings and a catalogue of crazy ideas. I hope this book will provide its readers with both. Throughout the book we include snippets of Python code and empirical evaluations, when relevant. 104 pp. Englisch. Bestandsnummer des Verkäufers 9783031010217
Anzahl: 2 verfügbar