Supervised Sequence Labelling with Recurrent Neural Networks
Alex Graves
Verkauft von AHA-BUCH GmbH, Einbeck, Deutschland
AbeBooks-Verkäufer seit 14. August 2006
Neu - Softcover
Zustand: Neu
Anzahl: 1 verfügbar
In den Warenkorb legenVerkauft von AHA-BUCH GmbH, Einbeck, Deutschland
AbeBooks-Verkäufer seit 14. August 2006
Zustand: Neu
Anzahl: 1 verfügbar
In den Warenkorb legenDruck auf Anfrage Neuware - Printed after ordering - Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning tools-robust to input noise and distortion, able to exploit long-range contextual information-that would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary. The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal. Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video. Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.
Bestandsnummer des Verkäufers 9783642432187
Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning tools—robust to input noise and distortion, able to exploit long-range contextual information—that would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary.
The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal. Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video.
Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.
Supervised sequence labelling is a vital area of machine learning, encompassing tasks such as speech, handwriting and gesture recognition, protein secondary structure prediction and part-of-speech tagging. Recurrent neural networks are powerful sequence learning tools--robust to input noise and distortion, able to exploit long-range contextual information--that would seem ideally suited to such problems. However their role in large-scale sequence labelling systems has so far been auxiliary.
The goal of this book is a complete framework for classifying and transcribing sequential data with recurrent neural networks only. Three main innovations are introduced in order to realise this goal. Firstly, the connectionist temporal classification output layer allows the framework to be trained with unsegmented target sequences, such as phoneme-level speech transcriptions; this is in contrast to previous connectionist approaches, which were dependent on error-prone prior segmentation. Secondly, multidimensional recurrent neural networks extend the framework in a natural way to data with more than one spatio-temporal dimension, such as images and videos. Thirdly, the use of hierarchical subsampling makes it feasible to apply the framework to very large or high resolution sequences, such as raw audio or video.
Experimental validation is provided by state-of-the-art results in speech and handwriting recognition.
„Über diesen Titel“ kann sich auf eine andere Ausgabe dieses Titels beziehen.
Allgemeine Geschäftsbedingungen und Kundeninformationen / Datenschutzerklärung
I. Allgemeine Geschäftsbedingungen
§ 1 Grundlegende Bestimmungen
(1) Die nachstehenden Geschäftsbedingungen gelten für alle Verträge, die Sie mit uns als Anbieter (AHA-BUCH GmbH) über die Internetplattformen AbeBooks und/oder ZVAB schließen. Soweit nicht anders vereinbart, wird der Einbeziehung gegebenenfalls von Ihnen verwendeter eigener Bedingungen widersprochen.
(2) Verbraucher im Sinne der nachstehenden Regelungen...
Mehr InformationWir liefern Lagerartikel innerhalb von 24 Stunden nach Erhalt der Bestellung aus.
Barsortimentsartikel, die wir über Nacht geliefert bekommen, am darauffolgenden Werktag.
Unser Ziel ist es Ihnen die Artikel in der ökonomischten und effizientesten Weise zu senden.