Apache Kafka is the platform that handles real-time data feeds with a high-throughput, and this book is all you need to harness its power, quickly and painlessly. A step by step tutorial with a practical approach.
Message publishing is a mechanism of connecting heterogeneous applications together with messages that are routed between them, for example by using a message broker like Apache Kafka. Such solutions deal with real-time volumes of information and route it to multiple consumers without letting information producers know who the final consumers are.
Apache Kafka is a practical, hands-on guide providing you with a series of step-by-step practical implementations, which will help you take advantage of the real power behind Kafka, and give you a strong grounding for using it in your publisher-subscriber based architectures.
Apache Kafka takes you through a number of clear, practical implementations that will help you to take advantage of the power of Apache Kafka, quickly and painlessly. You will learn everything you need to know for setting up Kafka clusters. This book explains how Kafka basic blocks like producers, brokers, and consumers actually work and fit together. You will then explore additional settings and configuration changes to achieve ever more complex goals. Finally you will learn how Kafka works with other tools like Hadoop, Storm, and so on.
You will learn everything you need to know to work with Apache Kafka in the right format, as well as how to leverage its power of handling hundreds of megabytes of messages per second from multiple clients.
What you will learn from this book
The book will follow a step-by-step tutorial approach which will show the readers how to use Apache Kafka for messaging from scratch.
Who this book is written for
Apache Kafka is for readers with software development experience, but no prior exposure to Apache Kafka or similar technologies is assumed. This book is also for enterprise application developers and big data enthusiasts who have worked with other publisher-subscriber based systems and now want to explore Apache Kafka as a futuristic scalable solution.
Die Inhaltsangabe kann sich auf eine andere Ausgabe dieses Titels beziehen.
Nishant Garg is a Technical Architect with more than 13 years' experience in various technologies such as Java Enterprise Edition, Spring, Hibernate, Hadoop, Hive, Flume, Sqoop, Oozie, Spark, Kafka, Storm, Mahout, and Solr/Lucene; NoSQL databases such as MongoDB, CouchDB, HBase and Cassandra, and MPP Databases such as GreenPlum and Vertica.
He has attained his M.S. in Software Systems from Birla Institute of Technology and Science, Pilani, India, and is currently a part of Big Data R&D team in innovation labs at Impetus Infotech Pvt. Ltd.
Nishant has enjoyed working with recognizable names in IT services and financial industries, employing full software lifecycle methodologies such as Agile and SCRUM. He has also undertaken many speaking engagements on Big Data technologies
„Über diesen Titel“ kann sich auf eine andere Ausgabe dieses Titels beziehen.