- Ocena:
- Bądź pierwszym, który oceni tę książkę
- Stron:
- 402
- Dostępne formaty:
-
ePubMobi
Opis ebooka: Kafka Connect
Used by more than 80% of Fortune 100 companies, Apache Kafka has become the de facto event streaming platform. Kafka Connect is a key component of Kafka that lets you flow data between your existing systems and Kafka to process data in real time.
With this practical guide, authors Mickael Maison and Kate Stanley show data engineers, site reliability engineers, and application developers how to build data pipelines between Kafka clusters and a variety of data sources and sinks. Kafka Connect allows you to quickly adopt Kafka by tapping into existing data and enabling many advanced use cases. No matter where you are in your event streaming journey, Kafka Connect is the ideal tool for building a modern data pipeline.
- Learn Kafka Connect's capabilities, main concepts, and terminology
- Design data and event streaming pipelines that use Kafka Connect
- Configure and operate Kafka Connect environments at scale
- Deploy secured and highly available Kafka Connect clusters
- Build sink and source connectors and single message transforms and converters
Wybrane bestsellery
-
Oprogramowanie Apache Kafka powstało jako broker wiadomości w LinkedIn. Obecnie pełni funkcję rozproszonego systemu przetwarzania strumieniowego danych, używanego do budowania aplikacji opracowujących duże ilości danych w czasie rzeczywistym. Z zalet tego oprogramowania korzystają firmy na całym ...
Apache Kafka. Kurs video. Przetwarzanie danych w czasie rzeczywistym Apache Kafka. Kurs video. Przetwarzanie danych w czasie rzeczywistym
(31.14 zł najniższa cena z 30 dni)40.05 zł
89.00 zł(-55%) -
This book describes both batch processing and real-time processing pipelines. You’ll learn how to implement basic and advanced big data use cases with ease and develop a deep understanding of the Apache Beam model. In addition to this, you’ll discover how the portability layer works...
Building Big Data Pipelines with Apache Beam. Use a single programming model for both batch and stream data processing Building Big Data Pipelines with Apache Beam. Use a single programming model for both batch and stream data processing
-
Every enterprise application creates data, including log messages, metrics, user activity, and outgoing messages. Learning how to move these items is almost as important as the data itself. If you're an application architect, developer, or production engineer new to Apache Pulsar, this practical ...(211.65 zł najniższa cena z 30 dni)
211.65 zł
249.00 zł(-15%) -
Data is bigger, arrives faster, and comes in a variety of formatsâ??and it all needs to be processed at scale for analytics or machine learning. But how can you process such varied workloads efficiently? Enter Apache Spark.Updated to include Spark 3.0, this second edition shows data engineer...(211.65 zł najniższa cena z 30 dni)
211.65 zł
249.00 zł(-15%) -
Serverless computing greatly simplifies software development. Your team can focus solely on your application while the cloud provider manages the servers you need. This practical guide shows you step-by-step how to build and deploy complex applications in a flexible multicloud, multilanguage envi...
Learning Apache OpenWhisk. Developing Open Serverless Solutions Learning Apache OpenWhisk. Developing Open Serverless Solutions
(211.65 zł najniższa cena z 30 dni)211.65 zł
249.00 zł(-15%) -
Before you can build analytics tools to gain quick insights, you first need to know how to process data in real time. With this practical guide, developers familiar with Apache Spark will learn how to put this in-memory framework to use for streaming data. You’ll discover how Spark enables ...
Stream Processing with Apache Spark. Mastering Structured Streaming and Spark Streaming Stream Processing with Apache Spark. Mastering Structured Streaming and Spark Streaming
(211.65 zł najniższa cena z 30 dni)211.65 zł
249.00 zł(-15%) -
This practical guide explains you to program and understand the power of Apache Cassandra 3.x. You will explore the integration and interaction of Cassandra components, and explore features such as the token allocation algorithm, CQL3, vnodes, lightweight transactions, and data modelling in detail.
Mastering Apache Cassandra 3.x. An expert guide to improving database scalability and availability without compromising performance - Third Edition Mastering Apache Cassandra 3.x. An expert guide to improving database scalability and availability without compromising performance - Third Edition
-
Apache Hive helps you deal with data summarization, queries, and analysis for huge amounts of data. This book will give you a background in big data, and familiarize you with your Hive working environment. Next you will cover advanced topics like performance and security in Hive and how to work e...
Apache Hive Essentials. Essential techniques to help you process, and get unique insights from, big data - Second Edition Apache Hive Essentials. Essential techniques to help you process, and get unique insights from, big data - Second Edition
-
Apache Spark is amazing when everything clicks. But if you haven’t seen the performance improvements you expected, or still don’t feel confident enough to use Spark in production, this practical book is for you. Authors Holden Karau and Rachel Warren demonstrate performance optimizati...
High Performance Spark. Best Practices for Scaling and Optimizing Apache Spark High Performance Spark. Best Practices for Scaling and Optimizing Apache Spark
(143.65 zł najniższa cena z 30 dni)143.65 zł
169.00 zł(-15%) -
More and more data-driven companies are looking to adopt stream processing and streaming analytics. With this concise ebook, you’ll learn best practices for designing a reliable architecture that supports this emerging big-data paradigm.Authors Ted Dunning and Ellen Friedman (Real World Had...
Streaming Architecture. New Designs Using Apache Kafka and MapR Streams Streaming Architecture. New Designs Using Apache Kafka and MapR Streams
(80.73 zł najniższa cena z 30 dni)80.73 zł
94.99 zł(-15%)
Ebooka "Kafka Connect" przeczytasz na:
-
czytnikach Inkbook, Kindle, Pocketbook, Onyx Boox i innych
-
systemach Windows, MacOS i innych
-
systemach Windows, Android, iOS, HarmonyOS
-
na dowolnych urządzeniach i aplikacjach obsługujących formaty: PDF, EPub, Mobi
Masz pytania? Zajrzyj do zakładki Pomoc »
Audiobooka "Kafka Connect" posłuchasz:
-
w aplikacji Ebookpoint na Android, iOS, HarmonyOs
-
na systemach Windows, MacOS i innych
-
na dowolnych urządzeniach i aplikacjach obsługujących format MP3 (pliki spakowane w ZIP)
Masz pytania? Zajrzyj do zakładki Pomoc »
Kurs Video "Kafka Connect" zobaczysz:
-
w aplikacjach Ebookpoint i Videopoint na Android, iOS, HarmonyOs
-
na systemach Windows, MacOS i innych z dostępem do najnowszej wersji Twojej przeglądarki internetowej
Szczegóły ebooka
- ISBN Ebooka:
- 978-10-981-2649-0, 9781098126490
- Data wydania ebooka:
- 2023-09-18 Data wydania ebooka często jest dniem wprowadzenia tytułu do sprzedaży i może nie być równoznaczna z datą wydania książki papierowej. Dodatkowe informacje możesz znaleźć w darmowym fragmencie. Jeśli masz wątpliwości skontaktuj się z nami sklep@ebookpoint.pl.
- Język publikacji:
- angielski
- Rozmiar pliku ePub:
- 3.9MB
- Rozmiar pliku Mobi:
- 9.6MB
Spis treści ebooka
- Foreword
- Preface
- Who Should Read This Book
- Kafka Versions
- Navigating This Book
- Conventions Used in This Book
- OReilly Online Learning
- How to Contact Us
- Acknowledgements
- I. Introduction to Kafka Connect
- 1. Meet Kafka Connect
- Kafka Connect Features
- Pluggable Architecture
- Scalability and Reliability
- Declarative Pipeline Definition
- Part of Apache Kafka
- Kafka Connect Features
- Use Cases
- Capturing Database Changes
- Mirroring Kafka Clusters
- Building Data Lakes
- Aggregating Logs
- Modernizing Legacy Systems
- Alternatives to Kafka Connect
- Summary
- 2. Apache Kafka Basics
- A Distributed Event Streaming Platform
- Open Source
- Distributed
- Event Streaming
- Platform
- A Distributed Event Streaming Platform
- Kafka Concepts
- Publish-Subscribe
- Brokers and Records
- Topics and Partitions
- Replication
- Retention and Compaction
- KRaft and ZooKeeper
- Interacting with Kafka
- Producers
- Consumers
- Kafka Streams
- Getting Started with Kafka
- Starting Kafka
- Kafka in KRaft mode (without ZooKeeper)
- Kafka with ZooKeeper
- Starting Kafka
- Sending and Receiving Records
- Running a Kafka Streams Application
- Summary
- II. Developing Data Pipelines with Kafka Connect
- 3. Components in a Kafka Connect Data Pipeline
- Kafka Connect Runtime
- Running Kafka Connect
- Kafka Connect REST API
- Installing Plug-Ins
- Deployment Modes
- Kafka Connect Runtime
- Source and Sink Connectors
- Connectors and Tasks
- Configuring Connectors
- Running Connectors
- Converters
- Data Format and Schemas
- Configuring Converters
- Using Converters
- Transformations and Predicates
- Transformation Use Cases
- Routing
- Sanitizing
- Formatting
- Enhancing
- Transformation Use Cases
- Predicates
- Configuring Transformations and Predicates
- Using Transformations and Predicates
- Summary
- 4. Designing Effective Data Pipelines
- Choosing a Connector
- Pipeline Direction
- Licensing and Support
- Connector Features
- Choosing a Connector
- Defining Data Models
- Data Transformation
- Mapping Data Between Systems
- Formatting Data
- Data Formats
- Schemas
- Kafka Connect record schemas
- Kafka record schemas
- Exploring Kafka Connect Internals
- Internal Topics
- Group Membership
- Rebalance Protocols
- Handling Failures in Kafka Connect
- Worker Failure
- Connector/Task Failure
- Kafka/External Systems Failure
- Dead Letter Queues
- Understanding Processing Semantics
- Sink Connectors
- Source Connectors
- Summary
- 5. Connectors in Action
- Confluent S3 Sink Connector
- Configuring the Connector
- Connectivity and S3 details
- Object partitioning
- Object naming
- Object formats
- Object upload
- Configuring the Connector
- Exactly-Once Semantics
- Running the Connector
- Using the field partitioner
- Using the time-based partitioner
- Confluent S3 Sink Connector
- Confluent JDBC Source Connector
- Configuring the Connector
- Connectivity
- Topic naming
- Table filtering
- Data collection mode
- Partitioning and parallelism
- Configuring the Connector
- Running the Connector
- Using the bulk mode
- Using an incrementing mode
- Debezium MySQL Source Connector
- Configuring the Connector
- Connectivity
- Database and table filtering
- Snapshotting
- Configuring the Connector
- Event Formats
- Running the Connector
- Summary
- 6. Mirroring Clusters with MirrorMaker
- Introduction to Mirroring
- Exploring Mirroring Use Cases
- Geo-replication
- Disaster recovery
- Migration
- Complex topologies
- Exploring Mirroring Use Cases
- Mirroring in Practice
- Introduction to Mirroring
- Introduction to MirrorMaker
- Common Concepts
- Local and remote topics
- Common configurations
- Replication policies
- Client overrides
- Common Concepts
- Deployment Modes
- MirrorMaker Connectors
- MirrorSourceConnector
- Configurations
- Topic configurations
- Offset-syncs configurations
- ACLs configurations
- Metrics configurations
- Configurations
- Permissions
- Source cluster ACLs
- Target cluster ACLs
- MirrorSourceConnector
- Metrics
- MirrorCheckpointConnector
- Configurations
- Permissions
- Source cluster ACLs
- Target cluster ACLs
- Metrics
- MirrorHeartbeatConnector
- Configurations
- Permissions
- Running MirrorMaker
- Disaster Recovery Example
- Geo-Replication Example
- Summary
- III. Running Kafka Connect in Production
- 7. Deploying and Operating Kafka Connect Clusters
- Preparing the Kafka Connect Environment
- Building a Kafka Connect Environment
- Installing Plug-Ins
- Networking and Permissions
- Preparing the Kafka Connect Environment
- Worker Plug-Ins
- Configuration Providers
- REST Extensions
- Connector Client Configuration Override Policies
- Sizing and Planning Capacity
- Understanding Kafka Connect Resource Utilization
- How Many Workers and Tasks?
- Single cluster versus separate clusters
- Maintainability
- Isolation
- Security
- Use case optimization
- Operating Kafka Connect Clusters
- Adding Workers
- Removing Workers
- Upgrading and Applying Maintenance to Workers
- Restarting Failed Tasks and Connectors
- Resetting Offsets of Connectors
- Sink connector offsets
- Source connector offsets
- Administering Kafka Connect Using the REST API
- Creating and Deleting a Connector
- Connector and Task Configuration
- Controlling the Lifecycle of Connectors
- Listing Connector Offsets
- Debugging Issues
- Summary
- 8. Configuring Kafka Connect
- Configuring the Runtime
- Configurations for Production
- Clients and connector overrides
- REST configurations
- Miscellaneous configuration
- Configurations for Production
- Fine-Tuning Configurations
- Connection configurations
- Inter-worker and rebalance configurations
- Topic tracking configurations
- Metrics configurations
- Offset flush configurations
- Configuring the Runtime
- Configuring Connectors
- Topic Configurations
- Client Overrides
- Configurations for Exactly-Once
- Configurations for Error Handling
- Configuring Kafka Connect Clusters for Security
- Securing the Connection to Kafka
- TLS configurations
- SASL configurations
- SASL OAUTHBEARER configurations
- SASL GSSAPI configurations
- Securing the Connection to Kafka
- Configuring Permissions
- Securing the REST API
- Summary
- 9. Monitoring Kafka Connect
- Monitoring Logs
- Logging Configuration
- Understanding Startup Logs
- Analyzing Logs
- Log contexts
- Key events
- Errors
- Monitoring Logs
- Monitoring Metrics
- Metrics Reporters
- Analyzing Metrics
- Exploring Metrics
- Key Metrics
- Kafka Connect Runtime Metrics
- Metadata metrics
- Network metrics
- Group protocol metrics
- Connector-level metrics
- Task-level metrics
- Kafka Connect Runtime Metrics
- Other System Metrics
- Internal Kafka client metrics
- Kafka and external system metrics
- Summary
- 10. Administering Kafka Connect on Kubernetes
- Introduction to Kubernetes
- Virtualization Technologies
- Kubernetes Fundamentals
- Introduction to Kubernetes
- Running Kafka Connect on Kubernetes
- Container Image
- Deploying Workers
- Networking and Monitoring
- Configuration
- Using a Kubernetes Operator to Deploy Kafka Connect
- Introduction to Kubernetes Operators
- Kubernetes Operators for Kafka Connect
- Strimzi
- Getting a Kubernetes Environment
- Starting the Operator
- Kafka Connect CRDs
- Deploying a Kafka Connect Cluster and Connectors
- MirrorMaker CRD
- Summary
- IV. Building Custom Connectors and Plug-Ins
- 11. Building Source and Sink Connectors
- Common Concepts and APIs
- Building a Custom Connector
- Implementing a connector
- Packaging a connector
- Building a Custom Connector
- The Connector API
- The version() method
- The config() method
- The initialize() method
- The start() method
- The taskClass() method
- The taskConfigs() method
- The stop() method
- The validate() method
- The context() methods
- Connector API lifecycle
- Common Concepts and APIs
- Configurations
- Configuration types
- Validators and recommenders
- Interacting with configurations at runtime
- The Task API
- The initialize() methods
- The start() method
- The stop() method
- Task API lifecycle
- Kafka Connect Records
- Schemas
- The ConnectorContext API
- The requestTaskReconfiguration() method
- The raiseError() method
- The configs() method
- Implementing Source Connectors
- The SourceTask API
- The poll() method
- The commit() and commitRecord() methods
- SourceTask API lifecycle
- The SourceTask API
- Source Records
- The SourceConnectorContext and SourceTaskContext APIs
- The offsetStorageReader() method
- The transactionContext() method
- Exactly-Once Support
- The exactlyOnceSupport() method
- The canDefineTransactionBoundaries() method
- The commitTransaction() methods
- The abortTransaction() methods
- Implementing Sink Connectors
- The SinkTask API
- The put() method
- The preCommit() method
- The flush() method
- The open() and close() methods
- The SinkTask API lifecycle
- The SinkTask API
- Sink Records
- The SinkConnectorContext and SinkTaskContext APIs
- The offset() methods
- The timeout() method
- The assignment() method
- The pause() and resume() methods
- The requestCommit() method
- The errantRecordReporter() method
- Summary
- 12. Extending Kafka Connect with Connector and Worker Plug-Ins
- Implementing Connector Plug-Ins
- The Transformation API
- The apply() method
- The config() method
- The configure() method
- The close() method
- The Transformation API
- The Predicate API
- The test() method
- The config() method
- The configure() method
- The close() method
- Implementing Connector Plug-Ins
- The Converter and HeaderConverter APIs
- The fromConnectData() methods
- The toConnectData() methods
- The fromConnectHeader() method
- The toConnectHeader() method
- The config() methods
- The configure() methods
- The close() method
- Implementing Worker Plug-Ins
- The ConfigProvider API
- The get() methods
- The configure() method
- The close() method
- The subscribe(), unsubscribe(), and unsubscribeAll() methods
- The ConfigProvider API
- The ConnectorClientConfigOverridePolicy API
- The validate() method
- The configure() method
- The close() method
- The ConnectRestExtension APIs
- The register() method
- The configure() method
- The close() method
- The version() method
- Summary
- Index
O'Reilly Media - inne książki
-
Software as a service (SaaS) is on the path to becoming the de facto model for building, delivering, and operating software solutions. Adopting a multi-tenant SaaS model requires builders to take on a broad range of new architecture, implementation, and operational challenges. How data is partiti...(237.15 zł najniższa cena z 30 dni)
245.65 zł
289.00 zł(-15%) -
Great engineers don't necessarily make great leaders—at least, not without a lot of work. Finding your path to becoming a strong leader is often fraught with challenges. It's not easy to figure out how to be strategic, successful, and considerate while also being firm. Whether you're on the...(126.65 zł najniższa cena z 30 dni)
126.65 zł
149.00 zł(-15%) -
Data science happens in code. The ability to write reproducible, robust, scaleable code is key to a data science project's success—and is absolutely essential for those working with production code. This practical book bridges the gap between data science and software engineering,and clearl...(211.65 zł najniższa cena z 30 dni)
211.65 zł
249.00 zł(-15%) -
With the massive adoption of microservices, operators and developers face far more complexity in their applications today. Service meshes can help you manage this problem by providing a unified control plane to secure, manage, and monitor your entire network. This practical guide shows you how th...(203.15 zł najniższa cena z 30 dni)
203.15 zł
239.00 zł(-15%) -
Get practical advice on how to leverage AI development tools for all stages of code creation, including requirements, planning, design, coding, debugging, testing, and documentation. With this book, beginners and experienced developers alike will learn how to use a wide range of tools, from gener...(164.25 zł najniższa cena z 30 dni)
186.15 zł
219.00 zł(-15%) -
Rust's popularity is growing, due in part to features like memory safety, type safety, and thread safety. But these same elements can also make learning Rust a challenge, even for experienced programmers. This practical guide helps you make the transition to writing idiomatic Rust—while als...(164.25 zł najniższa cena z 30 dni)
186.15 zł
219.00 zł(-15%) -
Advance your Power BI skills by adding AI to your repertoire at a practice level. With this practical book, business-oriented software engineers and developers will learn the terminologies, practices, and strategy necessary to successfully incorporate AI into your business intelligence estate. Je...(211.65 zł najniższa cena z 30 dni)
211.65 zł
249.00 zł(-15%) -
Microservices can be a very effective approach for delivering value to your organization and to your customers. If you get them right, microservices help you to move fast by making changes to small parts of your system hundreds of times a day. But if you get them wrong, microservices will just ma...(203.15 zł najniższa cena z 30 dni)
203.15 zł
239.00 zł(-15%) -
JavaScript gives web developers great power to create rich interactive browser experiences, and much of that power is provided by the browser itself. Modern web APIs enable web-based applications to come to life like never before, supporting actions that once required browser plug-ins. Some are s...(186.15 zł najniższa cena z 30 dni)
186.15 zł
219.00 zł(-15%) -
How will software development and operations have to change to meet the sustainability and green needs of the planet? And what does that imply for development organizations? In this eye-opening book, sustainable software advocates Anne Currie, Sarah Hsu, and Sara Bergman provide a unique overview...(169.14 zł najniższa cena z 30 dni)
169.14 zł
199.00 zł(-15%)
Dzieki opcji "Druk na żądanie" do sprzedaży wracają tytuły Grupy Helion, które cieszyły sie dużym zainteresowaniem, a których nakład został wyprzedany.
Dla naszych Czytelników wydrukowaliśmy dodatkową pulę egzemplarzy w technice druku cyfrowego.
Co powinieneś wiedzieć o usłudze "Druk na żądanie":
- usługa obejmuje tylko widoczną poniżej listę tytułów, którą na bieżąco aktualizujemy;
- cena książki może być wyższa od początkowej ceny detalicznej, co jest spowodowane kosztami druku cyfrowego (wyższymi niż koszty tradycyjnego druku offsetowego). Obowiązująca cena jest zawsze podawana na stronie WWW książki;
- zawartość książki wraz z dodatkami (płyta CD, DVD) odpowiada jej pierwotnemu wydaniu i jest w pełni komplementarna;
- usługa nie obejmuje książek w kolorze.
Masz pytanie o konkretny tytuł? Napisz do nas: sklep[at]helion.pl.
Książka, którą chcesz zamówić pochodzi z końcówki nakładu. Oznacza to, że mogą się pojawić drobne defekty (otarcia, rysy, zagięcia).
Co powinieneś wiedzieć o usłudze "Końcówka nakładu":
- usługa obejmuje tylko książki oznaczone tagiem "Końcówka nakładu";
- wady o których mowa powyżej nie podlegają reklamacji;
Masz pytanie o konkretny tytuł? Napisz do nas: sklep[at]helion.pl.
Książka drukowana
Oceny i opinie klientów: Kafka Connect Mickael Maison, Kate Stanley (0) Weryfikacja opinii następuję na podstawie historii zamówień na koncie Użytkownika umieszczającego opinię. Użytkownik mógł otrzymać punkty za opublikowanie opinii uprawniające do uzyskania rabatu w ramach Programu Punktowego.