- Ocena:
- Bądź pierwszym, który oceni tę książkę
- Stron:
- 402
- Dostępne formaty:
-
ePubMobi
Opis ebooka: Kafka Connect
Used by more than 80% of Fortune 100 companies, Apache Kafka has become the de facto event streaming platform. Kafka Connect is a key component of Kafka that lets you flow data between your existing systems and Kafka to process data in real time.
With this practical guide, authors Mickael Maison and Kate Stanley show data engineers, site reliability engineers, and application developers how to build data pipelines between Kafka clusters and a variety of data sources and sinks. Kafka Connect allows you to quickly adopt Kafka by tapping into existing data and enabling many advanced use cases. No matter where you are in your event streaming journey, Kafka Connect is the ideal tool for building a modern data pipeline.
- Learn Kafka Connect's capabilities, main concepts, and terminology
- Design data and event streaming pipelines that use Kafka Connect
- Configure and operate Kafka Connect environments at scale
- Deploy secured and highly available Kafka Connect clusters
- Build sink and source connectors and single message transforms and converters
Wybrane bestsellery
-
Oprogramowanie Apache Kafka powstało jako broker wiadomości w LinkedIn. Obecnie pełni funkcję rozproszonego systemu przetwarzania strumieniowego danych, używanego do budowania aplikacji opracowujących duże ilości danych w czasie rzeczywistym. Z zalet tego oprogramowania korzystają firmy na całym ...
Apache Kafka. Kurs video. Przetwarzanie danych w czasie rzeczywistym Apache Kafka. Kurs video. Przetwarzanie danych w czasie rzeczywistym
(39.90 zł najniższa cena z 30 dni)39.90 zł
89.00 zł(-55%) -
This book describes both batch processing and real-time processing pipelines. You’ll learn how to implement basic and advanced big data use cases with ease and develop a deep understanding of the Apache Beam model. In addition to this, you’ll discover how the portability layer works...
Building Big Data Pipelines with Apache Beam. Use a single programming model for both batch and stream data processing Building Big Data Pipelines with Apache Beam. Use a single programming model for both batch and stream data processing
-
Every enterprise application creates data, including log messages, metrics, user activity, and outgoing messages. Learning how to move these items is almost as important as the data itself. If you're an application architect, developer, or production engineer new to Apache Pulsar, this practical ...(211.65 zł najniższa cena z 30 dni)
220.15 zł
259.00 zł(-15%) -
Data is bigger, arrives faster, and comes in a variety of formatsâ??and it all needs to be processed at scale for analytics or machine learning. But how can you process such varied workloads efficiently? Enter Apache Spark.Updated to include Spark 3.0, this second edition shows data engineer...(211.65 zł najniższa cena z 30 dni)
220.15 zł
259.00 zł(-15%) -
Serverless computing greatly simplifies software development. Your team can focus solely on your application while the cloud provider manages the servers you need. This practical guide shows you step-by-step how to build and deploy complex applications in a flexible multicloud, multilanguage envi...
Learning Apache OpenWhisk. Developing Open Serverless Solutions Learning Apache OpenWhisk. Developing Open Serverless Solutions
(211.65 zł najniższa cena z 30 dni)220.15 zł
259.00 zł(-15%) -
Before you can build analytics tools to gain quick insights, you first need to know how to process data in real time. With this practical guide, developers familiar with Apache Spark will learn how to put this in-memory framework to use for streaming data. You’ll discover how Spark enables ...
Stream Processing with Apache Spark. Mastering Structured Streaming and Spark Streaming Stream Processing with Apache Spark. Mastering Structured Streaming and Spark Streaming
(211.65 zł najniższa cena z 30 dni)220.15 zł
259.00 zł(-15%) -
This practical guide explains you to program and understand the power of Apache Cassandra 3.x. You will explore the integration and interaction of Cassandra components, and explore features such as the token allocation algorithm, CQL3, vnodes, lightweight transactions, and data modelling in detail.
Mastering Apache Cassandra 3.x. An expert guide to improving database scalability and availability without compromising performance - Third Edition Mastering Apache Cassandra 3.x. An expert guide to improving database scalability and availability without compromising performance - Third Edition
-
Apache Hive helps you deal with data summarization, queries, and analysis for huge amounts of data. This book will give you a background in big data, and familiarize you with your Hive working environment. Next you will cover advanced topics like performance and security in Hive and how to work e...
Apache Hive Essentials. Essential techniques to help you process, and get unique insights from, big data - Second Edition Apache Hive Essentials. Essential techniques to help you process, and get unique insights from, big data - Second Edition
-
Apache Spark is amazing when everything clicks. But if you haven’t seen the performance improvements you expected, or still don’t feel confident enough to use Spark in production, this practical book is for you. Authors Holden Karau and Rachel Warren demonstrate performance optimizati...
High Performance Spark. Best Practices for Scaling and Optimizing Apache Spark High Performance Spark. Best Practices for Scaling and Optimizing Apache Spark
(143.65 zł najniższa cena z 30 dni)143.65 zł
169.00 zł(-15%) -
More and more data-driven companies are looking to adopt stream processing and streaming analytics. With this concise ebook, you’ll learn best practices for designing a reliable architecture that supports this emerging big-data paradigm.Authors Ted Dunning and Ellen Friedman (Real World Had...
Streaming Architecture. New Designs Using Apache Kafka and MapR Streams Streaming Architecture. New Designs Using Apache Kafka and MapR Streams
(80.73 zł najniższa cena z 30 dni)80.73 zł
94.99 zł(-15%)
Ebooka "Kafka Connect" przeczytasz na:
-
czytnikach Inkbook, Kindle, Pocketbook, Onyx Boox i innych
-
systemach Windows, MacOS i innych
-
systemach Windows, Android, iOS, HarmonyOS
-
na dowolnych urządzeniach i aplikacjach obsługujących formaty: PDF, EPub, Mobi
Masz pytania? Zajrzyj do zakładki Pomoc »
Audiobooka "Kafka Connect" posłuchasz:
-
w aplikacji Ebookpoint na Android, iOS, HarmonyOs
-
na systemach Windows, MacOS i innych
-
na dowolnych urządzeniach i aplikacjach obsługujących format MP3 (pliki spakowane w ZIP)
Masz pytania? Zajrzyj do zakładki Pomoc »
Kurs Video "Kafka Connect" zobaczysz:
-
w aplikacjach Ebookpoint i Videopoint na Android, iOS, HarmonyOs
-
na systemach Windows, MacOS i innych z dostępem do najnowszej wersji Twojej przeglądarki internetowej
Szczegóły ebooka
- ISBN Ebooka:
- 978-10-981-2649-0, 9781098126490
- Data wydania ebooka:
- 2023-09-18 Data wydania ebooka często jest dniem wprowadzenia tytułu do sprzedaży i może nie być równoznaczna z datą wydania książki papierowej. Dodatkowe informacje możesz znaleźć w darmowym fragmencie. Jeśli masz wątpliwości skontaktuj się z nami sklep@ebookpoint.pl.
- Język publikacji:
- angielski
- Rozmiar pliku ePub:
- 3.9MB
- Rozmiar pliku Mobi:
- 9.6MB
Spis treści ebooka
- Foreword
- Preface
- Who Should Read This Book
- Kafka Versions
- Navigating This Book
- Conventions Used in This Book
- OReilly Online Learning
- How to Contact Us
- Acknowledgements
- I. Introduction to Kafka Connect
- 1. Meet Kafka Connect
- Kafka Connect Features
- Pluggable Architecture
- Scalability and Reliability
- Declarative Pipeline Definition
- Part of Apache Kafka
- Kafka Connect Features
- Use Cases
- Capturing Database Changes
- Mirroring Kafka Clusters
- Building Data Lakes
- Aggregating Logs
- Modernizing Legacy Systems
- Alternatives to Kafka Connect
- Summary
- 2. Apache Kafka Basics
- A Distributed Event Streaming Platform
- Open Source
- Distributed
- Event Streaming
- Platform
- A Distributed Event Streaming Platform
- Kafka Concepts
- Publish-Subscribe
- Brokers and Records
- Topics and Partitions
- Replication
- Retention and Compaction
- KRaft and ZooKeeper
- Interacting with Kafka
- Producers
- Consumers
- Kafka Streams
- Getting Started with Kafka
- Starting Kafka
- Kafka in KRaft mode (without ZooKeeper)
- Kafka with ZooKeeper
- Starting Kafka
- Sending and Receiving Records
- Running a Kafka Streams Application
- Summary
- II. Developing Data Pipelines with Kafka Connect
- 3. Components in a Kafka Connect Data Pipeline
- Kafka Connect Runtime
- Running Kafka Connect
- Kafka Connect REST API
- Installing Plug-Ins
- Deployment Modes
- Kafka Connect Runtime
- Source and Sink Connectors
- Connectors and Tasks
- Configuring Connectors
- Running Connectors
- Converters
- Data Format and Schemas
- Configuring Converters
- Using Converters
- Transformations and Predicates
- Transformation Use Cases
- Routing
- Sanitizing
- Formatting
- Enhancing
- Transformation Use Cases
- Predicates
- Configuring Transformations and Predicates
- Using Transformations and Predicates
- Summary
- 4. Designing Effective Data Pipelines
- Choosing a Connector
- Pipeline Direction
- Licensing and Support
- Connector Features
- Choosing a Connector
- Defining Data Models
- Data Transformation
- Mapping Data Between Systems
- Formatting Data
- Data Formats
- Schemas
- Kafka Connect record schemas
- Kafka record schemas
- Exploring Kafka Connect Internals
- Internal Topics
- Group Membership
- Rebalance Protocols
- Handling Failures in Kafka Connect
- Worker Failure
- Connector/Task Failure
- Kafka/External Systems Failure
- Dead Letter Queues
- Understanding Processing Semantics
- Sink Connectors
- Source Connectors
- Summary
- 5. Connectors in Action
- Confluent S3 Sink Connector
- Configuring the Connector
- Connectivity and S3 details
- Object partitioning
- Object naming
- Object formats
- Object upload
- Configuring the Connector
- Exactly-Once Semantics
- Running the Connector
- Using the field partitioner
- Using the time-based partitioner
- Confluent S3 Sink Connector
- Confluent JDBC Source Connector
- Configuring the Connector
- Connectivity
- Topic naming
- Table filtering
- Data collection mode
- Partitioning and parallelism
- Configuring the Connector
- Running the Connector
- Using the bulk mode
- Using an incrementing mode
- Debezium MySQL Source Connector
- Configuring the Connector
- Connectivity
- Database and table filtering
- Snapshotting
- Configuring the Connector
- Event Formats
- Running the Connector
- Summary
- 6. Mirroring Clusters with MirrorMaker
- Introduction to Mirroring
- Exploring Mirroring Use Cases
- Geo-replication
- Disaster recovery
- Migration
- Complex topologies
- Exploring Mirroring Use Cases
- Mirroring in Practice
- Introduction to Mirroring
- Introduction to MirrorMaker
- Common Concepts
- Local and remote topics
- Common configurations
- Replication policies
- Client overrides
- Common Concepts
- Deployment Modes
- MirrorMaker Connectors
- MirrorSourceConnector
- Configurations
- Topic configurations
- Offset-syncs configurations
- ACLs configurations
- Metrics configurations
- Configurations
- Permissions
- Source cluster ACLs
- Target cluster ACLs
- MirrorSourceConnector
- Metrics
- MirrorCheckpointConnector
- Configurations
- Permissions
- Source cluster ACLs
- Target cluster ACLs
- Metrics
- MirrorHeartbeatConnector
- Configurations
- Permissions
- Running MirrorMaker
- Disaster Recovery Example
- Geo-Replication Example
- Summary
- III. Running Kafka Connect in Production
- 7. Deploying and Operating Kafka Connect Clusters
- Preparing the Kafka Connect Environment
- Building a Kafka Connect Environment
- Installing Plug-Ins
- Networking and Permissions
- Preparing the Kafka Connect Environment
- Worker Plug-Ins
- Configuration Providers
- REST Extensions
- Connector Client Configuration Override Policies
- Sizing and Planning Capacity
- Understanding Kafka Connect Resource Utilization
- How Many Workers and Tasks?
- Single cluster versus separate clusters
- Maintainability
- Isolation
- Security
- Use case optimization
- Operating Kafka Connect Clusters
- Adding Workers
- Removing Workers
- Upgrading and Applying Maintenance to Workers
- Restarting Failed Tasks and Connectors
- Resetting Offsets of Connectors
- Sink connector offsets
- Source connector offsets
- Administering Kafka Connect Using the REST API
- Creating and Deleting a Connector
- Connector and Task Configuration
- Controlling the Lifecycle of Connectors
- Listing Connector Offsets
- Debugging Issues
- Summary
- 8. Configuring Kafka Connect
- Configuring the Runtime
- Configurations for Production
- Clients and connector overrides
- REST configurations
- Miscellaneous configuration
- Configurations for Production
- Fine-Tuning Configurations
- Connection configurations
- Inter-worker and rebalance configurations
- Topic tracking configurations
- Metrics configurations
- Offset flush configurations
- Configuring the Runtime
- Configuring Connectors
- Topic Configurations
- Client Overrides
- Configurations for Exactly-Once
- Configurations for Error Handling
- Configuring Kafka Connect Clusters for Security
- Securing the Connection to Kafka
- TLS configurations
- SASL configurations
- SASL OAUTHBEARER configurations
- SASL GSSAPI configurations
- Securing the Connection to Kafka
- Configuring Permissions
- Securing the REST API
- Summary
- 9. Monitoring Kafka Connect
- Monitoring Logs
- Logging Configuration
- Understanding Startup Logs
- Analyzing Logs
- Log contexts
- Key events
- Errors
- Monitoring Logs
- Monitoring Metrics
- Metrics Reporters
- Analyzing Metrics
- Exploring Metrics
- Key Metrics
- Kafka Connect Runtime Metrics
- Metadata metrics
- Network metrics
- Group protocol metrics
- Connector-level metrics
- Task-level metrics
- Kafka Connect Runtime Metrics
- Other System Metrics
- Internal Kafka client metrics
- Kafka and external system metrics
- Summary
- 10. Administering Kafka Connect on Kubernetes
- Introduction to Kubernetes
- Virtualization Technologies
- Kubernetes Fundamentals
- Introduction to Kubernetes
- Running Kafka Connect on Kubernetes
- Container Image
- Deploying Workers
- Networking and Monitoring
- Configuration
- Using a Kubernetes Operator to Deploy Kafka Connect
- Introduction to Kubernetes Operators
- Kubernetes Operators for Kafka Connect
- Strimzi
- Getting a Kubernetes Environment
- Starting the Operator
- Kafka Connect CRDs
- Deploying a Kafka Connect Cluster and Connectors
- MirrorMaker CRD
- Summary
- IV. Building Custom Connectors and Plug-Ins
- 11. Building Source and Sink Connectors
- Common Concepts and APIs
- Building a Custom Connector
- Implementing a connector
- Packaging a connector
- Building a Custom Connector
- The Connector API
- The version() method
- The config() method
- The initialize() method
- The start() method
- The taskClass() method
- The taskConfigs() method
- The stop() method
- The validate() method
- The context() methods
- Connector API lifecycle
- Common Concepts and APIs
- Configurations
- Configuration types
- Validators and recommenders
- Interacting with configurations at runtime
- The Task API
- The initialize() methods
- The start() method
- The stop() method
- Task API lifecycle
- Kafka Connect Records
- Schemas
- The ConnectorContext API
- The requestTaskReconfiguration() method
- The raiseError() method
- The configs() method
- Implementing Source Connectors
- The SourceTask API
- The poll() method
- The commit() and commitRecord() methods
- SourceTask API lifecycle
- The SourceTask API
- Source Records
- The SourceConnectorContext and SourceTaskContext APIs
- The offsetStorageReader() method
- The transactionContext() method
- Exactly-Once Support
- The exactlyOnceSupport() method
- The canDefineTransactionBoundaries() method
- The commitTransaction() methods
- The abortTransaction() methods
- Implementing Sink Connectors
- The SinkTask API
- The put() method
- The preCommit() method
- The flush() method
- The open() and close() methods
- The SinkTask API lifecycle
- The SinkTask API
- Sink Records
- The SinkConnectorContext and SinkTaskContext APIs
- The offset() methods
- The timeout() method
- The assignment() method
- The pause() and resume() methods
- The requestCommit() method
- The errantRecordReporter() method
- Summary
- 12. Extending Kafka Connect with Connector and Worker Plug-Ins
- Implementing Connector Plug-Ins
- The Transformation API
- The apply() method
- The config() method
- The configure() method
- The close() method
- The Transformation API
- The Predicate API
- The test() method
- The config() method
- The configure() method
- The close() method
- Implementing Connector Plug-Ins
- The Converter and HeaderConverter APIs
- The fromConnectData() methods
- The toConnectData() methods
- The fromConnectHeader() method
- The toConnectHeader() method
- The config() methods
- The configure() methods
- The close() method
- Implementing Worker Plug-Ins
- The ConfigProvider API
- The get() methods
- The configure() method
- The close() method
- The subscribe(), unsubscribe(), and unsubscribeAll() methods
- The ConfigProvider API
- The ConnectorClientConfigOverridePolicy API
- The validate() method
- The configure() method
- The close() method
- The ConnectRestExtension APIs
- The register() method
- The configure() method
- The close() method
- The version() method
- Summary
- Index
O'Reilly Media - inne książki
-
JavaScript gives web developers great power to create rich interactive browser experiences, and much of that power is provided by the browser itself. Modern web APIs enable web-based applications to come to life like never before, supporting actions that once required browser plug-ins. Some are s...(186.15 zł najniższa cena z 30 dni)
186.15 zł
219.00 zł(-15%) -
How will software development and operations have to change to meet the sustainability and green needs of the planet? And what does that imply for development organizations? In this eye-opening book, sustainable software advocates Anne Currie, Sarah Hsu, and Sara Bergman provide a unique overview...(160.65 zł najniższa cena z 30 dni)
169.14 zł
199.00 zł(-15%) -
OpenTelemetry is a revolution in observability data. Instead of running multiple uncoordinated pipelines, OpenTelemetry provides users with a single integrated stream of data, providing multiple sources of high-quality telemetry data: tracing, metrics, logs, RUM, eBPF, and more. This practical gu...(143.65 zł najniższa cena z 30 dni)
143.65 zł
169.00 zł(-15%) -
What will you learn from this book?If you're a software developer looking for a quick on-ramp to software architecture, this handy guide is a great place to start. From the authors of Fundamentals of Software Architecture, Head First Software Architecture teaches you how to think architecturally ...(245.65 zł najniższa cena z 30 dni)
245.65 zł
289.00 zł(-15%) -
If you use Linux in your day-to-day work, then Linux Pocket Guide is the perfect on-the-job reference. This thoroughly updated 20th anniversary edition explains more than 200 Linux commands, including new commands for file handling, package management, version control, file format conversions, an...(92.65 zł najniższa cena z 30 dni)
101.15 zł
119.00 zł(-15%) -
Interested in developing embedded systems? Since they don't tolerate inefficiency, these systems require a disciplined approach to programming. This easy-to-read guide helps you cultivate good development practices based on classic software design patterns and new patterns unique to embedded prog...(152.15 zł najniższa cena z 30 dni)
160.65 zł
189.00 zł(-15%) -
Gain the valuable skills and techniques you need to accelerate the delivery of machine learning solutions. With this practical guide, data scientists, ML engineers, and their leaders will learn how to bridge the gap between data science and Lean product delivery in a practical and simple way. Dav...(245.65 zł najniższa cena z 30 dni)
245.65 zł
289.00 zł(-15%) -
This practical book provides a detailed explanation of the zero trust security model. Zero trust is a security paradigm shift that eliminates the concept of traditional perimeter-based security and requires you to "always assume breach" and "never trust but always verify." The updated edition off...(203.15 zł najniższa cena z 30 dni)
211.65 zł
249.00 zł(-15%) -
Decentralized finance (DeFi) is a rapidly growing field in fintech, having grown from $700 million to $100 billion over the past three years alone. But the lack of reliable information makes this area both risky and murky. In this practical book, experienced securities attorney Alexandra Damsker ...(203.15 zł najniższa cena z 30 dni)
211.65 zł
249.00 zł(-15%) -
Whether you're a startup founder trying to disrupt an industry or an entrepreneur trying to provoke change from within, your biggest challenge is creating a product people actually want. Lean Analytics steers you in the right direction.This book shows you how to validate your initial idea, find t...(126.65 zł najniższa cena z 30 dni)
126.65 zł
149.00 zł(-15%)
Dzieki opcji "Druk na żądanie" do sprzedaży wracają tytuły Grupy Helion, które cieszyły sie dużym zainteresowaniem, a których nakład został wyprzedany.
Dla naszych Czytelników wydrukowaliśmy dodatkową pulę egzemplarzy w technice druku cyfrowego.
Co powinieneś wiedzieć o usłudze "Druk na żądanie":
- usługa obejmuje tylko widoczną poniżej listę tytułów, którą na bieżąco aktualizujemy;
- cena książki może być wyższa od początkowej ceny detalicznej, co jest spowodowane kosztami druku cyfrowego (wyższymi niż koszty tradycyjnego druku offsetowego). Obowiązująca cena jest zawsze podawana na stronie WWW książki;
- zawartość książki wraz z dodatkami (płyta CD, DVD) odpowiada jej pierwotnemu wydaniu i jest w pełni komplementarna;
- usługa nie obejmuje książek w kolorze.
Masz pytanie o konkretny tytuł? Napisz do nas: sklep[at]helion.pl.
Książka, którą chcesz zamówić pochodzi z końcówki nakładu. Oznacza to, że mogą się pojawić drobne defekty (otarcia, rysy, zagięcia).
Co powinieneś wiedzieć o usłudze "Końcówka nakładu":
- usługa obejmuje tylko książki oznaczone tagiem "Końcówka nakładu";
- wady o których mowa powyżej nie podlegają reklamacji;
Masz pytanie o konkretny tytuł? Napisz do nas: sklep[at]helion.pl.
Książka drukowana
Oceny i opinie klientów: Kafka Connect Mickael Maison, Kate Stanley (0) Weryfikacja opinii następuję na podstawie historii zamówień na koncie Użytkownika umieszczającego opinię. Użytkownik mógł otrzymać punkty za opublikowanie opinii uprawniające do uzyskania rabatu w ramach Programu Punktowego.