Kafka Connect – Externalizing Secrets – KIP 297 In order to connect with a data source or sink, we need to use credentials. Kafka Connect added support for specifying credential using config providers. The support for file config provider is available with the installation package. This is discussed in KIP 297. The KIP was released… Continue reading »
Securing JMX on Confluent Kafka Confluent kafka process start with these default arguments. You can see that JMX authentication is disabled by default. This is a security vulnerability and might lead to possible issues. Confluent kafka consists of the following services. We need to enable authentication for all of these services for JMX: Kafka Broker… Continue reading »
Confluent Hub Client Confluent Hub is an online repository for extensions and components for Kafka. Kafka is based on extensible model for many of its services. It allows plug-ins and extensions which makes it generic enough to be suitable for many real world streaming based applications. They include both Confluent and 3rd party components. Generally,… Continue reading »
Kafka Connect – Single Message Transforms (SMTs) Single Message Transforms were released with 0.10.2 release [ Release notes ]. It provides us the ability to transform a message before they get in or out of a connector using Kafka Connect. Source Transforms Source connector can be configured with a list of transforms. These transforms are… Continue reading »
Custom Log4J Configuration file for Kafka In order to use a custom Log4J configuration file and to override the default one, we need to add the details in -Dlog4j.configuration. Here is how you can update the configuration path: As you can notice, we have also update the directory where logs are generated by overriding LOG_DIR.
Deleting Connectors from Kafka Connect A connector can be deleting using Kafka Connect’s REST API. Here is the curl command:
Kafka Connect Issues: Found null value for non-optional schema In a custom connector, you might encounter the following issue: Here we have a schema defined for one of our custom connector: Now you would see the error when you create a SourceRecord with a value missing the non-optional fields.
Confluent – Using off-the-shelf-connectors with Landoop’s Kafka Connect UI In this post we are going to see how easy it is to run a Kafka Connector in order to source data in Kafka using a well designed tool like Landoop’s Kafka Connect UI. We can use Kafka Connect’s REST API to determine what connectors are… Continue reading »
Setting up Landoop’s Kafka-Connect-ui on Mac In this post, we are going to setup the third of the series of tools by Landoop to manage Confluent services for Kafka. Landoop provides Kafka-connect-ui to manage connectors in Kafka-connect. The tools is available as a Docker image on DockerHub, so let’s first pull it on our machine…. Continue reading »
Using Landoop’s Schema Registry UI In this post, we are going to setup Schema Registry UI for Confluent’s schema registry using Docker Image for the tool. Schema Registry is an amazing tool by Landoop. It’s available in Github. Let’s first pull the image from Docker Hub. Now we can simply run the docker image by… Continue reading »