Apache Kafka connection
To access your data with Apache Kafka, create a connection asset for it.
Apache Kafka is a distributed event streaming platform. Connect to an Apache Kafka real-time processing server to write and to read streams of events from and into topics.
Supported versions
Apache Kafka versions 0.11 - 2.x
Create a connection to Apache Kafka
To create the connection asset, you need these connection details.
Kafka server hostname: Hostname and port number for the Kafka server. Use this format: hostname:port-number
. To connect to a Kafka cluster, separate the values with commas: hostname1:port-number1,hostname2:port-number2,hostname3:port-number3
.
If you connect to a cluster, the connection uses all the servers irrespective of which servers are specified for bootstrapping. Because these servers are used for the initial connection to discover the full cluster membership, which can change
dynamically, this list does not need to contain the full set of servers. But if the Kafka cluster has three hosts or fewer, include all the hosts in this list to prevent data loss.
Secure connection
Select the network authentication protocol that is set up on the Kafka server. None is the default. These are the other selections and their properties:
SASL OAUTHBEARER
Select OAuth 2.0
- Server URL: Token endpoint URL. Example:
https://example.com/oauth2/default/v1/token
. - Client ID: The OAuth client ID.
- Client secret: The OAuth client secret
- Scope Optional scope to reference in the call to the OAuth server.
- Additional properties: Additional OAuth configuration options. Enter as
key=value
pairs. The value of this multiline property must conform to Java Properties class requirements. - Truststore certificates: Trusted certificates in PEM format. Only X.509 certificates are supported.
SASL_Plain
User principal name: The authenticated user in the Kafka server or cluster.
Password: Password for the user principal name.
SASL_SSL
User principal name: The authenticated user in the Kafka server or cluster.
Password: Password for the user principal name.
Truststore certificates: Truststore certificates in PEM format. Only X.509 certificates are supported.
SCRAM-SHA-256 or SCRAM-SHA-512
User principal name: The authenticated user in the Kafka server or cluster.
Password: Password for the user principal name.
Truststore certificates: Truststore certificates in PEM format. Only X.509 certificates are supported.
SSL
Truststore certificates: Truststore certificates in PEM format. Only X.509 certificates are supported.
Key: Private key in PEM format. The key must use PKCS #8 syntax.
Key certificates chain: Certificate chain for the private key in PEM format. Only X.509 certificates are supported.
Key password: This value is required if the key is encrypted.
Message format
A schema registry is third-party software that manages the messages and maps the schemas to topics so that producers know which topics are accepting which types (schemas) of messages and consumers know how to read and parse messages in a topic. If you select Use Schema Registry for message format, you can select these additional details to securely connect to the schema registry service.
Prerequisite
Set up the schema registry for your Kafka server with Confluent (Confluent versions 6.x and 7.x are supported) or with IBM Event Streams.
Schema Registry URL: URL to the schema registry service.
Authentication
Select the authentication method to the schema registry service. None is the default. These are the other selections and their properties:
-
Use Kafka server SASL user credentials: You can choose this selection if you entered properties for SASL_Plain or SASL_SSL for the secure connection to the Kafka server. The username and password for the SASL security settings will be used for authentication to schema registry service.
-
User credentials: Username and password to the schema registry service.
Secure connection
Select the secure network authentication protocol to the schema registry service. None is the default. These are the other selections and their properties:
-
Use Kafka server SSL user credentials: You can choose this selection if you entered properties for SSL for the secure connection to the Kafka server. The certificates configuration from the Kafka server connection will be used for the secure connection to schema registry service.
-
SSL
- Truststore certificates: Truststore certificates in PEM format. Only X.509 certificates are supported.
- Key: Private key in PEM format. The key must use PKCS #8 syntax.
- Key certificates chain: Certificate chain for the private key in PEM format. Only X.509 certificates are supported.
- Key password: This value is required if the key is encrypted.
Schema Registry type
Select the schema registry type:
- Confluent
- IBM Event Streams (Confluent-compatible API)
Choose the method for creating a connection based on where you are in the platform
In a project
Click Assets > New asset > Connect to a data source. See Adding a connection to a project.
In a catalog
Click Add to catalog > Connection. See Adding a connection asset to a catalog.
In the Platform assets catalog
Click New connection. See Adding platform connections.
Next step: Add data assets from the connection
Where you can use this connection
You can use the Apache Kafka connection in the following workspaces and tools:
Projects
- Data Replication (Data Replication service). You can replicate data from other databases to Apache Kafka using Data Replication. See Replicating Apache Kafka data.
- DataStage (DataStage service). See Connecting to a data source in DataStage.
Catalogs
-
Platform assets catalog
-
Other catalogs (IBM Knowledge Catalog)
Note:Preview, profile, and masking are not certified for this connection in IBM Knowledge Catalog.
Apache Kafka setup
Learn more
Parent topic: Supported connections