Connect Self-Managed REST Proxy to Confluent Cloud

You can configure a local Confluent REST Proxy to produce to and consume from an Apache Kafka® topic in a Kafka cluster in Confluent Cloud.

See also

To see a hands-on example that uses Confluent REST Proxy to produce and consume data from a Kafka cluster, see the Confluent REST Proxy tutorial.

To connect REST Proxy to Confluent Cloud, you must download the Confluent Platform tarball and then start REST Proxy by using a customized properties file.

Prerequisites

Procedure

  1. Download Confluent Platform and extract the contents.

  2. Create a topic named rest-proxy-test by using the Confluent CLI:

    confluent kafka topic create --partitions 4 rest-proxy-test 
  3. Create a properties file.

    1. Find the client settings for your cluster by clicking CLI & client configuration from the Cloud Console interface.

    2. Click the Clients tab.

    3. Click the Java client selection. This example uses the Java client.

      Java client configuration properties

      Java client configuration properties

    4. Create a properties file named ccloud-kafka-rest.properties where the Confluent Platform files are location.

      cd <path-to-confluent> 
      touch ccloud-kafka-rest.properties 
    5. Copy and paste the Java client configuration properties into the file. Add the client. prefix to each of security properties. For example:

      # Kafka bootstrap.servers=<myproject>.cloud:9092 security.protocol=SASL_SSL sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule \ required username="<kafka-cluster-api-key>" password="<kafka-cluster-api-secret>"; ssl.endpoint.identification.algorithm=https sasl.mechanism=PLAIN client.bootstrap.servers=<myproject>.cloud:9092 client.security.protocol=SASL_SSL client.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule \ required username="<kafka-cluster-api-key>" password="<kafka-cluster-api-secret>"; client.ssl.endpoint.identification.algorithm=https client.sasl.mechanism=PLAIN # Confluent Cloud Schema Registry schema.registry.url=<schema-registry-endpoint> client.basic.auth.credentials.source=USER_INFO client.schema.registry.basic.auth.user.info=<schema-registry-api-key>:<schema-registry-api-secret> 

    Producers, consumers, and the admin client share the client. properties. Refer to the following table to specify additional properties for the producer, consumer, or admin client.

    Component Prefix Example
    Admin Client admin. admin.request.timeout.ms
    Consumer consumer. consumer.request.timeout.ms
    Producer producer. producer.acks

    An example of adding these properties is shown below:

    # Kafka bootstrap.servers=<myproject>.cloud:9092 security.protocol=SASL_SSL client.security.protocol=SASL_SSL client.sasl.jaas.config=org.apache.kafka.common.security.plain.PlainLoginModule required username="<kafka-cluster-api-key>" password="<kafka-cluster-api-secret>"; client.ssl.endpoint.identification.algorithm=https sasl.mechanism=PLAIN client.sasl.mechanism=PLAIN # Confluent Cloud Schema Registry schema.registry.url=<schema-registry-endpoint> client.basic.auth.credentials.source=USER_INFO client.schema.registry.basic.auth.user.info=<schema-registry-api-key>:<schema-registry-api-secret> # consumer only properties must be prefixed with consumer. consumer.retry.backoff.ms=600 consumer.request.timeout.ms=25000 # producer only properties must be prefixed with producer. producer.acks=1 # admin client only properties must be prefixed with admin. admin.request.timeout.ms=50000 

    For details about how to create a Confluent Cloud API key and API secret so that you can communicate with the REST API, refer to Create credentials to access the Kafka cluster resources.

  4. Start the REST Proxy.

    ./bin/kafka-rest-start ccloud-kafka-rest.properties 
  5. Make REST calls using REST API v2.

    Example request:

    GET /topics/test HTTP/1.1 Accept: application/vnd.kafka.v2+json 

    Caution

    Do not use API v1. API v1 has a ZooKeeper dependency that does not work in Confluent Cloud.

Docker environment

You can run a mix of fully-managed services in Confluent Cloud and self-managed components running in Docker. For a Docker environment that connects any Confluent Platform component to Confluent Cloud, see cp-all-in-one-cloud.