Confluent Kafka assets package¶
The Confluent Kafka assets package crawls Confluent Kafka assets and publishes them to Atlan for discovery.
Direct extraction¶
Will create a new connection
This should only be used to create the workflow the first time. Each time you run this method it will create a new connection and new assets within that connection — which could lead to duplicate assets if you run the workflow this way multiple times with the same settings.
Instead, when you want to re-crawl assets, re-run the existing workflow (see Re-run existing workflow below).
To crawl assets directly from Confluent Kafka:
Direct extraction from Confluent Kafka | |
---|---|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 |
|
- The
ConfluentKafkaCrawler
package will create a workflow to crawl assets from Confluent Kafka. - You must provide Atlan client.
- You must provide a name for the connection that the Confluent Kafka assets will exist within.
-
You must specify at least one connection admin, either:
- everyone in a role (in this example, all
$admin
users). - a list of groups (names) that will be connection admins.
- a list of users (names) that will be connection admins.
- everyone in a role (in this example, all
-
When crawling assets directly from Confluent Kafka, you are required to provide the following information:
- hostname and port number (host.example.com:9092) for the Kafka bootstrap server.
- whether to use an encrypted SSL connection (
true
), or plaintext (false
)
-
You must provide an API key through which to access Kafka.
- You must provide API secret through which to access Kafka.
- You can also optionally set whether to skip internal
topics when crawling (
true
) or include them (false
). - You can also optionally provide the regular expression to use for including topics when crawling.
- You can also optionally provide the regular expression to use for excluding topics when crawling.
- Build the minimal package object.
- Now, you can convert the package into a
Workflow
object. -
You can then run the workflow using the
run()
method on the object you've created.Workflows run asynchronously
Remember that workflows run asynchronously. See the packages and workflows introduction for details on how you can check the status and wait until the workflow has been completed.
Direct extraction from Confluent Kafka | |
---|---|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 |
|
- Base configuration for a new Confluent Kafka crawler.
- You must provide a name for the connection that the Confluent Kafka assets will exist within.
-
You must specify at least one connection admin, either:
- everyone in a role (in this example, all
$admin
users). - a list of groups (names) that will be connection admins.
- a list of users (names) that will be connection admins.
- everyone in a role (in this example, all
-
When crawling assets directly from Confluent Kafka, you are required to provide the following information:
- hostname and port number (host.example.com:9092) for the Kafka bootstrap server.
- whether to use an encrypted SSL connection (
True
), or plaintext (False
)
-
You must provide an API key through which to access Kafka.
- You must provide API secret through which to access Kafka.
- You can also optionally set whether to skip internal
topics when crawling (
True
) or include them (False
). - You can also optionally provide the regular expression to use for including topics when crawling. (If set to None, all topics will be crawled.)
- You can also optionally provide the regular expression to use for excluding topics when crawling. (If set to None, no topics will be excluded.)
- Now, you can convert the package into a
Workflow
object. -
Run the workflow by invoking the
run()
method on the workflow client, passing the created object.Workflows run asynchronously
Remember that workflows run asynchronously. See the packages and workflows introduction for details on how you can check the status and wait until the workflow has been completed.
Direct extraction from Confluent Kafka | |
---|---|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 |
|
- The
ConfluentKafkaCrawler
package will create a workflow to crawl assets from Confluent Kafka. - You must provide Atlan client.
- You must provide a name for the connection that the Confluent Kafka assets will exist within.
-
You must specify at least one connection admin, either:
- everyone in a role (in this example, all
$admin
users). - a list of groups (names) that will be connection admins.
- a list of users (names) that will be connection admins.
- everyone in a role (in this example, all
-
When crawling assets directly from Confluent Kafka, you are required to provide the following information:
- hostname and port number (host.example.com:9092) for the Kafka bootstrap server.
- whether to use an encrypted SSL connection (
true
), or plaintext (false
)
-
You must provide an API key through which to access Kafka.
- You must provide API secret through which to access Kafka.
- You can also optionally set whether to skip internal
topics when crawling (
true
) or include them (false
). - You can also optionally provide the regular expression to use for including topics when crawling.
- You can also optionally provide the regular expression to use for excluding topics when crawling.
- Build the minimal package object.
- Now, you can convert the package into a
Workflow
object. -
You can then run the workflow using the
run()
method on the object you've created.Workflows run asynchronously
Remember that workflows run asynchronously. See the packages and workflows introduction for details on how you can check the status and wait until the workflow has been completed.
Create the workflow via UI only
We recommend creating the workflow only via the UI. To rerun an existing workflow, see the steps below.
Re-run existing workflow¶
To re-run an existing workflow for Confluent Kafka assets:
Re-run existing Confluent Kafka workflow | |
---|---|
1 2 3 4 |
|
- You can search for existing workflows through the
WorkflowSearchRequest
class. - You can find workflows by their type using the
findByType()
helper method and providing the prefix for one of the packages. In this example, we do so for theConfluentKafkaCrawler
. (You can also specify the maximum number of resulting workflows you want to retrieve as results.) -
Once you've found the workflow you want to re-run, you can simply call the
rerun()
helper method on the workflow search result. TheWorkflowRunResponse
is just a subtype ofWorkflowResponse
so has the same helper method to monitor progress of the workflow run.- Optionally, you can use the
rerun(true)
method with idempotency to avoid re-running a workflow that is already in running or in a pending state. This will return details of the already running workflow if found, and by default, it is set tofalse
Workflows run asynchronously
Remember that workflows run asynchronously. See the packages and workflows introduction for details on how you can check the status and wait until the workflow has been completed.
- Optionally, you can use the
Re-run existing Confluent Kafka workflow | |
---|---|
1 2 3 4 5 6 7 8 9 10 11 12 |
|
- You can find workflows by their type using the workflow client
find_by_type()
method and providing the prefix for one of the packages. In this example, we do so for theConfluentKafkaCrawler
. (You can also specify the maximum number of resulting workflows you want to retrieve as results.) -
Once you've found the workflow you want to re-run, you can simply call the workflow client
rerun()
method.- Optionally, you can use
rerun(idempotent=True)
to avoid re-running a workflow that is already in running or in a pending state. This will return details of the already running workflow if found, and by default, it is set toFalse
.
Workflows run asynchronously
Remember that workflows run asynchronously. See the packages and workflows introduction for details on how you can check the status and wait until the workflow has been completed.
- Optionally, you can use
Re-run existing Confluent Kafka workflow | |
---|---|
1 2 3 4 5 |
|
- You can search for existing workflows through the
WorkflowSearchRequest
class. - You can find workflows by their type using the
findByType()
helper method and providing the prefix for one of the packages. In this example, we do so for theConfluentKafkaCrawler
. (You can also specify the maximum number of resulting workflows you want to retrieve as results.) -
Once you've found the workflow you want to re-run, you can simply call the
rerun()
helper method on the workflow search result. TheWorkflowRunResponse
is just a subtype ofWorkflowResponse
so has the same helper method to monitor progress of the workflow run.- Optionally, you can use the
rerun(true)
method with idempotency to avoid re-running a workflow that is already in running or in a pending state. This will return details of the already running workflow if found, and by default, it is set tofalse
Workflows run asynchronously
Remember that workflows run asynchronously. See the packages and workflows introduction for details on how you can check the status and wait until the workflow has been completed.
- Optionally, you can use the
Requires multiple steps through the raw REST API
- Find the existing workflow.
- Send through the resulting re-run request.
POST /api/service/workflows/indexsearch | |
---|---|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
|
-
Searching by the
atlan-kafka-confluent-cloud
prefix will ensure you only find existing Confluent Kafka assets workflows.Name of the workflow
The name of the workflow will be nested within the
_source.metadata.name
property of the response object. (Remember since this is a search, there could be multiple results, so you may want to use the other details in each result to determine which workflow you really want.)
POST /api/service/workflows/submit | |
---|---|
100 101 102 103 104 |
|
- Send the name of the workflow as the
resourceName
to rerun it.