camel-spark-kafka-connector sink configuration
Connector Description: Send RDD or DataFrame jobs to Apache Spark clusters.
When using camel-spark-kafka-connector as sink make sure to use the following Maven dependency to have support for the connector:
<dependency>
<groupId>org.apache.camel.kafkaconnector</groupId>
<artifactId>camel-spark-kafka-connector</artifactId>
<version>x.x.x</version>
<!-- use the same version as your Camel Kafka connector version -->
</dependency>
To use this sink connector in Kafka connect you’ll need to set the following connector.class
connector.class=org.apache.camel.kafkaconnector.spark.CamelSparkSinkConnector
The camel-spark sink connector supports 11 options, which are listed below.
Name | Description | Default | Priority |
---|---|---|---|
Required Type of the endpoint (rdd, dataframe, hive). One of: [rdd] [dataframe] [hive]. Enum values:
|
HIGH |
||
Indicates if results should be collected or counted. |
true |
MEDIUM |
|
DataFrame to compute against. |
MEDIUM |
||
Function performing action against an DataFrame. |
MEDIUM |
||
Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel’s routing error handlers. Beware that when the first message is processed then creating and starting the producer may take a little time and prolong the total processing time of the processing. |
false |
MEDIUM |
|
RDD to compute against. |
MEDIUM |
||
Function performing action against an RDD. |
MEDIUM |
||
Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel’s routing error handlers. Beware that when the first message is processed then creating and starting the producer may take a little time and prolong the total processing time of the processing. |
false |
MEDIUM |
|
RDD to compute against. |
MEDIUM |
||
Function performing action against an RDD. |
MEDIUM |
||
Whether autowiring is enabled. This is used for automatic autowiring options (the option must be marked as autowired) by looking up in the registry to find if there is a single instance of matching type, which then gets configured on the component. This can be used for automatic configuring JDBC data sources, JMS connection factories, AWS Clients, etc. |
true |
MEDIUM |
The camel-spark sink connector has no converters out of the box.
The camel-spark sink connector has no transforms out of the box.
The camel-spark sink connector has no aggregation strategies out of the box.