This is an automated email from the ASF dual-hosted git repository. acosentino pushed a commit to branch main in repository https://gitbox.apache.org/repos/asf/camel-kafka-connector.git
The following commit(s) were added to refs/heads/main by this push: new 2c668e0 [create-pull-request] automated change 2c668e0 is described below commit 2c668e0c049d26ba466049511762428041498dcc Author: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com> AuthorDate: Thu Apr 29 03:29:52 2021 +0000 [create-pull-request] automated change --- .../resources/connectors/camel-atlasmap-sink.json | 12 ++++++------ .../resources/connectors/camel-aws2-ddb-sink.json | 14 ++++++++++++++ .../generated/resources/connectors/camel-jt400-source.json | 7 +++++++ .../src/generated/resources/camel-atlasmap-sink.json | 12 ++++++------ .../src/main/docs/camel-atlasmap-kafka-sink-connector.adoc | 2 +- .../atlasmap/CamelAtlasmapSinkConnectorConfig.java | 8 ++++---- .../src/generated/resources/camel-aws2-ddb-sink.json | 14 ++++++++++++++ .../src/main/docs/camel-aws2-ddb-kafka-sink-connector.adoc | 4 +++- .../aws2ddb/CamelAws2ddbSinkConnectorConfig.java | 8 ++++++++ .../src/generated/resources/camel-jt400-source.json | 7 +++++++ .../src/main/docs/camel-jt400-kafka-source-connector.adoc | 3 ++- .../jt400/CamelJt400SourceConnectorConfig.java | 4 ++++ .../connectors/camel-atlasmap-kafka-sink-connector.adoc | 2 +- .../connectors/camel-aws2-ddb-kafka-sink-connector.adoc | 4 +++- .../connectors/camel-jt400-kafka-source-connector.adoc | 3 ++- 15 files changed, 82 insertions(+), 22 deletions(-) diff --git a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-atlasmap-sink.json b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-atlasmap-sink.json index 293ae3d..ec1e6bc 100644 --- a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-atlasmap-sink.json +++ b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-atlasmap-sink.json @@ -60,12 +60,6 @@ "priority": "MEDIUM", "required": "false" }, - "camel.sink.endpoint.propertiesFile": { - "name": "camel.sink.endpoint.propertiesFile", - "description": "The URI of the properties file which is used for AtlasContextFactory initialization.", - "priority": "MEDIUM", - "required": "false" - }, "camel.component.atlasmap.lazyStartProducer": { "name": "camel.component.atlasmap.lazyStartProducer", "description": "Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel's routing error handlers. Beware that when the first message is processed then creating and starting the pr [...] @@ -85,6 +79,12 @@ "defaultValue": "true", "priority": "MEDIUM", "required": "false" + }, + "camel.component.atlasmap.propertiesFile": { + "name": "camel.component.atlasmap.propertiesFile", + "description": "The URI of the properties file which is used for AtlasContextFactory initialization.", + "priority": "MEDIUM", + "required": "false" } } } \ No newline at end of file diff --git a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-ddb-sink.json b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-ddb-sink.json index f89a8ae..b64ecb0 100644 --- a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-ddb-sink.json +++ b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-aws2-ddb-sink.json @@ -127,6 +127,13 @@ "priority": "MEDIUM", "required": "false" }, + "camel.sink.endpoint.useDefaultCredentialsProvider": { + "name": "camel.sink.endpoint.useDefaultCredentialsProvider", + "description": "Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in.", + "defaultValue": "false", + "priority": "MEDIUM", + "required": "false" + }, "camel.sink.endpoint.writeCapacity": { "name": "camel.sink.endpoint.writeCapacity", "description": "The provisioned throughput to reserved for writing resources to your table", @@ -263,6 +270,13 @@ "priority": "MEDIUM", "required": "false" }, + "camel.component.aws2-ddb.useDefaultCredentialsProvider": { + "name": "camel.component.aws2-ddb.useDefaultCredentialsProvider", + "description": "Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in.", + "defaultValue": "false", + "priority": "MEDIUM", + "required": "false" + }, "camel.component.aws2-ddb.writeCapacity": { "name": "camel.component.aws2-ddb.writeCapacity", "description": "The provisioned throughput to reserved for writing resources to your table", diff --git a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-jt400-source.json b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-jt400-source.json index 56a31a2..21d384c 100644 --- a/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-jt400-source.json +++ b/camel-kafka-connector-catalog/src/generated/resources/connectors/camel-jt400-source.json @@ -130,6 +130,13 @@ "priority": "MEDIUM", "required": "false" }, + "camel.source.endpoint.sendingReply": { + "name": "camel.source.endpoint.sendingReply", + "description": "If true, the consumer endpoint will set the Jt400Constants.MESSAGE_REPLYTO_KEY header of the camel message for any IBM i inquiry messages received. If that message is then routed to a producer endpoint, the action will not be processed as sending a message to the queue, but rather a reply to the specific inquiry message.", + "defaultValue": "true", + "priority": "MEDIUM", + "required": "false" + }, "camel.source.endpoint.exceptionHandler": { "name": "camel.source.endpoint.exceptionHandler", "description": "To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored.", diff --git a/connectors/camel-atlasmap-kafka-connector/src/generated/resources/camel-atlasmap-sink.json b/connectors/camel-atlasmap-kafka-connector/src/generated/resources/camel-atlasmap-sink.json index 293ae3d..ec1e6bc 100644 --- a/connectors/camel-atlasmap-kafka-connector/src/generated/resources/camel-atlasmap-sink.json +++ b/connectors/camel-atlasmap-kafka-connector/src/generated/resources/camel-atlasmap-sink.json @@ -60,12 +60,6 @@ "priority": "MEDIUM", "required": "false" }, - "camel.sink.endpoint.propertiesFile": { - "name": "camel.sink.endpoint.propertiesFile", - "description": "The URI of the properties file which is used for AtlasContextFactory initialization.", - "priority": "MEDIUM", - "required": "false" - }, "camel.component.atlasmap.lazyStartProducer": { "name": "camel.component.atlasmap.lazyStartProducer", "description": "Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel's routing error handlers. Beware that when the first message is processed then creating and starting the pr [...] @@ -85,6 +79,12 @@ "defaultValue": "true", "priority": "MEDIUM", "required": "false" + }, + "camel.component.atlasmap.propertiesFile": { + "name": "camel.component.atlasmap.propertiesFile", + "description": "The URI of the properties file which is used for AtlasContextFactory initialization.", + "priority": "MEDIUM", + "required": "false" } } } \ No newline at end of file diff --git a/connectors/camel-atlasmap-kafka-connector/src/main/docs/camel-atlasmap-kafka-sink-connector.adoc b/connectors/camel-atlasmap-kafka-connector/src/main/docs/camel-atlasmap-kafka-sink-connector.adoc index 0220636..45c966b 100644 --- a/connectors/camel-atlasmap-kafka-connector/src/main/docs/camel-atlasmap-kafka-sink-connector.adoc +++ b/connectors/camel-atlasmap-kafka-connector/src/main/docs/camel-atlasmap-kafka-sink-connector.adoc @@ -38,10 +38,10 @@ The camel-atlasmap sink connector supports 11 options, which are listed below. | *camel.sink.endpoint.sourceMapName* | The Exchange property name for a source message map which hold java.util.Map<String, Message> where the key is AtlasMap Document ID. AtlasMap consumes Message bodies as source documents, as well as message headers as source properties where the scope equals to Document ID. | null | false | MEDIUM | *camel.sink.endpoint.targetMapMode* | TargetMapMode enum value to specify how multiple target documents are delivered if exist. 'MAP': Stores them into a java.util.Map, and the java.util.Map is set to an exchange property if 'targetMapName' is specified, otherwise message body. 'MESSAGE_HEADER': Stores them into message headers. 'EXCHANGE_PROPERTY': Stores them into exchange properties. ) One of: [MAP] [MESSAGE_HEADER] [EXCHANGE_PROPERTY] | "MAP" | false | MEDIUM | *camel.sink.endpoint.targetMapName* | The Exchange property name for a target document map which hold java.util.Map<String, Object> where the key is AtlasMap Document ID. AtlasMap populates multiple target documents into this map. | null | false | MEDIUM -| *camel.sink.endpoint.propertiesFile* | The URI of the properties file which is used for AtlasContextFactory initialization. | null | false | MEDIUM | *camel.component.atlasmap.lazyStartProducer* | Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel's routing error handlers. Beware that when the first message is processed the [...] | *camel.component.atlasmap.atlasContextFactory* | To use the AtlasContextFactory otherwise a new engine is created. | null | false | MEDIUM | *camel.component.atlasmap.autowiredEnabled* | Whether autowiring is enabled. This is used for automatic autowiring options (the option must be marked as autowired) by looking up in the registry to find if there is a single instance of matching type, which then gets configured on the component. This can be used for automatic configuring JDBC data sources, JMS connection factories, AWS Clients, etc. | true | false | MEDIUM +| *camel.component.atlasmap.propertiesFile* | The URI of the properties file which is used for AtlasContextFactory initialization. | null | false | MEDIUM |=== diff --git a/connectors/camel-atlasmap-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/atlasmap/CamelAtlasmapSinkConnectorConfig.java b/connectors/camel-atlasmap-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/atlasmap/CamelAtlasmapSinkConnectorConfig.java index 81504d1..c76b50d 100644 --- a/connectors/camel-atlasmap-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/atlasmap/CamelAtlasmapSinkConnectorConfig.java +++ b/connectors/camel-atlasmap-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/atlasmap/CamelAtlasmapSinkConnectorConfig.java @@ -47,9 +47,6 @@ public class CamelAtlasmapSinkConnectorConfig public static final String CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_NAME_CONF = "camel.sink.endpoint.targetMapName"; public static final String CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_NAME_DOC = "The Exchange property name for a target document map which hold java.util.Map<String, Object> where the key is AtlasMap Document ID. AtlasMap populates multiple target documents into this map."; public static final String CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_NAME_DEFAULT = null; - public static final String CAMEL_SINK_ATLASMAP_ENDPOINT_PROPERTIES_FILE_CONF = "camel.sink.endpoint.propertiesFile"; - public static final String CAMEL_SINK_ATLASMAP_ENDPOINT_PROPERTIES_FILE_DOC = "The URI of the properties file which is used for AtlasContextFactory initialization."; - public static final String CAMEL_SINK_ATLASMAP_ENDPOINT_PROPERTIES_FILE_DEFAULT = null; public static final String CAMEL_SINK_ATLASMAP_COMPONENT_LAZY_START_PRODUCER_CONF = "camel.component.atlasmap.lazyStartProducer"; public static final String CAMEL_SINK_ATLASMAP_COMPONENT_LAZY_START_PRODUCER_DOC = "Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel's routing error handlers. Beware that [...] public static final Boolean CAMEL_SINK_ATLASMAP_COMPONENT_LAZY_START_PRODUCER_DEFAULT = false; @@ -59,6 +56,9 @@ public class CamelAtlasmapSinkConnectorConfig public static final String CAMEL_SINK_ATLASMAP_COMPONENT_AUTOWIRED_ENABLED_CONF = "camel.component.atlasmap.autowiredEnabled"; public static final String CAMEL_SINK_ATLASMAP_COMPONENT_AUTOWIRED_ENABLED_DOC = "Whether autowiring is enabled. This is used for automatic autowiring options (the option must be marked as autowired) by looking up in the registry to find if there is a single instance of matching type, which then gets configured on the component. This can be used for automatic configuring JDBC data sources, JMS connection factories, AWS Clients, etc."; public static final Boolean CAMEL_SINK_ATLASMAP_COMPONENT_AUTOWIRED_ENABLED_DEFAULT = true; + public static final String CAMEL_SINK_ATLASMAP_COMPONENT_PROPERTIES_FILE_CONF = "camel.component.atlasmap.propertiesFile"; + public static final String CAMEL_SINK_ATLASMAP_COMPONENT_PROPERTIES_FILE_DOC = "The URI of the properties file which is used for AtlasContextFactory initialization."; + public static final String CAMEL_SINK_ATLASMAP_COMPONENT_PROPERTIES_FILE_DEFAULT = null; public CamelAtlasmapSinkConnectorConfig( ConfigDef config, @@ -79,10 +79,10 @@ public class CamelAtlasmapSinkConnectorConfig conf.define(CAMEL_SINK_ATLASMAP_ENDPOINT_SOURCE_MAP_NAME_CONF, ConfigDef.Type.STRING, CAMEL_SINK_ATLASMAP_ENDPOINT_SOURCE_MAP_NAME_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_ENDPOINT_SOURCE_MAP_NAME_DOC); conf.define(CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_MODE_CONF, ConfigDef.Type.STRING, CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_MODE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_MODE_DOC); conf.define(CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_NAME_CONF, ConfigDef.Type.STRING, CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_NAME_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_ENDPOINT_TARGET_MAP_NAME_DOC); - conf.define(CAMEL_SINK_ATLASMAP_ENDPOINT_PROPERTIES_FILE_CONF, ConfigDef.Type.STRING, CAMEL_SINK_ATLASMAP_ENDPOINT_PROPERTIES_FILE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_ENDPOINT_PROPERTIES_FILE_DOC); conf.define(CAMEL_SINK_ATLASMAP_COMPONENT_LAZY_START_PRODUCER_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_ATLASMAP_COMPONENT_LAZY_START_PRODUCER_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_COMPONENT_LAZY_START_PRODUCER_DOC); conf.define(CAMEL_SINK_ATLASMAP_COMPONENT_ATLAS_CONTEXT_FACTORY_CONF, ConfigDef.Type.STRING, CAMEL_SINK_ATLASMAP_COMPONENT_ATLAS_CONTEXT_FACTORY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_COMPONENT_ATLAS_CONTEXT_FACTORY_DOC); conf.define(CAMEL_SINK_ATLASMAP_COMPONENT_AUTOWIRED_ENABLED_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_ATLASMAP_COMPONENT_AUTOWIRED_ENABLED_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_COMPONENT_AUTOWIRED_ENABLED_DOC); + conf.define(CAMEL_SINK_ATLASMAP_COMPONENT_PROPERTIES_FILE_CONF, ConfigDef.Type.STRING, CAMEL_SINK_ATLASMAP_COMPONENT_PROPERTIES_FILE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_ATLASMAP_COMPONENT_PROPERTIES_FILE_DOC); return conf; } } \ No newline at end of file diff --git a/connectors/camel-aws2-ddb-kafka-connector/src/generated/resources/camel-aws2-ddb-sink.json b/connectors/camel-aws2-ddb-kafka-connector/src/generated/resources/camel-aws2-ddb-sink.json index f89a8ae..b64ecb0 100644 --- a/connectors/camel-aws2-ddb-kafka-connector/src/generated/resources/camel-aws2-ddb-sink.json +++ b/connectors/camel-aws2-ddb-kafka-connector/src/generated/resources/camel-aws2-ddb-sink.json @@ -127,6 +127,13 @@ "priority": "MEDIUM", "required": "false" }, + "camel.sink.endpoint.useDefaultCredentialsProvider": { + "name": "camel.sink.endpoint.useDefaultCredentialsProvider", + "description": "Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in.", + "defaultValue": "false", + "priority": "MEDIUM", + "required": "false" + }, "camel.sink.endpoint.writeCapacity": { "name": "camel.sink.endpoint.writeCapacity", "description": "The provisioned throughput to reserved for writing resources to your table", @@ -263,6 +270,13 @@ "priority": "MEDIUM", "required": "false" }, + "camel.component.aws2-ddb.useDefaultCredentialsProvider": { + "name": "camel.component.aws2-ddb.useDefaultCredentialsProvider", + "description": "Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in.", + "defaultValue": "false", + "priority": "MEDIUM", + "required": "false" + }, "camel.component.aws2-ddb.writeCapacity": { "name": "camel.component.aws2-ddb.writeCapacity", "description": "The provisioned throughput to reserved for writing resources to your table", diff --git a/connectors/camel-aws2-ddb-kafka-connector/src/main/docs/camel-aws2-ddb-kafka-sink-connector.adoc b/connectors/camel-aws2-ddb-kafka-connector/src/main/docs/camel-aws2-ddb-kafka-sink-connector.adoc index 7f8a826..3e73ea1 100644 --- a/connectors/camel-aws2-ddb-kafka-connector/src/main/docs/camel-aws2-ddb-kafka-sink-connector.adoc +++ b/connectors/camel-aws2-ddb-kafka-connector/src/main/docs/camel-aws2-ddb-kafka-sink-connector.adoc @@ -24,7 +24,7 @@ connector.class=org.apache.camel.kafkaconnector.aws2ddb.CamelAws2ddbSinkConnecto ---- -The camel-aws2-ddb sink connector supports 39 options, which are listed below. +The camel-aws2-ddb sink connector supports 41 options, which are listed below. @@ -47,6 +47,7 @@ The camel-aws2-ddb sink connector supports 39 options, which are listed below. | *camel.sink.endpoint.region* | The region in which DDB client needs to work | null | false | MEDIUM | *camel.sink.endpoint.trustAllCertificates* | If we want to trust all certificates in case of overriding the endpoint | false | false | MEDIUM | *camel.sink.endpoint.uriEndpointOverride* | Set the overriding uri endpoint. This option needs to be used in combination with overrideEndpoint option | null | false | MEDIUM +| *camel.sink.endpoint.useDefaultCredentialsProvider* | Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in. | false | false | MEDIUM | *camel.sink.endpoint.writeCapacity* | The provisioned throughput to reserved for writing resources to your table | null | false | MEDIUM | *camel.sink.endpoint.accessKey* | Amazon AWS Access Key | null | false | MEDIUM | *camel.sink.endpoint.secretKey* | Amazon AWS Secret Key | null | false | MEDIUM @@ -66,6 +67,7 @@ The camel-aws2-ddb sink connector supports 39 options, which are listed below. | *camel.component.aws2-ddb.region* | The region in which DDB client needs to work | null | false | MEDIUM | *camel.component.aws2-ddb.trustAllCertificates* | If we want to trust all certificates in case of overriding the endpoint | false | false | MEDIUM | *camel.component.aws2-ddb.uriEndpointOverride* | Set the overriding uri endpoint. This option needs to be used in combination with overrideEndpoint option | null | false | MEDIUM +| *camel.component.aws2-ddb.useDefaultCredentials Provider* | Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in. | false | false | MEDIUM | *camel.component.aws2-ddb.writeCapacity* | The provisioned throughput to reserved for writing resources to your table | null | false | MEDIUM | *camel.component.aws2-ddb.autowiredEnabled* | Whether autowiring is enabled. This is used for automatic autowiring options (the option must be marked as autowired) by looking up in the registry to find if there is a single instance of matching type, which then gets configured on the component. This can be used for automatic configuring JDBC data sources, JMS connection factories, AWS Clients, etc. | true | false | MEDIUM | *camel.component.aws2-ddb.accessKey* | Amazon AWS Access Key | null | false | MEDIUM diff --git a/connectors/camel-aws2-ddb-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2ddb/CamelAws2ddbSinkConnectorConfig.java b/connectors/camel-aws2-ddb-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2ddb/CamelAws2ddbSinkConnectorConfig.java index 46de65d..7c94eb6 100644 --- a/connectors/camel-aws2-ddb-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2ddb/CamelAws2ddbSinkConnectorConfig.java +++ b/connectors/camel-aws2-ddb-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/aws2ddb/CamelAws2ddbSinkConnectorConfig.java @@ -72,6 +72,9 @@ public class CamelAws2ddbSinkConnectorConfig extends CamelSinkConnectorConfig { public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_URI_ENDPOINT_OVERRIDE_CONF = "camel.sink.endpoint.uriEndpointOverride"; public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_URI_ENDPOINT_OVERRIDE_DOC = "Set the overriding uri endpoint. This option needs to be used in combination with overrideEndpoint option"; public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_URI_ENDPOINT_OVERRIDE_DEFAULT = null; + public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF = "camel.sink.endpoint.useDefaultCredentialsProvider"; + public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC = "Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in."; + public static final Boolean CAMEL_SINK_AWS2DDB_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT = false; public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_WRITE_CAPACITY_CONF = "camel.sink.endpoint.writeCapacity"; public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_WRITE_CAPACITY_DOC = "The provisioned throughput to reserved for writing resources to your table"; public static final String CAMEL_SINK_AWS2DDB_ENDPOINT_WRITE_CAPACITY_DEFAULT = null; @@ -129,6 +132,9 @@ public class CamelAws2ddbSinkConnectorConfig extends CamelSinkConnectorConfig { public static final String CAMEL_SINK_AWS2DDB_COMPONENT_URI_ENDPOINT_OVERRIDE_CONF = "camel.component.aws2-ddb.uriEndpointOverride"; public static final String CAMEL_SINK_AWS2DDB_COMPONENT_URI_ENDPOINT_OVERRIDE_DOC = "Set the overriding uri endpoint. This option needs to be used in combination with overrideEndpoint option"; public static final String CAMEL_SINK_AWS2DDB_COMPONENT_URI_ENDPOINT_OVERRIDE_DEFAULT = null; + public static final String CAMEL_SINK_AWS2DDB_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF = "camel.component.aws2-ddb.useDefaultCredentialsProvider"; + public static final String CAMEL_SINK_AWS2DDB_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC = "Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in."; + public static final Boolean CAMEL_SINK_AWS2DDB_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT = false; public static final String CAMEL_SINK_AWS2DDB_COMPONENT_WRITE_CAPACITY_CONF = "camel.component.aws2-ddb.writeCapacity"; public static final String CAMEL_SINK_AWS2DDB_COMPONENT_WRITE_CAPACITY_DOC = "The provisioned throughput to reserved for writing resources to your table"; public static final String CAMEL_SINK_AWS2DDB_COMPONENT_WRITE_CAPACITY_DEFAULT = null; @@ -170,6 +176,7 @@ public class CamelAws2ddbSinkConnectorConfig extends CamelSinkConnectorConfig { conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_REGION_CONF, ConfigDef.Type.STRING, CAMEL_SINK_AWS2DDB_ENDPOINT_REGION_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_REGION_DOC); conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_TRUST_ALL_CERTIFICATES_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_AWS2DDB_ENDPOINT_TRUST_ALL_CERTIFICATES_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_TRUST_ALL_CERTIFICATES_DOC); conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_URI_ENDPOINT_OVERRIDE_CONF, ConfigDef.Type.STRING, CAMEL_SINK_AWS2DDB_ENDPOINT_URI_ENDPOINT_OVERRIDE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_URI_ENDPOINT_OVERRIDE_DOC); + conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_AWS2DDB_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC); conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_WRITE_CAPACITY_CONF, ConfigDef.Type.STRING, CAMEL_SINK_AWS2DDB_ENDPOINT_WRITE_CAPACITY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_WRITE_CAPACITY_DOC); conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_ACCESS_KEY_CONF, ConfigDef.Type.PASSWORD, CAMEL_SINK_AWS2DDB_ENDPOINT_ACCESS_KEY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_ACCESS_KEY_DOC); conf.define(CAMEL_SINK_AWS2DDB_ENDPOINT_SECRET_KEY_CONF, ConfigDef.Type.PASSWORD, CAMEL_SINK_AWS2DDB_ENDPOINT_SECRET_KEY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_ENDPOINT_SECRET_KEY_DOC); @@ -189,6 +196,7 @@ public class CamelAws2ddbSinkConnectorConfig extends CamelSinkConnectorConfig { conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_REGION_CONF, ConfigDef.Type.STRING, CAMEL_SINK_AWS2DDB_COMPONENT_REGION_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_REGION_DOC); conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_TRUST_ALL_CERTIFICATES_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_AWS2DDB_COMPONENT_TRUST_ALL_CERTIFICATES_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_TRUST_ALL_CERTIFICATES_DOC); conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_URI_ENDPOINT_OVERRIDE_CONF, ConfigDef.Type.STRING, CAMEL_SINK_AWS2DDB_COMPONENT_URI_ENDPOINT_OVERRIDE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_URI_ENDPOINT_OVERRIDE_DOC); + conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_AWS2DDB_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_USE_DEFAULT_CREDENTIALS_PROVIDER_DOC); conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_WRITE_CAPACITY_CONF, ConfigDef.Type.STRING, CAMEL_SINK_AWS2DDB_COMPONENT_WRITE_CAPACITY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_WRITE_CAPACITY_DOC); conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_AUTOWIRED_ENABLED_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SINK_AWS2DDB_COMPONENT_AUTOWIRED_ENABLED_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_AUTOWIRED_ENABLED_DOC); conf.define(CAMEL_SINK_AWS2DDB_COMPONENT_ACCESS_KEY_CONF, ConfigDef.Type.PASSWORD, CAMEL_SINK_AWS2DDB_COMPONENT_ACCESS_KEY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SINK_AWS2DDB_COMPONENT_ACCESS_KEY_DOC); diff --git a/connectors/camel-jt400-kafka-connector/src/generated/resources/camel-jt400-source.json b/connectors/camel-jt400-kafka-connector/src/generated/resources/camel-jt400-source.json index 56a31a2..21d384c 100644 --- a/connectors/camel-jt400-kafka-connector/src/generated/resources/camel-jt400-source.json +++ b/connectors/camel-jt400-kafka-connector/src/generated/resources/camel-jt400-source.json @@ -130,6 +130,13 @@ "priority": "MEDIUM", "required": "false" }, + "camel.source.endpoint.sendingReply": { + "name": "camel.source.endpoint.sendingReply", + "description": "If true, the consumer endpoint will set the Jt400Constants.MESSAGE_REPLYTO_KEY header of the camel message for any IBM i inquiry messages received. If that message is then routed to a producer endpoint, the action will not be processed as sending a message to the queue, but rather a reply to the specific inquiry message.", + "defaultValue": "true", + "priority": "MEDIUM", + "required": "false" + }, "camel.source.endpoint.exceptionHandler": { "name": "camel.source.endpoint.exceptionHandler", "description": "To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored.", diff --git a/connectors/camel-jt400-kafka-connector/src/main/docs/camel-jt400-kafka-source-connector.adoc b/connectors/camel-jt400-kafka-connector/src/main/docs/camel-jt400-kafka-source-connector.adoc index 9e4e737..0778765 100644 --- a/connectors/camel-jt400-kafka-connector/src/main/docs/camel-jt400-kafka-source-connector.adoc +++ b/connectors/camel-jt400-kafka-connector/src/main/docs/camel-jt400-kafka-source-connector.adoc @@ -24,7 +24,7 @@ connector.class=org.apache.camel.kafkaconnector.jt400.CamelJt400SourceConnector ---- -The camel-jt400 source connector supports 36 options, which are listed below. +The camel-jt400 source connector supports 37 options, which are listed below. @@ -46,6 +46,7 @@ The camel-jt400 source connector supports 36 options, which are listed below. | *camel.source.endpoint.readTimeout* | Timeout in millis the consumer will wait while trying to read a new message of the data queue. | 30000 | false | MEDIUM | *camel.source.endpoint.searchType* | Search type such as EQ for equal etc. One of: [EQ] [NE] [LT] [LE] [GT] [GE] | "EQ" | false | MEDIUM | *camel.source.endpoint.sendEmptyMessageWhenIdle* | If the polling consumer did not poll any files, you can enable this option to send an empty message (no body) instead. | false | false | MEDIUM +| *camel.source.endpoint.sendingReply* | If true, the consumer endpoint will set the Jt400Constants.MESSAGE_REPLYTO_KEY header of the camel message for any IBM i inquiry messages received. If that message is then routed to a producer endpoint, the action will not be processed as sending a message to the queue, but rather a reply to the specific inquiry message. | true | false | MEDIUM | *camel.source.endpoint.exceptionHandler* | To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored. | null | false | MEDIUM | *camel.source.endpoint.exchangePattern* | Sets the exchange pattern when the consumer creates an exchange. One of: [InOnly] [InOut] [InOptionalOut] | null | false | MEDIUM | *camel.source.endpoint.pollStrategy* | A pluggable org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your custom implementation to control error handling usually occurred during the poll operation before an Exchange have been created and being routed in Camel. | null | false | MEDIUM diff --git a/connectors/camel-jt400-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/jt400/CamelJt400SourceConnectorConfig.java b/connectors/camel-jt400-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/jt400/CamelJt400SourceConnectorConfig.java index 6602768..c2dbaae 100644 --- a/connectors/camel-jt400-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/jt400/CamelJt400SourceConnectorConfig.java +++ b/connectors/camel-jt400-kafka-connector/src/main/java/org/apache/camel/kafkaconnector/jt400/CamelJt400SourceConnectorConfig.java @@ -71,6 +71,9 @@ public class CamelJt400SourceConnectorConfig public static final String CAMEL_SOURCE_JT400_ENDPOINT_SEND_EMPTY_MESSAGE_WHEN_IDLE_CONF = "camel.source.endpoint.sendEmptyMessageWhenIdle"; public static final String CAMEL_SOURCE_JT400_ENDPOINT_SEND_EMPTY_MESSAGE_WHEN_IDLE_DOC = "If the polling consumer did not poll any files, you can enable this option to send an empty message (no body) instead."; public static final Boolean CAMEL_SOURCE_JT400_ENDPOINT_SEND_EMPTY_MESSAGE_WHEN_IDLE_DEFAULT = false; + public static final String CAMEL_SOURCE_JT400_ENDPOINT_SENDING_REPLY_CONF = "camel.source.endpoint.sendingReply"; + public static final String CAMEL_SOURCE_JT400_ENDPOINT_SENDING_REPLY_DOC = "If true, the consumer endpoint will set the Jt400Constants.MESSAGE_REPLYTO_KEY header of the camel message for any IBM i inquiry messages received. If that message is then routed to a producer endpoint, the action will not be processed as sending a message to the queue, but rather a reply to the specific inquiry message."; + public static final Boolean CAMEL_SOURCE_JT400_ENDPOINT_SENDING_REPLY_DEFAULT = true; public static final String CAMEL_SOURCE_JT400_ENDPOINT_EXCEPTION_HANDLER_CONF = "camel.source.endpoint.exceptionHandler"; public static final String CAMEL_SOURCE_JT400_ENDPOINT_EXCEPTION_HANDLER_DOC = "To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored."; public static final String CAMEL_SOURCE_JT400_ENDPOINT_EXCEPTION_HANDLER_DEFAULT = null; @@ -162,6 +165,7 @@ public class CamelJt400SourceConnectorConfig conf.define(CAMEL_SOURCE_JT400_ENDPOINT_READ_TIMEOUT_CONF, ConfigDef.Type.INT, CAMEL_SOURCE_JT400_ENDPOINT_READ_TIMEOUT_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_READ_TIMEOUT_DOC); conf.define(CAMEL_SOURCE_JT400_ENDPOINT_SEARCH_TYPE_CONF, ConfigDef.Type.STRING, CAMEL_SOURCE_JT400_ENDPOINT_SEARCH_TYPE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_SEARCH_TYPE_DOC); conf.define(CAMEL_SOURCE_JT400_ENDPOINT_SEND_EMPTY_MESSAGE_WHEN_IDLE_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SOURCE_JT400_ENDPOINT_SEND_EMPTY_MESSAGE_WHEN_IDLE_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_SEND_EMPTY_MESSAGE_WHEN_IDLE_DOC); + conf.define(CAMEL_SOURCE_JT400_ENDPOINT_SENDING_REPLY_CONF, ConfigDef.Type.BOOLEAN, CAMEL_SOURCE_JT400_ENDPOINT_SENDING_REPLY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_SENDING_REPLY_DOC); conf.define(CAMEL_SOURCE_JT400_ENDPOINT_EXCEPTION_HANDLER_CONF, ConfigDef.Type.STRING, CAMEL_SOURCE_JT400_ENDPOINT_EXCEPTION_HANDLER_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_EXCEPTION_HANDLER_DOC); conf.define(CAMEL_SOURCE_JT400_ENDPOINT_EXCHANGE_PATTERN_CONF, ConfigDef.Type.STRING, CAMEL_SOURCE_JT400_ENDPOINT_EXCHANGE_PATTERN_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_EXCHANGE_PATTERN_DOC); conf.define(CAMEL_SOURCE_JT400_ENDPOINT_POLL_STRATEGY_CONF, ConfigDef.Type.STRING, CAMEL_SOURCE_JT400_ENDPOINT_POLL_STRATEGY_DEFAULT, ConfigDef.Importance.MEDIUM, CAMEL_SOURCE_JT400_ENDPOINT_POLL_STRATEGY_DOC); diff --git a/docs/modules/ROOT/pages/connectors/camel-atlasmap-kafka-sink-connector.adoc b/docs/modules/ROOT/pages/connectors/camel-atlasmap-kafka-sink-connector.adoc index 0220636..45c966b 100644 --- a/docs/modules/ROOT/pages/connectors/camel-atlasmap-kafka-sink-connector.adoc +++ b/docs/modules/ROOT/pages/connectors/camel-atlasmap-kafka-sink-connector.adoc @@ -38,10 +38,10 @@ The camel-atlasmap sink connector supports 11 options, which are listed below. | *camel.sink.endpoint.sourceMapName* | The Exchange property name for a source message map which hold java.util.Map<String, Message> where the key is AtlasMap Document ID. AtlasMap consumes Message bodies as source documents, as well as message headers as source properties where the scope equals to Document ID. | null | false | MEDIUM | *camel.sink.endpoint.targetMapMode* | TargetMapMode enum value to specify how multiple target documents are delivered if exist. 'MAP': Stores them into a java.util.Map, and the java.util.Map is set to an exchange property if 'targetMapName' is specified, otherwise message body. 'MESSAGE_HEADER': Stores them into message headers. 'EXCHANGE_PROPERTY': Stores them into exchange properties. ) One of: [MAP] [MESSAGE_HEADER] [EXCHANGE_PROPERTY] | "MAP" | false | MEDIUM | *camel.sink.endpoint.targetMapName* | The Exchange property name for a target document map which hold java.util.Map<String, Object> where the key is AtlasMap Document ID. AtlasMap populates multiple target documents into this map. | null | false | MEDIUM -| *camel.sink.endpoint.propertiesFile* | The URI of the properties file which is used for AtlasContextFactory initialization. | null | false | MEDIUM | *camel.component.atlasmap.lazyStartProducer* | Whether the producer should be started lazy (on the first message). By starting lazy you can use this to allow CamelContext and routes to startup in situations where a producer may otherwise fail during starting and cause the route to fail being started. By deferring this startup to be lazy then the startup failure can be handled during routing messages via Camel's routing error handlers. Beware that when the first message is processed the [...] | *camel.component.atlasmap.atlasContextFactory* | To use the AtlasContextFactory otherwise a new engine is created. | null | false | MEDIUM | *camel.component.atlasmap.autowiredEnabled* | Whether autowiring is enabled. This is used for automatic autowiring options (the option must be marked as autowired) by looking up in the registry to find if there is a single instance of matching type, which then gets configured on the component. This can be used for automatic configuring JDBC data sources, JMS connection factories, AWS Clients, etc. | true | false | MEDIUM +| *camel.component.atlasmap.propertiesFile* | The URI of the properties file which is used for AtlasContextFactory initialization. | null | false | MEDIUM |=== diff --git a/docs/modules/ROOT/pages/connectors/camel-aws2-ddb-kafka-sink-connector.adoc b/docs/modules/ROOT/pages/connectors/camel-aws2-ddb-kafka-sink-connector.adoc index 7f8a826..3e73ea1 100644 --- a/docs/modules/ROOT/pages/connectors/camel-aws2-ddb-kafka-sink-connector.adoc +++ b/docs/modules/ROOT/pages/connectors/camel-aws2-ddb-kafka-sink-connector.adoc @@ -24,7 +24,7 @@ connector.class=org.apache.camel.kafkaconnector.aws2ddb.CamelAws2ddbSinkConnecto ---- -The camel-aws2-ddb sink connector supports 39 options, which are listed below. +The camel-aws2-ddb sink connector supports 41 options, which are listed below. @@ -47,6 +47,7 @@ The camel-aws2-ddb sink connector supports 39 options, which are listed below. | *camel.sink.endpoint.region* | The region in which DDB client needs to work | null | false | MEDIUM | *camel.sink.endpoint.trustAllCertificates* | If we want to trust all certificates in case of overriding the endpoint | false | false | MEDIUM | *camel.sink.endpoint.uriEndpointOverride* | Set the overriding uri endpoint. This option needs to be used in combination with overrideEndpoint option | null | false | MEDIUM +| *camel.sink.endpoint.useDefaultCredentialsProvider* | Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in. | false | false | MEDIUM | *camel.sink.endpoint.writeCapacity* | The provisioned throughput to reserved for writing resources to your table | null | false | MEDIUM | *camel.sink.endpoint.accessKey* | Amazon AWS Access Key | null | false | MEDIUM | *camel.sink.endpoint.secretKey* | Amazon AWS Secret Key | null | false | MEDIUM @@ -66,6 +67,7 @@ The camel-aws2-ddb sink connector supports 39 options, which are listed below. | *camel.component.aws2-ddb.region* | The region in which DDB client needs to work | null | false | MEDIUM | *camel.component.aws2-ddb.trustAllCertificates* | If we want to trust all certificates in case of overriding the endpoint | false | false | MEDIUM | *camel.component.aws2-ddb.uriEndpointOverride* | Set the overriding uri endpoint. This option needs to be used in combination with overrideEndpoint option | null | false | MEDIUM +| *camel.component.aws2-ddb.useDefaultCredentials Provider* | Set whether the S3 client should expect to load credentials through a default credentials provider or to expect static credentials to be passed in. | false | false | MEDIUM | *camel.component.aws2-ddb.writeCapacity* | The provisioned throughput to reserved for writing resources to your table | null | false | MEDIUM | *camel.component.aws2-ddb.autowiredEnabled* | Whether autowiring is enabled. This is used for automatic autowiring options (the option must be marked as autowired) by looking up in the registry to find if there is a single instance of matching type, which then gets configured on the component. This can be used for automatic configuring JDBC data sources, JMS connection factories, AWS Clients, etc. | true | false | MEDIUM | *camel.component.aws2-ddb.accessKey* | Amazon AWS Access Key | null | false | MEDIUM diff --git a/docs/modules/ROOT/pages/connectors/camel-jt400-kafka-source-connector.adoc b/docs/modules/ROOT/pages/connectors/camel-jt400-kafka-source-connector.adoc index 9e4e737..0778765 100644 --- a/docs/modules/ROOT/pages/connectors/camel-jt400-kafka-source-connector.adoc +++ b/docs/modules/ROOT/pages/connectors/camel-jt400-kafka-source-connector.adoc @@ -24,7 +24,7 @@ connector.class=org.apache.camel.kafkaconnector.jt400.CamelJt400SourceConnector ---- -The camel-jt400 source connector supports 36 options, which are listed below. +The camel-jt400 source connector supports 37 options, which are listed below. @@ -46,6 +46,7 @@ The camel-jt400 source connector supports 36 options, which are listed below. | *camel.source.endpoint.readTimeout* | Timeout in millis the consumer will wait while trying to read a new message of the data queue. | 30000 | false | MEDIUM | *camel.source.endpoint.searchType* | Search type such as EQ for equal etc. One of: [EQ] [NE] [LT] [LE] [GT] [GE] | "EQ" | false | MEDIUM | *camel.source.endpoint.sendEmptyMessageWhenIdle* | If the polling consumer did not poll any files, you can enable this option to send an empty message (no body) instead. | false | false | MEDIUM +| *camel.source.endpoint.sendingReply* | If true, the consumer endpoint will set the Jt400Constants.MESSAGE_REPLYTO_KEY header of the camel message for any IBM i inquiry messages received. If that message is then routed to a producer endpoint, the action will not be processed as sending a message to the queue, but rather a reply to the specific inquiry message. | true | false | MEDIUM | *camel.source.endpoint.exceptionHandler* | To let the consumer use a custom ExceptionHandler. Notice if the option bridgeErrorHandler is enabled then this option is not in use. By default the consumer will deal with exceptions, that will be logged at WARN or ERROR level and ignored. | null | false | MEDIUM | *camel.source.endpoint.exchangePattern* | Sets the exchange pattern when the consumer creates an exchange. One of: [InOnly] [InOut] [InOptionalOut] | null | false | MEDIUM | *camel.source.endpoint.pollStrategy* | A pluggable org.apache.camel.PollingConsumerPollingStrategy allowing you to provide your custom implementation to control error handling usually occurred during the poll operation before an Exchange have been created and being routed in Camel. | null | false | MEDIUM