arjun180 commented on issue #1241: URL: https://github.com/apache/camel-kafka-connector/issues/1241#issuecomment-893675388
Thanks @oscerd . I changed the Splunk url to `my server.com:<port>` and I don't get the URL error anymore. I also removed parameters `camel.sink.endpoint.bodyOnly: false` and `camel.sink.endpoint.headersOnly: false` because they weren't valid. Now I get : ``` 2021-08-05 17:41:04,041 ERROR Failed delivery for (MessageId: 6413467DD725059-0000000000000000 on ExchangeId: 6413467DD725059-0000000000000000). Exhausted after delivery attempt: 1 caught: java.lang.RuntimeException: HTTP/1.1 503 Service Unavailable: Back-end server is at capacity Message History (complete message history is disabled) --------------------------------------------------------------------------------------------------------------------------------------- RouteId ProcessorId Processor Elapsed (ms) [route6 ] [route6 ] [ ] [ 400] ... [route6 ] [toD6 ] [splunk-hec:myserver:8088/<token>?htt] [ 0] Stacktrace --------------------------------------------------------------------------------------------------------------------------------------- (org.apache.camel.processor.errorhandler.DefaultErrorHandler) [task-thread-splunk-sink-connector-0] java.lang.RuntimeException: HTTP/1.1 503 Service Unavailable: Back-end server is at capacity at org.apache.camel.component.splunkhec.SplunkHECProducer.process(SplunkHECProducer.java:86) at org.apache.camel.support.AsyncProcessorConverterHelper$ProcessorToAsyncProcessorBridge.process(AsyncProcessorConverterHelper.java:66) at org.apache.camel.processor.SendDynamicProcessor.lambda$process$0(SendDynamicProcessor.java:197) at org.apache.camel.support.cache.DefaultProducerCache.doInAsyncProducer(DefaultProducerCache.java:318) at org.apache.camel.processor.SendDynamicProcessor.process(SendDynamicProcessor.java:182) at org.apache.camel.processor.errorhandler.RedeliveryErrorHandler$SimpleTask.run(RedeliveryErrorHandler.java:439) at org.apache.camel.impl.engine.DefaultReactiveExecutor$Worker.schedule(DefaultReactiveExecutor.java:181) at org.apache.camel.impl.engine.DefaultReactiveExecutor.scheduleMain(DefaultReactiveExecutor.java:62) at org.apache.camel.processor.Pipeline.process(Pipeline.java:167) at org.apache.camel.impl.engine.CamelInternalProcessor.process(CamelInternalProcessor.java:388) at org.apache.camel.component.direct.DirectProducer.process(DirectProducer.java:96) at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:217) at org.apache.camel.impl.engine.SharedCamelInternalProcessor$1.process(SharedCamelInternalProcessor.java:111) at org.apache.camel.impl.engine.DefaultAsyncProcessorAwaitManager.process(DefaultAsyncProcessorAwaitManager.java:83) at org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:108) at org.apache.camel.support.cache.DefaultProducerCache.send(DefaultProducerCache.java:190) at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:176) at org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:148) at org.apache.camel.kafkaconnector.CamelSinkTask.put(CamelSinkTask.java:194) at org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:581) at org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:329) at org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:232) at org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:201) at org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:182) at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:231) at java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515) at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264) at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128) at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628) at java.base/java.lang.Thread.run(Thread.java:829) ``` But I curled the Splunk URL from the Kafka Connect pod to check if there was an issue and it came out healthy. The `Service Unavailable: Back-end server is at capacity` error seems to be an issue with the Splunk LB, but that doesn't seem to be a problem in this case. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: commits-unsubscr...@camel.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org