arjun180 removed a comment on issue #1241:
URL: 
https://github.com/apache/camel-kafka-connector/issues/1241#issuecomment-893675388


   Thanks @oscerd . I changed the Splunk url to `my server.com:<port>` and I 
don't get the URL error anymore. I also removed parameters 
`camel.sink.endpoint.bodyOnly: false` and `camel.sink.endpoint.headersOnly: 
false` because they weren't valid. 
   
   Now I get : 
   
   ```
   2021-08-05 17:41:04,041 ERROR Failed delivery for (MessageId: 
6413467DD725059-0000000000000000 on ExchangeId: 
6413467DD725059-0000000000000000). Exhausted after delivery attempt: 1 caught: 
java.lang.RuntimeException: HTTP/1.1 503 Service Unavailable: Back-end server 
is at capacity
   
   
   Message History (complete message history is disabled)
   
---------------------------------------------------------------------------------------------------------------------------------------
   RouteId              ProcessorId          Processor                          
                                              Elapsed (ms)
   [route6            ] [route6            ] [                                  
                                            ] [       400]
        ...
   [route6            ] [toD6              ] 
[splunk-hec:myserver:8088/<token>?htt] [         0]
   
   Stacktrace
   
---------------------------------------------------------------------------------------------------------------------------------------
    (org.apache.camel.processor.errorhandler.DefaultErrorHandler) 
[task-thread-splunk-sink-connector-0]
   java.lang.RuntimeException: HTTP/1.1 503 Service Unavailable: Back-end 
server is at capacity
   
        at 
org.apache.camel.component.splunkhec.SplunkHECProducer.process(SplunkHECProducer.java:86)
        at 
org.apache.camel.support.AsyncProcessorConverterHelper$ProcessorToAsyncProcessorBridge.process(AsyncProcessorConverterHelper.java:66)
        at 
org.apache.camel.processor.SendDynamicProcessor.lambda$process$0(SendDynamicProcessor.java:197)
        at 
org.apache.camel.support.cache.DefaultProducerCache.doInAsyncProducer(DefaultProducerCache.java:318)
        at 
org.apache.camel.processor.SendDynamicProcessor.process(SendDynamicProcessor.java:182)
        at 
org.apache.camel.processor.errorhandler.RedeliveryErrorHandler$SimpleTask.run(RedeliveryErrorHandler.java:439)
        at 
org.apache.camel.impl.engine.DefaultReactiveExecutor$Worker.schedule(DefaultReactiveExecutor.java:181)
        at 
org.apache.camel.impl.engine.DefaultReactiveExecutor.scheduleMain(DefaultReactiveExecutor.java:62)
        at org.apache.camel.processor.Pipeline.process(Pipeline.java:167)
        at 
org.apache.camel.impl.engine.CamelInternalProcessor.process(CamelInternalProcessor.java:388)
        at 
org.apache.camel.component.direct.DirectProducer.process(DirectProducer.java:96)
        at 
org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:217)
        at 
org.apache.camel.impl.engine.SharedCamelInternalProcessor$1.process(SharedCamelInternalProcessor.java:111)
        at 
org.apache.camel.impl.engine.DefaultAsyncProcessorAwaitManager.process(DefaultAsyncProcessorAwaitManager.java:83)
        at 
org.apache.camel.impl.engine.SharedCamelInternalProcessor.process(SharedCamelInternalProcessor.java:108)
        at 
org.apache.camel.support.cache.DefaultProducerCache.send(DefaultProducerCache.java:190)
        at 
org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:176)
        at 
org.apache.camel.impl.engine.DefaultProducerTemplate.send(DefaultProducerTemplate.java:148)
        at 
org.apache.camel.kafkaconnector.CamelSinkTask.put(CamelSinkTask.java:194)
        at 
org.apache.kafka.connect.runtime.WorkerSinkTask.deliverMessages(WorkerSinkTask.java:581)
        at 
org.apache.kafka.connect.runtime.WorkerSinkTask.poll(WorkerSinkTask.java:329)
        at 
org.apache.kafka.connect.runtime.WorkerSinkTask.iteration(WorkerSinkTask.java:232)
        at 
org.apache.kafka.connect.runtime.WorkerSinkTask.execute(WorkerSinkTask.java:201)
        at 
org.apache.kafka.connect.runtime.WorkerTask.doRun(WorkerTask.java:182)
        at org.apache.kafka.connect.runtime.WorkerTask.run(WorkerTask.java:231)
        at 
java.base/java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:515)
        at java.base/java.util.concurrent.FutureTask.run(FutureTask.java:264)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
        at 
java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
        at java.base/java.lang.Thread.run(Thread.java:829)
   ```
   
   But I curled the Splunk URL from the Kafka Connect pod to check if there was 
an issue and it came out healthy. The `Service Unavailable: Back-end server is 
at capacity` error seems to be an issue with the Splunk LB, but that doesn't 
seem to be a problem in this case.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: commits-unsubscr...@camel.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


Reply via email to