Interesting, thanks for the data. It does seem like the process is spending
a lot of time doing GC like I thought.

One trick you could try is to increase the memory allocated to the prober,
which would reduce the time spent on GC.

The default setting is is GOGC=100.

You could try increasing this by setting the environment variable, GOGC.

Try something like GOGC=200 or GOGC=300.

This will make the process use more memory, but it should reduce the CPU
time spent.

On Sun, Feb 25, 2024 at 11:18 PM Alexander Wilke <[email protected]>
wrote:

> Hello,
>
> I attached a few screenshots showing the results and graphs for 1h and 6h.
> In addition I added a screenshot from node_exporter metrics to give you an
> overview of the system itself.
> On the same system there is prometheus, grafana, snmp_exporter (200-800%
> CPU), smokeping prober, node_exporter, blackbox_exporter.
> The main CPU consumers are snmp_exporter and smokeping.
>
> Ben Kochie schrieb am Sonntag, 25. Februar 2024 um 19:22:35 UTC+1:
>
>> Looking at the CPU profile, I'm seeing almost all the time spent in the
>> Go runtime. Mostly the ICMP packet receiving code and garbage collection.
>> I'm not sure there's a lot we can optimize here as it's core Go code for
>> ICMP packet handling.
>>
>> Can you also post me a graph of a few metrics queries?
>>
>> rate(process_cpu_seconds_total{job="smokeping_prober"}[30s])
>> rate(go_gc_duration_seconds_count{job="smokeping_prober"}[5m])
>> rate(go_gc_duration_seconds_sum{job="smokeping_prober"}[5m])
>>
>>
>> On Sun, Feb 25, 2024 at 7:08 PM Alexander Wilke <[email protected]>
>> wrote:
>>
>>> Hello,
>>> any Chance to investigate the Reports and any suggestions?
>>>
>>> Alexander Wilke schrieb am Donnerstag, 22. Februar 2024 um 12:40:09
>>> UTC+1:
>>>
>>>> Hello,
>>>>
>>>> sorry for the delay. here are the results. to be honest - I do not
>>>> understand anything of it.
>>>>
>>>> Smokeping_Prober Heap:
>>>>
>>>>
>>>> https://pprof.me/a1e7400d32859dbc217e2182398485df/?profileType=profile%3Aalloc_objects%3Acount%3Aspace%3Abytes&dashboard_items=icicle
>>>>
>>>>
>>>>
>>>> smokeping_prober profile30s
>>>>
>>>>
>>>> https://pprof.me/340674b335e114e4b0df6b4582f0644e/?profileType=profile%3Asamples%3Acount%3Acpu%3Ananoseconds%3Adelta
>>>>
>>>> Ben Kochie schrieb am Dienstag, 20. Februar 2024 um 10:27:10 UTC+1:
>>>>
>>>>> Best thing you can do is capture some pprof data. That will show you
>>>>> what it's spending the time on.
>>>>>
>>>>> :9374/debug/pprof/heap
>>>>> :9374/debug/pprof/profile?seconds=30
>>>>>
>>>>> You can post the results to https://pprof.me/ for sharing.
>>>>>
>>>>> On Tue, Feb 20, 2024 at 6:22 AM Alexander Wilke <[email protected]>
>>>>> wrote:
>>>>>
>>>>>> Hello,
>>>>>> I am running smokeping_prober from one VM to Monitor around 500
>>>>>> destinations.
>>>>>> Around 30 devices are monitored with 0.2s Intervall and Others with
>>>>>> 1.65s Intervall.
>>>>>>
>>>>>> Prometheus scrapes every 5s.
>>>>>>
>>>>>> So there are roughly 600 icmp ipv4 24byte pings per Seconds.
>>>>>> CPU usage jumps between 700-1200% using "top"
>>>>>>
>>>>>> What Else except reducing Interval or Host Count could Help to reduce
>>>>>> CPU usage?
>>>>>> Is the UDP Socket "better" or any other optimization which could be
>>>>>> relevant for that Type of Traffic? Running on RHEL8
>>>>>>
>>>>>> Someone with similar CPU usage and this amount of pings per Seconds?
>>>>>> Maybe Others Ping 6.000 Destination every 10s?
>>>>>>
>>>>>> --
>>>>>> You received this message because you are subscribed to the Google
>>>>>> Groups "Prometheus Users" group.
>>>>>> To unsubscribe from this group and stop receiving emails from it,
>>>>>> send an email to [email protected].
>>>>>> To view this discussion on the web visit
>>>>>> https://groups.google.com/d/msgid/prometheus-users/d803c1a2-64ee-48d1-8513-b864856f53c8n%40googlegroups.com
>>>>>> <https://groups.google.com/d/msgid/prometheus-users/d803c1a2-64ee-48d1-8513-b864856f53c8n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>>>>> .
>>>>>>
>>>>> --
>>> You received this message because you are subscribed to the Google
>>> Groups "Prometheus Users" group.
>>> To unsubscribe from this group and stop receiving emails from it, send
>>> an email to [email protected].
>>>
>> To view this discussion on the web visit
>>> https://groups.google.com/d/msgid/prometheus-users/62aeaacb-9fd6-4d64-8bec-d7171f592766n%40googlegroups.com
>>> <https://groups.google.com/d/msgid/prometheus-users/62aeaacb-9fd6-4d64-8bec-d7171f592766n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>> .
>>>
>> --
> You received this message because you are subscribed to the Google Groups
> "Prometheus Users" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/prometheus-users/045a6fb0-b82c-4425-8bae-e2c99403b2adn%40googlegroups.com
> <https://groups.google.com/d/msgid/prometheus-users/045a6fb0-b82c-4425-8bae-e2c99403b2adn%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/prometheus-users/CABbyFmrDGh4cJWHCtmD_-SvJDEiTtA-kOnafmu0oe57F4rWz_w%40mail.gmail.com.

Reply via email to