Failing the Manager has become my go-to when things seem “stuck”.

> On Jun 19, 2025, at 9:11 AM, Niklaus Hofer <[email protected]> 
> wrote:
> 
> Dear Eugen
> 
> My hero! This resolved this issue - the warnings are now gone.
> 
> We did try to restart the mons before, but never thought to restart the 
> mgrs...
> 
> Sincerely
> 
> Niklaus Hofer
> 
> On 19/06/2025 14.52, Eugen Block wrote:
>> Default question: have you tried to fail the mgr? ;-)
>> ceph mgr fail
>> Zitat von Niklaus Hofer <[email protected]>:
>>> Dear all
>>> 
>>> After upgrading to Pacific, we are now getting health warnings from the 
>>> auto scaler:
>>> 
>>>>            10 pools have too few placement groups
>>>>            8 pools have too many placement groups
>>> 
>>> The thing is, I do not particularly care about the recommendations from the 
>>> auto scaler. So I checked and saw that it was set to warning on all pools:
>>> 
>>> | ceph osd pool autoscale-status
>>>> POOL                                SIZE  TARGET SIZE  RATE  RAW CAPACITY  
>>>>  RATIO  TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE 
>>>>  BULK   openstack-glance                  902.7G                3.0        
>>>> 618.2T  0.0043                                  1.0     128              
>>>> warn       False  openstack-cinder                  99.48T                
>>>> 3.0        618.2T  0.4827                                  1.0    2048     
>>>>          warn       False  openstack-nova                    169.6G        
>>>>         3.0        618.2T  0.0008                                  1.0     
>>>> 128              warn       False  [...]
>>> 
>>> No worries, I just turn it off:
>>> 
>>> | ceph osd pool set noautoscale
>>> 
>>> Check it's off:
>>> 
>>> | ceph osd pool autoscale-status
>>>> POOL                                SIZE  TARGET SIZE  RATE  RAW CAPACITY  
>>>>  RATIO  TARGET RATIO  EFFECTIVE RATIO  BIAS  PG_NUM  NEW PG_NUM  AUTOSCALE 
>>>>  BULK   openstack-glance                  902.7G                3.0        
>>>> 618.2T  0.0043                                  1.0     128              
>>>> off        False  openstack-cinder                  99.48T                
>>>> 3.0        618.2T  0.4827                                  1.0    2048     
>>>>          off        False  openstack-nova                    169.6G        
>>>>         3.0        618.2T  0.0008                                  1.0     
>>>> 128              off        False
>>>> [...]
>>> 
>>> That looks as expected. But despite that, I'm still getting the health 
>>> warnings:
>>> 
>>> | ceph --status
>>>> [...]
>>>>            10 pools have too few placement groups
>>>>            8 pools have too many placement groups
>>> 
>>> So how can I get rid of the warnings (without changing the number of PGs on 
>>> my pools!)?
>>> 
>>> Sincerely
>>> 
>>> Niklaus Hofer
>>> -- 
>>> stepping stone AG
>>> Wasserwerkgasse 7
>>> CH-3011 Bern
>>> 
>>> Telefon: +41 31 332 53 63
>>> www.stepping-stone.ch
>>> [email protected]
>>> _______________________________________________
>>> ceph-users mailing list -- [email protected]
>>> To unsubscribe send an email to [email protected]
>> _______________________________________________
>> ceph-users mailing list -- [email protected]
>> To unsubscribe send an email to [email protected]
> 
> -- 
> Managed Kubernetes (Container as a Service) kostenlos testen!
> 
> Jetzt bestellen und 500 Franken Startguthaben sichern:
> https://www.stoney-cloud.com/caas/
> 
> Arbeitszeiten: Dienstag bis Freitag
> 
> stepping stone AG
> Wasserwerkgasse 7
> CH-3011 Bern
> 
> Telefon: +41 31 332 53 63
> www.stepping-stone.ch
> [email protected]
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to