Looks like you have one device class and the same replication on all pools, 
which makes that simpler.

Your MAX AVAIL figures are lower than I would expect if you're using size=3, so 
I'd check if you have the balancer enabled, if it's working properly.

Run

ceph osd df

and look at the VAR column, 

[rook@rook-ceph-tools-5ff8d58445-p9npl /]$ ceph osd df | head
ID   CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE   DATA      OMAP     META     
AVAIL    %USE   VAR   PGS  STATUS

Ideally the numbers should all be close to 1.00 + / -

> On Mar 20, 2024, at 16:55, Michael Worsham <[email protected]> 
> wrote:
> 
> I had a request from the upper management wanting to use SolarWinds to be 
> able to extract what I am looking at and have SolarWinds track it in terms of 
> total available space, remaining space of the overall cluster, and I guess 
> would be the current RGW pools/buckets we have and their allocated sizes and 
> space remaining in it as well. I am sort of in the dark when it comes to 
> trying to break things down to make it readable/understandable for those that 
> are non-technical.
> 
> I was told that when it comes to pools and buckets, you sort of have to see 
> it this way:
> - Bucket is like a folder
> - Pool is like a hard drive.
> - You can create many folders in a hard drive and you can add quota to each 
> folder.
> - But if you want to know the remaining space, you need to check the hard 
> drive.
> 
> I did the "ceph df" command on the ceph monitor and we have something that 
> looks like this:
> 
>>> sudo ceph df
> --- RAW STORAGE ---
> CLASS     SIZE    AVAIL     USED  RAW USED  %RAW USED
> ssd    873 TiB  346 TiB  527 TiB   527 TiB      60.40
> TOTAL  873 TiB  346 TiB  527 TiB   527 TiB      60.40
> 
> --- POOLS ---
> POOL                             ID   PGS   STORED  OBJECTS     USED  %USED  
> MAX AVAIL
> .mgr                              1     1  449 KiB        2  1.3 MiB      0   
>   61 TiB
> default.rgw.buckets.data          2  2048  123 TiB   41.86M  371 TiB  66.76   
>   61 TiB
> default.rgw.control               3     2      0 B        8      0 B      0   
>   61 TiB
> default.rgw.data.root             4     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.gc                    5     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.log                   6     2   41 KiB      209  732 KiB      0   
>   61 TiB
> default.rgw.intent-log            7     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.meta                  8     2   20 KiB       96  972 KiB      0   
>   61 TiB
> default.rgw.otp                   9     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.usage                10     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.users.keys           11     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.users.email          12     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.users.swift          13     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.users.uid            14     2      0 B        0      0 B      0   
>   61 TiB
> default.rgw.buckets.extra        15    16      0 B        0      0 B      0   
>   61 TiB
> default.rgw.buckets.index        16    64  6.3 GiB      184   19 GiB   0.01   
>   61 TiB
> .rgw.root                        17     2  2.3 KiB        4   48 KiB      0   
>   61 TiB
> ceph-benchmarking                18   128  596 GiB  302.20k  1.7 TiB   0.94   
>   61 TiB
> ceph-fs_data                     19    64  438 MiB      110  1.3 GiB      0   
>   61 TiB
> ceph-fs_metadata                 20    16   37 MiB       32  111 MiB      0   
>   61 TiB
> test                             21    32   21 TiB    5.61M   64 TiB  25.83   
>   61 TiB
> DD-Test                          22    32   11 MiB       13   32 MiB      0   
>   61 TiB
> nativesqlbackup                  24    32  539 MiB      147  1.6 GiB      0   
>   61 TiB
> default.rgw.buckets.non-ec       25    32  1.7 MiB        0  5.0 MiB      0   
>   61 TiB
> ceph-fs_sql_backups              26    32      0 B        0      0 B      0   
>   61 TiB
> ceph-fs_sql_backups_metadata     27    32      0 B        0      0 B      0   
>   61 TiB
> dd-drs-backups                   28    32      0 B        0      0 B      0   
>   61 TiB
> default.rgw.jv-corp-pool.data    59    32   16 TiB   63.90M   49 TiB  21.12   
>   61 TiB
> default.rgw.jv-corp-pool.index   60    32  108 GiB    1.19k  323 GiB   0.17   
>   61 TiB
> default.rgw.jv-corp-pool.non-ec  61    32      0 B        0      0 B      0   
>   61 TiB
> default.rgw.jv-comm-pool.data    62    32  8.1 TiB   44.20M   24 TiB  11.65   
>   61 TiB
> default.rgw.jv-comm-pool.index   63    32   83 GiB      811  248 GiB   0.13   
>   61 TiB
> default.rgw.jv-comm-pool.non-ec  64    32      0 B        0      0 B      0   
>   61 TiB
> default.rgw.jv-va-pool.data      65    32  4.8 TiB   22.17M   14 TiB   7.28   
>   61 TiB
> default.rgw.jv-va-pool.index     66    32   38 GiB      401  113 GiB   0.06   
>   61 TiB
> default.rgw.jv-va-pool.non-ec    67    32      0 B        0      0 B      0   
>   61 TiB
> jv-edi-pool                      68    32      0 B        0      0 B      0   
>   61 TiB
> 
> -- Michael
> 
> -----Original Message-----
> From: Anthony D'Atri <[email protected]>
> Sent: Wednesday, March 20, 2024 2:48 PM
> To: Michael Worsham <[email protected]>
> Cc: [email protected]
> Subject: Re: [ceph-users] Need easy way to calculate Ceph cluster space for 
> SolarWinds
> 
> This is an external email. Please take care when clicking links or opening 
> attachments. When in doubt, check with the Help Desk or Security.
> 
> 
>> On Mar 20, 2024, at 14:42, Michael Worsham <[email protected]> 
>> wrote:
>> 
>> Is there an easy way to poll a Ceph cluster to see how much space is
>> available
> 
> `ceph df`
> 
> The exporter has percentages per pool as well.
> 
> 
>> and how much space is available per bucket?
> 
> Are you using RGW quotas?
> 
>> 
>> Looking for a way to use SolarWinds to monitor the entire Ceph cluster space 
>> utilization and then also be able to break down each RGW bucket to see how 
>> much space it was provisioned for and how much is available.
> 
> RGW buckets do not provision space.  Optionally there may be some RGW quotas 
> but they're a different thing than you're implying.
> 
> 
>> 
>> -- Michael
>> 
>> 
>> Get Outlook for Android<https://aka.ms/AAb9ysg> This message and its
>> attachments are from Data Dimensions and are intended only for the use of 
>> the individual or entity to which it is addressed, and may contain 
>> information that is privileged, confidential, and exempt from disclosure 
>> under applicable law. If the reader of this message is not the intended 
>> recipient, or the employee or agent responsible for delivering the message 
>> to the intended recipient, you are hereby notified that any dissemination, 
>> distribution, or copying of this communication is strictly prohibited. If 
>> you have received this communication in error, please notify the sender 
>> immediately and permanently delete the original email and destroy any copies 
>> or printouts of this email as well as any attachments.
>> _______________________________________________
>> ceph-users mailing list -- [email protected] To unsubscribe send an
>> email to [email protected]
> 
> This message and its attachments are from Data Dimensions and are intended 
> only for the use of the individual or entity to which it is addressed, and 
> may contain information that is privileged, confidential, and exempt from 
> disclosure under applicable law. If the reader of this message is not the 
> intended recipient, or the employee or agent responsible for delivering the 
> message to the intended recipient, you are hereby notified that any 
> dissemination, distribution, or copying of this communication is strictly 
> prohibited. If you have received this communication in error, please notify 
> the sender immediately and permanently delete the original email and destroy 
> any copies or printouts of this email as well as any attachments.
> 
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to