It seems to be relatively close to that +/- 1.00 range.
ubuntu@juju-5dcfd8-3-lxd-2:~$ sudo ceph osd df
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META
AVAIL %USE VAR PGS STATUS
1 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 11 GiB 38 GiB 8.1
TiB 55.59 0.92 174 up
4 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 20 GiB 34 GiB 7.3
TiB 59.90 0.99 175 up
9 ssd 18.19040 1.00000 18 TiB 9.7 TiB 9.6 TiB 23 GiB 35 GiB 8.5
TiB 53.11 0.88 185 up
13 ssd 18.19040 1.00000 18 TiB 13 TiB 13 TiB 19 GiB 48 GiB 4.8
TiB 73.87 1.22 199 up
17 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 14 GiB 55 GiB 5.9
TiB 67.49 1.12 185 up
21 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 24 GiB 38 GiB 7.8
TiB 57.27 0.95 179 up
25 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 20 GiB 57 GiB 6.1
TiB 66.70 1.10 192 up
29 ssd 18.19040 1.00000 18 TiB 9.2 TiB 9.2 TiB 15 GiB 34 GiB 9.0
TiB 50.61 0.84 170 up
33 ssd 18.19040 1.00000 18 TiB 9.2 TiB 9.1 TiB 13 GiB 36 GiB 9.0
TiB 50.56 0.84 180 up
39 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 4.3 GiB 45 GiB 6.0
TiB 66.84 1.11 188 up
44 ssd 18.19040 1.00000 18 TiB 13 TiB 13 TiB 10 GiB 56 GiB 5.3
TiB 70.59 1.17 187 up
46 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 14 GiB 44 GiB 7.8
TiB 57.24 0.95 174 up
0 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 5.0 GiB 38 GiB 6.9
TiB 62.13 1.03 172 up
5 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 11 GiB 44 GiB 7.3
TiB 59.69 0.99 177 up
10 ssd 18.19040 1.00000 18 TiB 12 TiB 11 TiB 18 GiB 47 GiB 6.7
TiB 63.34 1.05 190 up
14 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 5.1 GiB 48 GiB 6.3
TiB 65.51 1.08 189 up
18 ssd 18.19040 1.00000 18 TiB 9.7 TiB 9.6 TiB 13 GiB 33 GiB 8.5
TiB 53.14 0.88 175 up
22 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 20 GiB 42 GiB 7.3
TiB 59.61 0.99 183 up
26 ssd 18.19040 1.00000 18 TiB 9.8 TiB 9.8 TiB 9.9 GiB 31 GiB 8.4
TiB 53.88 0.89 186 up
30 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 4.5 GiB 56 GiB 6.4
TiB 64.65 1.07 179 up
34 ssd 18.19040 1.00000 18 TiB 13 TiB 13 TiB 18 GiB 49 GiB 4.9
TiB 73.17 1.21 192 up
38 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 16 GiB 51 GiB 6.4
TiB 64.67 1.07 186 up
40 ssd 18.19040 1.00000 18 TiB 13 TiB 12 TiB 23 GiB 52 GiB 5.7
TiB 68.91 1.14 184 up
42 ssd 18.19040 1.00000 18 TiB 9.7 TiB 9.7 TiB 14 GiB 26 GiB 8.5
TiB 53.35 0.88 171 up
3 ssd 18.19040 1.00000 18 TiB 9.9 TiB 9.8 TiB 15 GiB 36 GiB 8.3
TiB 54.42 0.90 184 up
7 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 5.4 GiB 36 GiB 7.9
TiB 56.73 0.94 184 up
11 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 9.6 GiB 54 GiB 6.6
TiB 63.76 1.06 188 up
15 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 7.2 GiB 47 GiB 7.5
TiB 58.51 0.97 192 up
19 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 19 GiB 41 GiB 7.3
TiB 59.87 0.99 181 up
23 ssd 18.19040 1.00000 18 TiB 13 TiB 12 TiB 20 GiB 54 GiB 5.7
TiB 68.89 1.14 181 up
27 ssd 18.19040 1.00000 18 TiB 9.0 TiB 9.0 TiB 15 GiB 31 GiB 9.2
TiB 49.63 0.82 173 up
32 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 20 GiB 26 GiB 7.3
TiB 59.92 0.99 183 up
36 ssd 18.19040 1.00000 18 TiB 8.3 TiB 8.3 TiB 11 GiB 17 GiB 9.8
TiB 45.86 0.76 177 up
41 ssd 18.19040 1.00000 18 TiB 13 TiB 13 TiB 25 GiB 49 GiB 5.2
TiB 71.30 1.18 191 up
45 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 13 GiB 42 GiB 7.9
TiB 56.37 0.93 163 up
47 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 13 GiB 38 GiB 7.2
TiB 60.45 1.00 167 up
2 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 5.2 GiB 43 GiB 7.2
TiB 60.42 1.00 179 up
6 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 28 GiB 47 GiB 7.1
TiB 60.99 1.01 184 up
8 ssd 18.19040 1.00000 18 TiB 13 TiB 13 TiB 20 GiB 59 GiB 5.5
TiB 69.95 1.16 184 up
12 ssd 18.19040 1.00000 18 TiB 11 TiB 11 TiB 6.2 GiB 39 GiB 7.4
TiB 59.22 0.98 180 up
16 ssd 18.19040 1.00000 18 TiB 9.8 TiB 9.7 TiB 14 GiB 37 GiB 8.4
TiB 53.63 0.89 187 up
20 ssd 18.19040 1.00000 18 TiB 9.7 TiB 9.6 TiB 21 GiB 33 GiB 8.5
TiB 53.14 0.88 181 up
24 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 10 GiB 46 GiB 6.3
TiB 65.24 1.08 180 up
28 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 13 GiB 45 GiB 8.1
TiB 55.41 0.92 192 up
31 ssd 18.19040 1.00000 18 TiB 12 TiB 12 TiB 22 GiB 48 GiB 6.2
TiB 65.92 1.09 186 up
35 ssd 18.19040 1.00000 18 TiB 10 TiB 10 TiB 15 GiB 33 GiB 8.0
TiB 56.11 0.93 175 up
37 ssd 18.19040 1.00000 18 TiB 13 TiB 13 TiB 13 GiB 53 GiB 5.0
TiB 72.78 1.21 179 up
43 ssd 18.19040 1.00000 18 TiB 8.9 TiB 8.8 TiB 17 GiB 23 GiB 9.3
TiB 48.71 0.81 178 up
TOTAL 873 TiB 527 TiB 525 TiB 704 GiB 2.0 TiB 346
TiB 60.40
MIN/MAX VAR: 0.76/1.22 STDDEV: 6.98
-----Original Message-----
From: Anthony D'Atri <[email protected]>
Sent: Wednesday, March 20, 2024 5:09 PM
To: Michael Worsham <[email protected]>
Cc: [email protected]
Subject: Re: [ceph-users] Need easy way to calculate Ceph cluster space for
SolarWinds
This is an external email. Please take care when clicking links or opening
attachments. When in doubt, check with the Help Desk or Security.
Looks like you have one device class and the same replication on all pools,
which makes that simpler.
Your MAX AVAIL figures are lower than I would expect if you're using size=3, so
I'd check if you have the balancer enabled, if it's working properly.
Run
ceph osd df
and look at the VAR column,
[rook@rook-ceph-tools-5ff8d58445-p9npl /]$ ceph osd df | head
ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP META
AVAIL %USE VAR PGS STATUS
Ideally the numbers should all be close to 1.00 + / -
> On Mar 20, 2024, at 16:55, Michael Worsham <[email protected]>
> wrote:
>
> I had a request from the upper management wanting to use SolarWinds to be
> able to extract what I am looking at and have SolarWinds track it in terms of
> total available space, remaining space of the overall cluster, and I guess
> would be the current RGW pools/buckets we have and their allocated sizes and
> space remaining in it as well. I am sort of in the dark when it comes to
> trying to break things down to make it readable/understandable for those that
> are non-technical.
>
> I was told that when it comes to pools and buckets, you sort of have to see
> it this way:
> - Bucket is like a folder
> - Pool is like a hard drive.
> - You can create many folders in a hard drive and you can add quota to each
> folder.
> - But if you want to know the remaining space, you need to check the hard
> drive.
>
> I did the "ceph df" command on the ceph monitor and we have something that
> looks like this:
>
>>> sudo ceph df
> --- RAW STORAGE ---
> CLASS SIZE AVAIL USED RAW USED %RAW USED
> ssd 873 TiB 346 TiB 527 TiB 527 TiB 60.40
> TOTAL 873 TiB 346 TiB 527 TiB 527 TiB 60.40
>
> --- POOLS ---
> POOL ID PGS STORED OBJECTS USED %USED
> MAX AVAIL
> .mgr 1 1 449 KiB 2 1.3 MiB 0
> 61 TiB
> default.rgw.buckets.data 2 2048 123 TiB 41.86M 371 TiB 66.76
> 61 TiB
> default.rgw.control 3 2 0 B 8 0 B 0
> 61 TiB
> default.rgw.data.root 4 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.gc 5 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.log 6 2 41 KiB 209 732 KiB 0
> 61 TiB
> default.rgw.intent-log 7 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.meta 8 2 20 KiB 96 972 KiB 0
> 61 TiB
> default.rgw.otp 9 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.usage 10 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.users.keys 11 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.users.email 12 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.users.swift 13 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.users.uid 14 2 0 B 0 0 B 0
> 61 TiB
> default.rgw.buckets.extra 15 16 0 B 0 0 B 0
> 61 TiB
> default.rgw.buckets.index 16 64 6.3 GiB 184 19 GiB 0.01
> 61 TiB
> .rgw.root 17 2 2.3 KiB 4 48 KiB 0
> 61 TiB
> ceph-benchmarking 18 128 596 GiB 302.20k 1.7 TiB 0.94
> 61 TiB
> ceph-fs_data 19 64 438 MiB 110 1.3 GiB 0
> 61 TiB
> ceph-fs_metadata 20 16 37 MiB 32 111 MiB 0
> 61 TiB
> test 21 32 21 TiB 5.61M 64 TiB 25.83
> 61 TiB
> DD-Test 22 32 11 MiB 13 32 MiB 0
> 61 TiB
> nativesqlbackup 24 32 539 MiB 147 1.6 GiB 0
> 61 TiB
> default.rgw.buckets.non-ec 25 32 1.7 MiB 0 5.0 MiB 0
> 61 TiB
> ceph-fs_sql_backups 26 32 0 B 0 0 B 0
> 61 TiB
> ceph-fs_sql_backups_metadata 27 32 0 B 0 0 B 0
> 61 TiB
> dd-drs-backups 28 32 0 B 0 0 B 0
> 61 TiB
> default.rgw.jv-corp-pool.data 59 32 16 TiB 63.90M 49 TiB 21.12
> 61 TiB
> default.rgw.jv-corp-pool.index 60 32 108 GiB 1.19k 323 GiB 0.17
> 61 TiB
> default.rgw.jv-corp-pool.non-ec 61 32 0 B 0 0 B 0
> 61 TiB
> default.rgw.jv-comm-pool.data 62 32 8.1 TiB 44.20M 24 TiB 11.65
> 61 TiB
> default.rgw.jv-comm-pool.index 63 32 83 GiB 811 248 GiB 0.13
> 61 TiB
> default.rgw.jv-comm-pool.non-ec 64 32 0 B 0 0 B 0
> 61 TiB
> default.rgw.jv-va-pool.data 65 32 4.8 TiB 22.17M 14 TiB 7.28
> 61 TiB
> default.rgw.jv-va-pool.index 66 32 38 GiB 401 113 GiB 0.06
> 61 TiB
> default.rgw.jv-va-pool.non-ec 67 32 0 B 0 0 B 0
> 61 TiB
> jv-edi-pool 68 32 0 B 0 0 B 0
> 61 TiB
>
> -- Michael
>
> -----Original Message-----
> From: Anthony D'Atri <[email protected]>
> Sent: Wednesday, March 20, 2024 2:48 PM
> To: Michael Worsham <[email protected]>
> Cc: [email protected]
> Subject: Re: [ceph-users] Need easy way to calculate Ceph cluster
> space for SolarWinds
>
> This is an external email. Please take care when clicking links or opening
> attachments. When in doubt, check with the Help Desk or Security.
>
>
>> On Mar 20, 2024, at 14:42, Michael Worsham <[email protected]>
>> wrote:
>>
>> Is there an easy way to poll a Ceph cluster to see how much space is
>> available
>
> `ceph df`
>
> The exporter has percentages per pool as well.
>
>
>> and how much space is available per bucket?
>
> Are you using RGW quotas?
>
>>
>> Looking for a way to use SolarWinds to monitor the entire Ceph cluster space
>> utilization and then also be able to break down each RGW bucket to see how
>> much space it was provisioned for and how much is available.
>
> RGW buckets do not provision space. Optionally there may be some RGW quotas
> but they're a different thing than you're implying.
>
>
>>
>> -- Michael
>>
>>
>> Get Outlook for Android<https://aka.ms/AAb9ysg> This message and its
>> attachments are from Data Dimensions and are intended only for the use of
>> the individual or entity to which it is addressed, and may contain
>> information that is privileged, confidential, and exempt from disclosure
>> under applicable law. If the reader of this message is not the intended
>> recipient, or the employee or agent responsible for delivering the message
>> to the intended recipient, you are hereby notified that any dissemination,
>> distribution, or copying of this communication is strictly prohibited. If
>> you have received this communication in error, please notify the sender
>> immediately and permanently delete the original email and destroy any copies
>> or printouts of this email as well as any attachments.
>> _______________________________________________
>> ceph-users mailing list -- [email protected] To unsubscribe send an
>> email to [email protected]
>
> This message and its attachments are from Data Dimensions and are intended
> only for the use of the individual or entity to which it is addressed, and
> may contain information that is privileged, confidential, and exempt from
> disclosure under applicable law. If the reader of this message is not the
> intended recipient, or the employee or agent responsible for delivering the
> message to the intended recipient, you are hereby notified that any
> dissemination, distribution, or copying of this communication is strictly
> prohibited. If you have received this communication in error, please notify
> the sender immediately and permanently delete the original email and destroy
> any copies or printouts of this email as well as any attachments.
>
This message and its attachments are from Data Dimensions and are intended only
for the use of the individual or entity to which it is addressed, and may
contain information that is privileged, confidential, and exempt from
disclosure under applicable law. If the reader of this message is not the
intended recipient, or the employee or agent responsible for delivering the
message to the intended recipient, you are hereby notified that any
dissemination, distribution, or copying of this communication is strictly
prohibited. If you have received this communication in error, please notify the
sender immediately and permanently delete the original email and destroy any
copies or printouts of this email as well as any attachments.
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]