Can you tell if the number of objects increases in your cephfs between those bursts? I noticed something similar in a 16.2.15 cluster as well. It's not that heavily used, but it contains home directories and development working directories etc. And when one user checked out a git project, the mds memory usage increased a lot, getting near its configured limit. Before there were around 3,7 Million objects in the cephfs, that user added more than a million more files with his checkout. It wasn't a real issue (yet) because the usage isn't very dynamical and the total number of files is relatively stable. This doesn't really help resolve anything, but if your total number of files grows, I'm not surprised that the mds requires more memory.

Zitat von Alexander Patrakov <[email protected]>:

As a workaround, to reduce the impact of the MDS slowed down by
excessive memory consumption, I would suggest installing earlyoom,
disabling swap, and configuring earlyoom as follows (usually through
/etc/sysconfig/earlyoom, but could be in a different place on your
distribution):

EARLYOOM_ARGS="-p -r 600 -m 4,4 -s 1,1"

On Sat, Aug 31, 2024 at 3:44 PM Sake Ceph <[email protected]> wrote:

Ow it got worse after the upgrade to Reef (was running Quincy). With Quincy the memory usage was also a lot of times around 95% and some swap usage, but never exceeding both to the point of crashing.

Kind regards,
Sake
> Op 31-08-2024 09:15 CEST schreef Alexander Patrakov <[email protected]>:
>
>
> Got it.
>
> However, to narrow down the issue, I suggest that you test whether it
> still exists after the following changes:
>
> 1. Reduce max_mds to 1.
> 2. Do not reduce max_mds to 1, but migrate all clients from a direct
> CephFS mount to NFS.
>
> On Sat, Aug 31, 2024 at 2:55 PM Sake Ceph <[email protected]> wrote:
> >
> > I was talking about the hosts where the MDS containers are running on. The clients are all RHEL 9.
> >
> > Kind regards,
> > Sake
> >
> > > Op 31-08-2024 08:34 CEST schreef Alexander Patrakov <[email protected]>:
> > >
> > >
> > > Hello Sake,
> > >
> > > The combination of two active MDSs and RHEL8 does ring a bell, and I
> > > have seen this with Quincy, too. However, what's relevant is the
> > > kernel version on the clients. If they run the default 4.18.x kernel
> > > from RHEL8, please either upgrade to the mainline kernel or decrease
> > > max_mds to 1. If they run a modern kernel, then it is something I do
> > > not know about.
> > >
> > > On Sat, Aug 31, 2024 at 1:21 PM Sake Ceph <[email protected]> wrote:
> > > >
> > > > @Anthony: it's a small virtualized cluster and indeed SWAP shouldn't be used, but this doesn't change the problem.
> > > >
> > > > @Alexander: the problem is in the active nodes, the standby replay don't have issues anymore.
> > > >
> > > > Last night's backup run increased the memory usage to 86% when rsync was running for app2. It dropped to 77,8% when it was done. When the rsync for app4 was running it increased to 84% and dropping to 80%. After a few hours it's now settled on 82%. > > > > It looks to me the MDS server is caching something forever while it isn't being used..
> > > >
> > > > The underlying host is running on RHEL 8. Upgrade to RHEL 9 is planned, but hit some issues with automatically upgrading hosts.
> > > >
> > > > Kind regards,
> > > > Sake
> > > > _______________________________________________
> > > > ceph-users mailing list -- [email protected]
> > > > To unsubscribe send an email to [email protected]
> > >
> > >
> > >
> > > --
> > > Alexander Patrakov
>
>
>
> --
> Alexander Patrakov
> _______________________________________________
> ceph-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]



--
Alexander Patrakov
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]


_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to