I wouldn't normally top post, but all I have is a generic question. Do you have a swap partition setup? (I'd use something like 6-8Gbytes for a 4Gbyte system.)
rick On Tue, Apr 22, 2025 at 8:23 AM Sulev-Madis Silber <freebsd-current-freebsd-org...@ketas.si.pri.ee> wrote: > > well i don't have those errors anymore so there's nothing to give > > i've tried to tune arc but it didn't do anything so i took those things off > again > > right now i'm looking at > > ARC: 1487M Total, 1102M MFU, 128M MRU, 1544K Anon, 56M Header, 199M Other > 942M Compressed, 18G Uncompressed, 19.36:1 Ratio > > and wonder wtf > > i bet there's issue somewhere and i somehow can't properly recreate it. on > memory pressure it does resize arc down properly so seems like i don't need > any limits > > and there's no tmpfs. it would be useless at that low memory sizes > > the problem is that i can't figure out what all those problems are, how to > recreate those conditions and how to workaround or maybe find bugs. also > don't have enough hw to solely test it on. unless i can maybe try it on tiny > 512m vm. and then i would need to know what to try > > i also don't know why those git settings help me: > > [core] > packedGitWindowSize = 32m > packedGitLimit = 128m > preloadIndex = false > [diff] > renameLimit = 16384 > > how to tune it from some global place. and so on. and why it would even need > fiddiling so much? zfs indeed has improved a lot, previously it was quite a > hell to use > > i don't even know if this is related to mmap. even then, i don't really get > what that function even does. hence then "zfs (?) issue". it might even not > be zfs at all > > there are probably multiple combined issues here > > i also don't really buy the idea that ton of ram would automatically fix this > > so yeah unsure what to think of this > > some of the issues i found that others also have. some of them seem new > > some fixes were like as if trial and errors and nobody seemed to know what's > wrong even. granted, that was forum so maybe here it's better here? > > i mean i have used below average equipment my entire life and usual case to > cope with this is to just give it more time. put more swap and just wait > > i think someone tested my git issues in 4g vm and found no issues at all? > other things seem like as i only i have them > > i also find kind of confusing that if this is hw, why i don't see any other > issues > > this is not the first time that i have found something confusing in fbsd that > later turned out to be bug and was further tested and fixed by other > > hence the current mailing list so maybe someone else has ideas. or if it has > already fix. and i hope there are people with much larger labs and could > easily tell / test things > > so in the end, > > 1) why should git on large repo cause machine to run out of memory, instead > of just being as slow as it would need to be > > 2) why / what are fs operations that could cause low power machine to > mysteriously fail on zfs, when expected results would be slow fs behaviour > > i don't know what really happens and it's way too complex me to get all > memory management that happens in kernel. i only have this wild guess that > any type of caching should happen in "leftover" ram and make things faster if > possible. and any fs operations that have already reported completed by > kernel can't be suddenly found incomplete later. whatever that fs-related > stray buildworld error was that resolved itself somehow. and what i can > recreate > > and i'm not expert in this so how do i even know? > > what's fun is how running rsync over several tb's of data doesn't seem to > cause any issues at all. this is still same machine, many would not recommend > using this. different workload? > > hell knows what's all this. maybe later i could figure it out or actually > save some logs or. those i didn't save as i assumed it repeats itself. didn't > and it went off tmux window history > > oh well. yes, this is questionable report but those are "heisenbugs" as well. > at least some? > > > > On April 22, 2025 3:52:11 PM GMT+03:00, Ronald Klop <ronald-li...@klop.ws> > wrote: > >Hi, > > > >First, instead of writing "it gives vague errors", it really helps others on > >this list if you can copy-paste the errors into your email. > > > >Second, as far as I can see FreeBSD 13.4 uses OpenZFS 2.1.14. FreeBSD 14 > >uses OpenZFS 2.2.X which has bugfixes and improved tuning, although I cannot > >claim that will fix your issues. > >What you can try is to limit the growth of the ARC. > > > >Set "sysctl vfs.zfs.arc_max=1073741824" or add this to /etc/sysctl.conf to > >set the value at boot. > > > >This will limit the ARC to 1GB. I used similar settings on small machines > >without really noticing a speed difference while usability increased. You > >can play a bit with the value. Maybe 512MB will be even enough for your use > >case. > > > >NB: sysctl vfs.zfs.arc_max was renamed to vfs.zfs.arc.max with arc_max as a > >legacy alias, but I don't know if that already happened in 13.4. > > > >Another thing to check is the usage of tmpfs. If you don't restrict the max > >size of a tmpfs filesystem it will compete for memory. Although this will > >also show an increase in swap usage. > > > >Regards, > >Ronald. > > > > > >Van: Sulev-Madis Silber <freebsd-current-freebsd-org...@ketas.si.pri.ee> > >Datum: maandag, 21 april 2025 03:25 > >Aan: freebsd-current <freebsd-current@freebsd.org> > >Onderwerp: zfs (?) issues? > >> > >> i have long running issue in my 13.4 box (amd64) > >> > >> others don't get it at all and only suggest adding more than 4g ram > >> > >> it manifests as some mmap or other problems i don't really get > >> > >> basically unrestricted git consumes all the memory. i had to turn watchdog > >> on because something a git pull on ports tree causes kernel to take 100% > >> of ram. it keeps killing userland off until it's just kernel running there > >> happily. it never panics and killing off userland obviously makes the > >> problem disappear and nothing will do any fs operations anymore > >> > >> dovecot without tuning or with some tuning tended to do this too > >> > >> what is it? > >> > >> now i noticed another issue. if i happen to do too many src git pulls in a > >> row, they never actually "pull" anything. and / or clean my obj tree out. > >> i can't run buildworld anymore. it gives vague errors > >> > >> if i wait a little before starting buildworld, it always works > >> > >> what could possibly happening here? the way the buildworld fails means > >> there's serious issue with fs. and how could it be fixed with waiting? it > >> means that some fs operations are still going on in background > >> > >> i have no idea what's happening here. zfs doesn't report any issues. nor > >> do storage. nothing was killed with out of memory but arc usage somehow > >> increased a lot. and it's compression ratio went weirdly high, like ~22:1 > >> or so > >> > >> i don't know if it's acceptable zfs behaviour if it runs low on memory or > >> not. how to test it. etc. and if this is fixed on 14, on stable, or on > >> current. i don't have enough hw to test it on all > >> > >> i have done other stuff on that box that might also improper for amoung of > >> ram i have there but then it's just slow, nothing fails like this > >> > >> unsure how this could be fixed or tuned or something else. or why does it > >> behave like this. as opposed to usual low resource issues that just mean > >> you need more time > >> > >> i mean it would be easy to add huge amounts of ram but people could also > >> want to use zfs in slightly less powerful embedded systems where lack of > >> power is expected but weird fails maybe not > >> > >> so is this a bug? a feature? something fixed? something that can't be > >> fixed? what could be acceptable ram size? 8g? 16g? and why can't it just > >> tune everything down and become slower as expected > >> > >> i tried to look up on any openzfs related bugs but zfs is huge and i'm not > >> fs expert either > >> > >> i also don't know what happens while i wait. it doesn't show any serious > >> io load. no cpu is taken. load is down. system is responsible > >> > >> it all feels like bug still > >> > >> i have wondered if this is second hand hw acting up but i checked and > >> tested it as best as i could and why would it only bug out when i try more > >> complex things on zfs? > >> > >> i'm curious about using zfs on super low memory systems too, because it > >> offers certain features. maybe we could fix this if whole issue is ram. or > >> if it's elsewhere, maybe that too > >> > >> i don't know what to think of this all. esp the last issue. i'm not really > >> alone here with earlier issues but unsure > >> > >> > >> > > >