Changed to infinity and the error seems gone. Will keep monitoring for a few days ..
Thanks guys! Rao On Tue, Feb 21, 2017 at 3:42 AM, Lennart Poettering <[email protected]> wrote: > On Mon, 20.02.17 16:44, Rao Vz ([email protected]) wrote: > > > Hi, Guys > > > > We have a Apache Spark cluster of 3 nodes, one is master and slave, the > > other two are slaves. When starting Spark worker with "systemctl start > > spark-worker", when running out apps, sometimes but not always it > generates > > "java.lang.OutOfMemoryError: unable to create new native thread" error in > > Spark worker logs. > > I figure the error is misleading and is not about memory at all, and > you need to bump the default TasksMax= field or even turn it off by > setting it to infinity. > > Lennart > > -- > Lennart Poettering, Red Hat >
_______________________________________________ systemd-devel mailing list [email protected] https://lists.freedesktop.org/mailman/listinfo/systemd-devel
