Dear R experts. I am running customized versions of nonlinear models in my package CatDyn. These are models with 140 parameters to estimate and composite likelihoods made of mixtures of adjusted profile normal, adjusted profile lognormal, and a robust version of the lognormal. There are 3^6 composite likelihoods, because of 3 core likelihoods and 6 agents acting to produce the data for the model, each one having one of the 3 likelihoods. The numerical methods I'm using are CG and spg, as these worked the best for these models in other, smaller optimization problems within the same set of models in CatDyn.
My motivation for this message is that the optimization is taking days for each of the 3^6 composite likelihoods on an Ubuntu 24.04 AMD Ryzen™ 7 8700G w/ Radeon™ 780M Graphics×16 with 128 GB RAM. I was expecting much faster optimization with 128 GB RAM. Some of you may have experience in running large nonlinear optimization problems in R. Is there any advice on how to speed up these rather large-ish optimization problems in R? Either software, hardware, or both? I apologize in advance if you consider this not a proper question for the mail list. Ruben --- Ruben H. Roa-Ureta, Ph. D. Consultant in Statistical Modeling ORCID ID 0000-0002-9620-5224 ______________________________________________ [email protected] mailing list -- To UNSUBSCRIBE and more, see https://stat.ethz.ch/mailman/listinfo/r-help PLEASE do read the posting guide https://www.R-project.org/posting-guide.html and provide commented, minimal, self-contained, reproducible code.

