Looking at my question I posted earlier how is it that java is so high 
performing when it comes to large data sets?

Regards,
Jonathan

From: Beowulf <beowulf-boun...@beowulf.org> On Behalf Of Oddo Da
Sent: 13 October 2020 14:38
To: Michael Di Domenico <mdidomeni...@gmail.com>
Cc: Beowulf Mailing List <beowulf@beowulf.org>
Subject: Re: [Beowulf] [External] Spark, Julia, OpenMPI etc. - all in one place

On Tue, Oct 13, 2020 at 8:33 AM Michael Di Domenico 
<mdidomeni...@gmail.com<mailto:mdidomeni...@gmail.com>> wrote:
i can't speak from a general industry sense, but i've had everything
run through my center over the past 11 years.  Hadoop seemed like
something that was going to take off.  it didn't with my group of
users.  we aren't counting clicks nor parsing text from huge files, so
its utility to us faded.  my understanding is the group behind hadoop
also made several industry missteps when trying to commercialize, i'm
not sure what happened after that.  i think a lot people realized that
hadoop made things easier, but the overhead was too high given the
limited functionality most people wanted to use it for

Michael, thank you for the insight. I think Hadoop in general is mostly dying, 
Spark is really the derivative that took off. Basically, what you are saying is 
that there is no demand on your infra for this kind of work. Do you have any 
insights as to why not? Do the AI/DS/ML guys just know that they cannot use 
your resources to run standard loads and go straight to the cloud or local 
ethernet clusters?

In your estimate, how many of your users write code in Julia vs MPI vs Python?
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
https://beowulf.org/cgi-bin/mailman/listinfo/beowulf

Reply via email to