DISCLOSURE: I work for AWS HPC Developer Relations in the services team. We 
developer AWS Batch, AWS ParallelCluster, NICE DCV, etc.

Lambda’s limits today are 128MB to 10,240MB (~10GB) and billed in 1MB per ms 
increments. 15 minute max runtime for the function invocation.

Would you all be interested in a hands-on self-paced workshop on creating (or 
porting) an application to serverless environment? E.g. Monte-Carlo simulation, 
a genome alignment or variant call, or some other problem? We have some basic 
data processing documentation but nothing that speaks to real-world HPC use 
case and that is a something I want to fill the gap on if folks are interested 
in it.

Dr. Denis Bauer at CSIRO is also doing interesting things with serverless.

-angel
--
Angel Pizarro | Principal Developer Advocate, HPC @ AWS

From: Beowulf <beowulf-boun...@beowulf.org> on behalf of Guy Coates 
<guy.coa...@gmail.com>
Date: Thursday, September 23, 2021 at 8:46 AM
To: Tim Cutts <t...@sanger.ac.uk>
Cc: Beowulf <beowulf@beowulf.org>
Subject: RE: [EXTERNAL] [Beowulf] Rant on why HPC isn't as easy as I'd like it 
to be. [EXT]


CAUTION: This email originated from outside of the organization. Do not click 
links or open attachments unless you can confirm the sender and know the 
content is safe.


Out of interest, how large are the compute jobs (memory, runtime etc)?  How 
easy to get them to fit into a serverless environment?

Thanks,

Guy

On Tue, 21 Sept 2021 at 13:02, Tim Cutts 
<t...@sanger.ac.uk<mailto:t...@sanger.ac.uk>> wrote:
I think that’s exactly the situation we’ve been in for a long time, especially 
in life sciences, and it’s becoming more entrenched.  My experience is that the 
average user of our scientific computing systems has been becoming less 
technically savvy for many years now.

The presence of the cloud makes that more acute, in particular because it makes 
it easy for the user to effectively throw more hardware at the problem, which 
reduces the incentive to make their code particularly fast or efficient.  Cost 
is the only brake on it, and in many cases I’m finding the PI doesn’t actually 
care about that.  They care that a result is being obtained (and it’s time to 
first result they care about, not time to complete all the analysis), and so 
they typically don’t have much time for those of us who are telling them they 
need to invest in time up front developing and optimising efficient code.

And cost is not necessarily the brake I thought it was going to be anyway.  One 
recent project we’ve done on AWS has impressed me a great deal.  It’s not 
terribly CPU efficient, and would doubtless, with sufficient effort, run much 
more efficiently on premise.  But it’s extremely elastic in its nature, and so 
a good fit for the cloud.   Once a week, the project has to completely 
re-analyse the 600,000+ COVID genomes we’e sequenced so far, looking for new 
branches in the phylogenetic tree, and to complete that analysis inside 8 
hours.   Initial attempts to naively convert the HPC implementation to run on 
AWS looked as though they were going to be very expensive (~$50k per weekly 
run).  But a fundamental reworking of the entire workflow to make it as cloud 
native as possible, by which I mean almost exclusively serverless, has 
succeeded beyond what I expected.  The total cost is <$5,000 a month, and 
because there is essentially no statically configured infrastructure at all, 
the security is fairly easy to be comfortable about.  And all of that was done 
with no detailed thinking about whether the actual algorithms running in the 
containers are at all optimised in a traditional HPC sense.  It’s just not 
needed for this particular piece of work.  Did it need software developers with 
hardcore knowledge of performance optimisation?  No.  Was it rapid to develop 
and deploy?  Yes.  Is the performance fast enough for UK national COVID variant 
surveillance?  Yes.  Is it cost effective?  Yes.  Sold!  The one thing it did 
need was knowledgeable cloud architects, but the cloud providers can and do 
help with that.

Tim

--
Tim Cutts
Head of Scientific Computing
Wellcome Sanger Institute


On 21 Sep 2021, at 12:24, John Hearns 
<hear...@gmail.com<mailto:hear...@gmail.com>> wrote:

Some points well made here. I have seen in the past job scripts passed on from 
graduate student to graduate student - the case I am thinking on was an Abaqus 
script for 8 core systems, being run on a new 32 core system. Why WOULD a 
graduate student question a script given to them - which works. They should be 
getting on with their science. I guess this is where Research Software 
Engineers come in.

Another point I would make is about modern processor architectures, for 
instance AMD Rome/Milan. You can have different Numa Per Socket options, which 
affect performance. We set the preferred IO path - which I have seen myself to 
have an effect on latency of MPI messages. IF you are not concerned about your 
hardware layout you would just go ahead and run, missing  a lot of performance.

I am now going to be controversial and common that over in Julia land the 
pattern seems to be these days people develop on their own laptops, or maybe 
local GPU systems. There is a lot of microbenchmarking going on. But there 
seems to be not a lot of thought given to CPU pinning or shat happens with 
hyperthreading. I guess topics like that are part of HPC 'Black Magic' - though 
I would imagine the low latency crowd are hot on them.

I often introduce people to the excellent lstopo/hwloc utilities which show the 
layout of a system. Most people are pleasantly surprised to find this.

-- The Wellcome Sanger Institute is operated by Genome Research Limited, a 
charity registered in England with number 1021457 and a company registered in 
England with number 2742969, whose registered office is 215 Euston Road, 
London, NW1 2BE.
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org<mailto:Beowulf@beowulf.org> sponsored 
by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
https://beowulf.org/cgi-bin/mailman/listinfo/beowulf


--
Dr. Guy Coates
+44(0)7801 710224

_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
https://beowulf.org/cgi-bin/mailman/listinfo/beowulf

Reply via email to