Would love to help you guys out in anyway i can in terms of hardware processing.
Have you guys thought of doing something like SETI@home and those projects to get idle compute power to help churn through the massive amounts of data? Regards, Jonathan ________________________________ From: Tim Cutts <t...@sanger.ac.uk> Sent: 04 February 2021 11:26 To: Jonathan Aquilina <jaquil...@eagleeyet.net> Cc: Beowulf <beowulf@beowulf.org> Subject: Re: [Beowulf] Project Heron at the Sanger Institute [EXT] On 4 Feb 2021, at 10:14, Jonathan Aquilina via Beowulf <beowulf@beowulf.org<mailto:beowulf@beowulf.org>> wrote: I am curious though to chunk out such large data is something like hadoop/HBase and the like of those platforms, are those whats being used? It’s a combination of our home-grown sequencing pipeline which we use across the board, and then a specific COG-UK analysis of the genomes themselves. This pipeline is common to all consortium members who are contributing sequence data. It’s a Nextflow pipeline, and the code is here: https://github.com/connor-lab/ncov2019-artic-nf Being nextflow, you can run it on anything for which nextflow has a backend scheduler. It supports data from both Illumina and Oxford Nanopore sequencers. Tim -- The Wellcome Sanger Institute is operated by Genome Research Limited, a charity registered in England with number 1021457 and a company registered in England with number 2742969, whose registered office is 215 Euston Road, London, NW1 2BE.
_______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit https://beowulf.org/cgi-bin/mailman/listinfo/beowulf