Good Morning all,

I am working on a project that I sadly cant go into much detail but there will 
be quite large amounts of data that will be ingested by this system and would 
need to be efficiently returned as output to the end user in around 10 min or 
so. I am in discussions with another partner involved in this project about the 
best way forward on this.

For me given the amount of data (and it is a huge amount of data) that an RDBMS 
such as postgresql would be a major bottle neck. Another thing that was 
considered flat files, and I think the best for that would be a Hadoop cluster 
with HDFS. But in the case of HPC how can such an environment help in terms of 
ingesting and analytics of large amounts of data? Would said flat files of data 
be put on a SAN/NAS or something and through an NFS share accessed that way for 
computational purposes?

Regards,
Jonathan
_______________________________________________
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit 
http://www.beowulf.org/mailman/listinfo/beowulf

Reply via email to