Re: [Beowulf] Clearing out scratch space

2018-06-12 Thread Nick Evans
at {$job -1} we used local scratch and tmpwatch. This had a wrapper script that would exclude files and folders for any user that currently running a job on the node. This way nothing got removed until the users job had finished even if they hadn't accessed the files for a while and you don't have

Re: [Beowulf] Cluster Authentication (LDAP,NIS,AD)

2017-12-27 Thread Nick Evans
Hi Robert, We are currently running our HPC, servers and desktops with storage needs serviced by an Isilon. We have CIFS and NFS capabilities both of which use the AD for authentication. Currently our cluster is Centos 6.8 NFS and SSH authenticating off of the AD using SSSD. We also have a number

Re: [Beowulf] slow jobs when run through queue

2017-12-19 Thread Nick Evans
I completely agree. We have a web page where people can see - where their jobs are running - what sort of resources were requested - the peak resources actually used - wall time remaining (orange highlighted at 20% remaining and red at 10% remaining) On 20 December 2017 at 03:41,

Re: [Beowulf] slow jobs when run through queue

2017-12-06 Thread Nick Evans
Hi all, Under normal circumstances I would agree that the request for all 48 cores in the machine is overkill but this particular machine has a highly specialised FPGA card in it to do most of the heavy lifting when running a specific set of analysis that has been tuned to run with the card. It ca

Re: [Beowulf] slow jobs when run through queue

2017-12-05 Thread Nick Evans
On 6 December 2017 at 12:58, Chris Samuel wrote: > On 6/12/17 11:44 am, Nick Evans wrote: > > We have found that if we submit a job to the queue then it takes a long >> time to process. ie. >4 hours >> If we are to run the exact same processing directly on the compute node >&

[Beowulf] slow jobs when run through queue

2017-12-05 Thread Nick Evans
HI All, Just wondering if anyone else has encountered a similar problem or thoughts on how they would try and track down the problem. Queue = PBS / Moab combination We have found that if we submit a job to the queue then it takes a long time to process. ie. >4 hours If we are to run the exact sa

Re: [Beowulf] Diagnosing Discovery issue xCat

2015-08-31 Thread Nick Evans
","SNMPv1","Cisco",, I did origionaly have just "2c" for the snmpversion and have now tried 1, 2c, SNMPv1, SNMPv2c... All with now luck. Will have to get wire shark onto it and find out what is happening. Thanks Nick On 1 September 2015 at 13:43, Christop

[Beowulf] Diagnosing Discovery issue xCat

2015-08-31 Thread Nick Evans
Hi All, I am sure i am just doing something silly as i haven't had an issue in the past getting nodes discovered via the switch port lookup method. Currently the newly booting node goes through the following steps Get IP Get the "xcat/xnba.kpxe" file Download the Genisis discovery environment an

Re: [Beowulf] 48-port 10gig switches?

2010-09-08 Thread Nick Evans
> > I could not find a price for the Voltaire, > but it isn't going to be cheap. Somewhere above $20K maybe? > > I saw in one of their PDF's that is $500 a port which i have worked out to $24K so you were close in your guess ___ Beowulf mailing list, Be