Re: [Beowulf] size of swap partition

2008-06-09 Thread Mark Hahn
We have the potential to have to swap whole jobs out of memory on a complete node. that was our intent as well. among other things, this scheme enables running the cluster "split-personality" - mostly shorter/smaller even interactive jobs during the day, with big/long jobs running at night. unf

Re: [Beowulf] User resource limits

2008-06-09 Thread Chris Samuel
- "Prentice Bisbal" <[EMAIL PROTECTED]> wrote: > I think David was assuming I was running Linux, and he was correct. > thanks for your help. I have to go read some man pages now. Be very aware that there are two different ulimits that affect memory allocations, *depending on the size of the

Re: [Beowulf] size of swap partition

2008-06-09 Thread Eric Thibodeau
Mikhail, Somewhat like Gerry said, ballpark figures have always been an arbitrary 1.5*RAM. This is completely ridiculous nowadays and should depend entirely on the applications you run. Typically, you should never swap out memory on a running application. I recommend you perform some m

Re: [Beowulf] size of swap partition

2008-06-09 Thread Gerry Creager
Misha, We have the potential to have to swap whole jobs out of memory on a complete node. As a result, I recommend 1.5-2.0 times memory in swap if this is a consideration. I do know there's likely to be a bit of discussion as this varies widely from site to site and based on requirements.

[Beowulf] size of swap partition

2008-06-09 Thread Mikhail Kuzminsky
A lot of time ago it was formulated simple rule for swap partition size (equal to main memory size). Currently we all have relative large RAM on the nodes (typically, I beleive, it is 2 or more GB per core; we have 16 GB per dual-socket quad-core Opteron node). What is typical modern swap size

Re: [Beowulf] User resource limits

2008-06-09 Thread Prentice Bisbal
Perry E. Metzger wrote: > "Lombard, David N" <[EMAIL PROTECTED]> writes: >> On Mon, Jun 09, 2008 at 11:41:29AM -0400, Prentice Bisbal wrote: >>> I would like to impose some CPU and memory limits on users that are hard >>> limits that can't be changed/overridden by the users. What is the best >>> wa

Re: [Beowulf] User resource limits

2008-06-09 Thread Perry E. Metzger
"Lombard, David N" <[EMAIL PROTECTED]> writes: > On Mon, Jun 09, 2008 at 11:41:29AM -0400, Prentice Bisbal wrote: >> >> I would like to impose some CPU and memory limits on users that are hard >> limits that can't be changed/overridden by the users. What is the best >> way to do this? All I know

Re: [Beowulf] User resource limits

2008-06-09 Thread Perry E. Metzger
Prentice Bisbal <[EMAIL PROTECTED]> writes: > I would like to impose some CPU and memory limits on users that are hard > limits that can't be changed/overridden by the users. What is the best > way to do this? All I know is environment variables or shell commands > done as the user (ulimit, for ex

Re: [Beowulf] User resource limits

2008-06-09 Thread Lombard, David N
On Mon, Jun 09, 2008 at 11:41:29AM -0400, Prentice Bisbal wrote: > > I would like to impose some CPU and memory limits on users that are hard > limits that can't be changed/overridden by the users. What is the best > way to do this? All I know is environment variables or shell commands > done as t

[Beowulf] User resource limits

2008-06-09 Thread Prentice Bisbal
This topic is slightly off topic, since it's not a beowulf specific problem, but it is HPC-related: I have several fat servers with 4 cores and 32 GB of RAM, for jobs that aren't very parallel and need large amounts of RAM. They are not clustered in any way. At the moment, users ssh into these sys

Re: [Beowulf] A couple of interesting comments

2008-06-09 Thread Matt Lawrence
On Mon, 9 Jun 2008, Ashley Pittman wrote: I can think of at least one cluster where the opposite has been true and PXE boot has been the default. The problem with this is if the head node PXE boots on the customers network and gets automatically re-installed as a windows workstation everybody g

Re: [Beowulf] A couple of interesting comments

2008-06-09 Thread Ashley Pittman
On Fri, 2008-06-06 at 10:39 -0500, Gerry Creager wrote: > > 2. BIOS had a couple of interesting defaults, including warn on > keyboard error (Keyboard? Not intentionally. This is a compute > node, > and should never require a keyboard. Ever.) We also find the BIOS > is > set to boot from

Re: [Beowulf] A couple of interesting comments

2008-06-09 Thread Chris Samuel
- "Tim Cutts" <[EMAIL PROTECTED]> wrote: > Wow. How many nodes were you buying? 95 nodes, each with two Barcelonas, so 760 cores all up. 32GB RAM (4GB/core) and 4x300GB SATA drives (RAID-0) per node. > And are we allowed to know who the vendor was? It's all public, so no reason why not.

Re: [Beowulf] A couple of interesting comments

2008-06-09 Thread Tim Cutts
Chris Samuel wrote: - "Tim Cutts" <[EMAIL PROTECTED]> wrote: On 9 Jun 2008, at 1:09 am, Chris Samuel wrote: Our most recent vendor went to the motherboard manufacturer and said "please can you cut us a BIOS with these default settings" and they did so. If you don't mind us

Re: [Beowulf] A couple of interesting comments

2008-06-09 Thread Chris Samuel
- "Tim Cutts" <[EMAIL PROTECTED]> wrote: > On 9 Jun 2008, at 1:09 am, Chris Samuel wrote: > > > Our most recent vendor went to the motherboard manufacturer > > and said "please can you cut us a BIOS with these default > > settings" and they did so. > > If you don't mind us asking, roughly ho