Re: [Beowulf] stace_analyzer.pl can't work.

2008-09-15 Thread Mark Hahn
It throws an error likes:Argument "" isn't numeric in addition (+) at ./strace_analyzer.pl line 310, <> line 409105. Have you experienced this? and resolved? I think the parser could be improved. in this case, it seems to be choking on lines like: 23:34:13.502857 read(7, 0xbfffc670, 32)

Re: [Beowulf] 10gig CX4 switches

2008-09-15 Thread Scott Atchley
On Sep 15, 2008, at 8:38 PM, Joe Landman wrote: Greg Lindahl wrote: I have a bunch of 1gig switches with CX4 10gig uplinks (and empty X2 ports) and it's time to buy a 10gig switch. Has anyone done a recent survey of the market? I don't need any layer-3 features, just layer-2. I see that HP h

Re: [Beowulf] 10gig CX4 switches

2008-09-15 Thread Bill Wichser
Have you looked at Fujitsu? Anything using the Fulcrum chip (this one does) will have the best latency numbers. And their switches are just layer 2. Bill Greg Lindahl wrote: I have a bunch of 1gig switches with CX4 10gig uplinks (and empty X2 ports) and it's time to buy a 10gig switch. Has

Re: [Beowulf] 10gig CX4 switches

2008-09-15 Thread Joe Landman
Greg Lindahl wrote: I have a bunch of 1gig switches with CX4 10gig uplinks (and empty X2 ports) and it's time to buy a 10gig switch. Has anyone done a recent survey of the market? I don't need any layer-3 features, just layer-2. I see that HP has a 6-port switch for ~ $4k, too small. Arastra lo

[Beowulf] 10gig CX4 switches

2008-09-15 Thread Greg Lindahl
I have a bunch of 1gig switches with CX4 10gig uplinks (and empty X2 ports) and it's time to buy a 10gig switch. Has anyone done a recent survey of the market? I don't need any layer-3 features, just layer-2. I see that HP has a 6-port switch for ~ $4k, too small. Arastra looks nice, except that

Re: [Beowulf] Lustre failover

2008-09-15 Thread Bernd Schubert
On Wednesday 10 September 2008 13:41:18 andrew holway wrote: > >From the Lustre manual:- > > With OST servers it is possible to have a load-balanced active/active > configuration. > Each node is the primary node for a group of OSTs, and the failover > node for other > groups. To expand the simple t

Re: [Beowulf] Re: GPU boards and cluster servers

2008-09-15 Thread Jonathan Dursi
2008/9/10 Jon Forrest <[EMAIL PROTECTED]>: > > What I'm going to try to do is to be able to show > the faculty and grad students around here how > easy it is to get a significant performance improvement > by using CUDA as compared to using their normal > i386 or x86_64 processors. So I agree with

Re: [Beowulf] Lustre failover

2008-09-15 Thread Bernd Schubert
On Wednesday 10 September 2008 15:02:17 Mark Hahn wrote: > > With OST servers it is possible to have a load-balanced active/active > > configuration. > > Each node is the primary node for a group of OSTs, and the failover > > node for other > > ... > > > Anyone done this on a production system? > >

[Beowulf] stace_analyzer.pl can't work.

2008-09-15 Thread Eric.L
Hi,Jeffery I've read the article you've post in LinuxMagazine: http://www.linux-mag.com/id/6711. I believe that the tool stace_analyzer.pl will be very useful,but I've got some problems. It throws an error likes:Argument "" isn't numeric in addition (+) at ./strace_analyzer.pl line 310

Re: [Beowulf] Re: GPU boards and cluster servers

2008-09-15 Thread Jonathan Dursi
2008/9/10 Jonathan Dursi <[EMAIL PROTECTED]>: > If you haven't found it already, gpgpu and the forumns there can be Uh, that should be gpgpu.org. ( and forums). Jonathan -- Jonathan Dursi [EMAIL PROTECTED] ___ Beowulf mailing list, Beowulf@beowulf.

[Beowulf] Re: Lustre failover

2008-09-15 Thread Greg Keller
Re: Lustre failover I've worked on a number of large'sh lustre configs over the years, and all of them have been configured with Active/Active type mappings. There are a few issues being confused here: 1) Active/Active does not mean both OSS are accessing the same luns at the same time.

Re: [Beowulf] Lustre failover

2008-09-15 Thread Bernd Schubert
On Wednesday 10 September 2008 16:34:45 Mark Hahn wrote: > >> active/active seems strange to me - it implies that the bottleneck > >> is the OSS (OST server), rather than the disk itself. and a/a means > >> each OSS has to do more locking for the shared disk, which would seem > >> to make the prob

[Beowulf] ethernet bonding performance comparison "802.3ad" vs Adaptive Load Balancing

2008-09-15 Thread Rahul Nabar
I was experimenting with using channel bonding my twin eth ports to get a combined bandwidth of (close to) 2 Gbps. The two relevant modes were 4 (802.3ad) and 6 (alb=Adaptive Load Balancing). I was trying to compare performance for both. Before running any sophisticated tests by netperf etc. I jus

[Beowulf] Hijacking topics

2008-09-15 Thread Maurice Hilarius
This is classic for this list. The topic: "Re: Re: GPU boards and cluster servers." Gets turned into a discourse on Dell hardware and related. Meanwhile, the (useful) question: "Subject: [Beowulf] Q: AMD Opteron (Barcelona) 2356 vs Intel Xeon 5460 Very likely a hopeless question, with this little