Can you use this switch? You won't lose a port for uplink since it has fiber and/or copper uplink ports.
Just my 10 cents... Forgot to paste the link: http://www.bestbuy.com/site/olspage.jsp?skuId=8891915&type=product&id=1212192931527&ref=06&loc=01&ci_src=14110944&ci_sku=8891915 -----Original Message----- From: beowulf-boun...@beowulf.org [mailto:beowulf-boun...@beowulf.org] On Behalf Of Jaime Requinton Sent: Wednesday, September 09, 2009 3:12 PM To: Mike Davis; psc Cc: beowulf@beowulf.org Subject: RE: [Beowulf] how large can we go with 1GB Ethernet? / Re: how large of an installation have people used NFS, with? Can you use this switch? You won't lose a port for uplink since it has fiber and/or copper uplink ports. Just my 10 cents... -----Original Message----- From: beowulf-boun...@beowulf.org [mailto:beowulf-boun...@beowulf.org] On Behalf Of Mike Davis Sent: Wednesday, September 09, 2009 2:10 PM To: psc Cc: beowulf@beowulf.org Subject: Re: [Beowulf] how large can we go with 1GB Ethernet? / Re: how large of an installation have people used NFS, with? psc wrote: > I wonder what would be the sensible biggest cluster possible based on > 1GB Ethernet network . And especially how would you connect those 1GB > switches together -- now we have (on one of our four clusters) Two 48 > ports gigabit switches connected together with 6 patch cables and I just > ran out of ports for expansion and wonder where to go from here as we > already have four clusters and it would be great to stop adding cluster > and start expending them beyond number of outlets on the switch/s .... > NFS and 1GB Ethernet works great for us and we want to stick with it , > but we would love to find a way how to overcome the current "switch > limitation". ... I heard that there are some "stackable switches" .. > in any case -- any idea , suggestion will be appreciated. > > thanks!! > psc > > When we started running clusters in 2000 we made the decision to use a flat networking model and a single switch if at all possible, We use 144 and 160 port Gig e switches for two of our clusters. The overall performance is better and the routing less complex. Larger switches are available as well. We try to go with a flat model as well for Infiniband. Right now we are using a 96 port Infiniband switch. When we additional nodes to that cluster we will either move up to a 144 or 288 port chassis. Running the numbers I found the cost of the large chassis to be on par with the extra switches required to network using 24 or 36 port switches. -- Mike Davis Technical Director (804) 828-3885 Center for High Performance Computing jmdav...@vcu.edu Virginia Commonwealth University "Never tell people how to do things. Tell them what to do and they will surprise you with their ingenuity." George S. Patton _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf _______________________________________________ Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf