| AMBER Archive (2008)Subject: RE: AMBER: gigabit ethernet cards and switches
From: Ross Walker (ross_at_rosswalker.co.uk)Date: Thu Nov 06 2008 - 12:37:04 CST
 
 
 
 
Hi Alessandro,
 > We are considering to buy a new 'low-cost' computer cluster dedicated
> to MD simulations. We already have two clusters interconnected by
 > gigabit ethernet. One of these is composed by intel quad-core
 > processors and, as discussed many times in this list, suffers from
 > scaling problems when running Amber in more than two nodes.
 >
 > That said, I would like to ask to users and developers about good
 > choices to be considered when buying gigabit Ethernet cards and
 > switches. My dream is to convince my PI to upgrade myrinet, but price
 > is also an important issue.
 
 In the 'good old' days when gigabit ethernet was still somewhat usable as an
interconnect I would have advised you to go for good quality cards and a top
 notch fully non-blocking switch but this advice doesn't really apply anymore
 since even in this scenario it doesn't really help with running in parallel.
 It can help a little bit with NFS say but to be honest most cheap gigabit
 switches even if not non-blocking will still outstrip disk performance so I
 would suggest going for a fairly cheap switch and cards that are compatible
 with Linux. This is probably the key here - making sure they are supported
 under Linux. I would recommend the Intel e1000 chip based cards which are
 effectively guaranteed to work under Linux.
 
 As for try to at least get some scalability - so you can run 2 node jobs
something you could try (given that extra network cards are cheap) is to buy
 2 cards for each machine. Get yourself 1 cheap switch and hook up one set of
 the cards to give you NFS, ssh access / management facilities etc. Then with
 the second card in each machine use cross over cables to hook up each pait
 of machines. E.g.
 
 Mach 1 eth0 -----------------------
|
 |
 Mach 2 eth0 -------------------   |
 |   |
 Switch ------- File Server / login node
 |   |
 Mach 3 eth0 ------------------    |
 |
 |
 Mach 4 eth0 -----------------------
 
 Mach 1 eth1  ---X---  Mach2 eth1
 Mach 3 eth1  ---X---  Mach4 eth1
 Where X is a cross over cable. Then you just need to assign a different
private subnet to the eth1 cards and make sure your machine miles are
 modified as necessary to send MPI traffic over the corssover cables.  This
 way you should be able to run 8 way jobs (as 2 x quad core) for only a small
 amount of extra outlay and I think the performance would be better than
 trying to run 2 node jobs over the switch. You would of course need to tweak
 your queuing software to hand out nodes in the correct pairs and also create
 the correct machine files but that shouldn't be too hard.
 
 Note though I have not actually tried this (not for AMBER anyway, just for
doing private NFS mounts / backups etc) but I figure it should work quite
 well - the thing I don't know is how well things will scale over the
 crossover cable and whether it is worth it based on this so you might want
 to just get a cross over cable and try it quickly on some of your existing
 machines.
 
 Anyway, just my 3c.
 Good luck,
Ross
 
 /\
\/
 |\oss Walker
 
 | Assistant Research Professor |
| San Diego Supercomputer Center |
 | Tel: +1 858 822 0854 | EMail:- ross_at_rosswalker.co.uk |
 | http://www.rosswalker.co.uk | PGP Key available on request |
 
 Note: Electronic Mail is not secure, has no guarantee of delivery, may not
be read every day, and should not be used for urgent or sensitive issues.
 
 -----------------------------------------------------------------------
The AMBER Mail Reflector
 To post, send mail to amber_at_scripps.edu
 To unsubscribe, send "unsubscribe amber" (in the *body* of the email)
 to majordomo_at_scripps.edu
 
 
 
 |