general
purpose cluster?
If there is no support possible with infiniband then can we plan for any
other high speed interconnect technology like Myrinet, Quadrics etc.
Regards,
Amjad Ali
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin
Hello All,
Is there any line/point to make distinction between accelerators and
co-processors (that are used in conjunction with the primary CPU to boost
up the performance)? or these terms can be used interchangeably?
Specifically, the word "accelerator" is used commonly with GPU. On the
other h
justifying this and comment/modify above two
justifications. Better if I you can suggent me to quote some reference of
any suitable publication in this regard.
best regards,
Amjad Ali
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin
Hi all,
What is the be the best way of using MPI persistent communication in an
iterative/repetative kind of code about calling MPI_Free(); Should we call
MPI_Free() in every iteration or
only once when all the iterations/repetitions are performed?
Means which one is the best out of following two:
ear, but Xeons get nearly stuck in terms of scaling
> when there are more than 2 processes (3 or 4) running in a single node.
>
> My two cents.
> Gus Correa
> -
> Gustavo Correa
> Lamont-Doherty Earth Observatory - Columbia University
> Palisades, NY, 10964-8000 - USA
>
Hi all,
I am usually running my parallel jobs on the university cluster of several
hundred nodes (accessible to all). So often I observe very different total
MPI_Wtime value when I run my program (even of same problem size) on
different occasion.
How should I make a reasonable performance measure
Hi all,
I have, with my group, a small cluster of about 16 nodes (each one with
single socket Xeon 3085 or 3110; And I face problem of poor scalability. Its
network is quite ordinary GiGE (perhaps DLink DGS-1024D 24-Port
10/100/1000), store and forward switch, of price about $250 only.
ftp://ftp10
, 2009 at 5:50 PM, David N. Lombard wrote:
> On Mon, Nov 30, 2009 at 01:24:34PM -0700, amjad ali wrote:
> > Hi,
> > Suppose we run a parallel MPI code with 64 processes on a cluster, say of
> 16 nodes. The cluster nodes has multicore CPU say 4 cores on each node.
> >
> &g
Hi,
Suppose we run a parallel MPI code with 64 processes on a cluster, say of 16
nodes. The cluster nodes has multicore CPU say 4 cores on each node.
Now all the 64 cores on the cluster running a process. Program is SPMD,
means all processes has the same workload.
Now if we had done auto-vectoriz
Hi,
while looking for high density (fat) nodes; one should keep in mind that
more number of cores striving simultaneously to access memory, cause memory
contention. Infact today CPU cores are fast and scalable but not the memory
bandwidth. Experts say that buying a CPU is infact buying "bandwidth"
Hi all,
Suppose a code is tuned on a specific system (e.g Intel Xeon based using
Vtune or Trace Collector). Then to how much extent this tuning will be
useful if this code is compiled and run on some other system (e.g. AMD
Opteron based)? Means whether the code tuning performed at one system
(poss
used (having arrays
declared fixed sizes/values)?
Thank You for your kind attention.
with best regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit
ht
Hi all and Thanks you all.
It is making quite good sense.
On Thu, Nov 12, 2009 at 8:50 AM, Michael H. Frese <
michael.fr...@numerex-llc.com> wrote:
> At 06:26 AM 11/12/2009, Stuart Barkley wrote:
>
>> At 03:40 PM 11/11/2009, Peter St. John wrote:
>> > The difference between:
>> > arr
benefit/efficiency?
Better if any body can refer to some tutorial/example-code on this.
Thank you for you attention.
With best regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription
each of these, Does MPI has the same (or nearly same)
overhead? or any significantly different overhead is involved in handling
each of these arrays (by MPI)?
with best regards,
Amjad Ali
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin
/portion in the recv_array.
Is this scheme is quite fine and correct.
I am in search of efficient one.
Request for help.
With best regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your
Hi,
please see below
On Wed, Sep 2, 2009 at 6:57 PM, Mark Hahn wrote:
> On Wed, 2 Sep 2009, amjad ali wrote:
>
> Hi All,
>> I have 4-Nodes ( 4 CPUs Xeon3085, total 8 cores) Beowulf cluster on
>> ROCKS-5
>> with GiG-Ethernet. I tested runs of a 1D CFD code both s
ication overhead! but how
to get better performance--lesser run time? My code is not too complicated
only 2 values are sent and 2 values are received by each process after each
stage.
Regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
way?
Does a program written for any one of the above will work fine on the other?
or we have to re-program for the other?
Regards.
On Sat, Aug 29, 2009 at 5:48 PM, wrote:
> On Sat, Aug 29, 2009 at 8:42 AM, amjad ali wrote:
> > Hello All,
> >
> >
> >
> &
these setups to get good efficiency?
Comments are welcome also for AMD/ATI FireStream.
With best regards,
AMJAD ALI.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe
partitioning routine from within the program run so that whatever -np
value would be we are at ease.
THANKS A LOT FOR YOUR ATTENTION.
Regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your
YOUR KIND ATTENTION.
With best regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf
display result online after 100 iterations/time-steps in my
parallel version so less "I/O" and/or "funneling I/O through master process"
will be required.
Any opinion/suggestion?
regards,
Amjad Ali.
On Wed, Jul 1, 2009 at 5:06 AM, Gus Correa wrote:
> Hi Bogdan, list
>
>
.
With best regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf
from the master process). Is it possible?.
Thank you very much.
Regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org
-machinefile option). Then I do not get correct result. The
convergence gets very much slow down after few iterations, ultimately
resulting in NaN values of problem variables.
I would be very grateful for having comments for the remedy of above
two difficulties/confusions.
with best regards,
Amjad Ali
deallaocated/freed automatically
on return from the subroutine?).
Thanks.
with best regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo
dd ssh key for the user but fails. Error
messages include:
cannot create /home/amjad/.ssh
and
/home/amjad/: no such file or directory.
What happened to my cluster? Solution?
regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
To change your
Hi,
Please reply me about followings:
1) Is there any significant performance difference between OpenMPI
built with GNU compilers and the one built with Intel compilers?
2) Once we have built mpi on our system and then we remove the C/C++
and Fortran compilers from the system. Will the mpicc an
More general term one could use is the "processing element".
For example if a cluster consists of
2 nodes with each having 1 single core CPU,
2 nodes with each having 1 dual core CPU,
2 nodes with each having 1 quad core CPU,
2 nodes with each having 2 dual core CPUs, and
2 nodes with each having 2
erhaps you could have easier simpler life.
Also have a look in using PETSc (that make use of MPI internally).
regards,
Amjad Ali.
On Fri, Feb 22, 2008 at 8:50 PM, John P. Kosky, PhD <[EMAIL PROTECTED]>
wrote:
> My company is taking it's first foray into the world of HPC with an
>
hours/days?
Which Board you recommend for this scenario?
Regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf
Hello all,
I want to develop and run my parallel code (MPI based) on a Beowulf cluster.
I have no problem as such that many user might log on to the cluster
simultaneously. Suppose that I am free to use cluster dedicatedly for my
single parallel application.
1) Do I really need a cluster schedule
Hello,
Because today the clusters with multicore nodes are quite common and the
cores within a node share memory.
Which Implementations of MPI (no matter commercial or free), make automatic
and efficient use of shared memory for message passing within a node. (means
which MPI librarries auomatica
y determined by the
speed of the memory, not the speed of the CPU.
regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf
Dual Core AMD Opterons, RAM 2 GB/core and
Infiniband is more suitable for these codes?
regards,
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/lis
Hi All,
Would any of you please like to share usage-experience/views/comments
about Windows Compute Cluster Server 2003 based Beowulf Clusters?
What in your opinion is the future of such clusters?
How you compare these with the LINUX CLUSTERS?
regards.
_
Hi All,
Which one of the following two choices are better (assuming both
clusters have nearly same cost):
1) A 4-node cluster having 8 AMD Opteron of 2.6GHz each and 2GB RAM/node
1) A 5-node cluster having 10 AMD Opteron of 2.2GHz each and 2GB RAM/node
Consider GiGE (max 1000mbps) as the inter
PSSC labs.
MOAB
Scali-Manage
Scyld Software.
regards,
AMJAD ALI.
___
Beowulf mailing list, Beowulf@beowulf.org
To change your subscription (digest mode or unsubscribe) visit
http://www.beowulf.org/mailman/listinfo/beowulf
oup.org
() Docs availaible at the website of ASPEN SYSTEMS
() "How to Build a Beowulf Linux Cluster" at teh website of The
Mississippi Center for Supercomputing Research
() "Configuring a Beowulf Cluster" by Forrest Hoffman, Extreme Li
for
having 2-CPU/node. As interconnect I opted GiGE (Net-Gear's GiGE
managed Switch with all ports 10/100/1000 Mbps).
Hope this will work in taking start.
Amjad Ali,
BZU, Mulatn, Pakistan.
___
Beowulf mailing list, Beowulf@beowulf.org
To change
cussion, I (opted) came up with the following design:
Each node having two AMD Opteron Dual Core processors and 4GB of main
memory on Tyan Thunder Board with GiGE as the interconnect.
regards,
AMJAD ALI.
BZU, Multan, Pakistan.
On 9/28/06, Peter Wainwright <[EMAIL PROTECTED]> wrote:
Pl
Hello, Hi All.
We are developing Beowulf Cluster for the first time at our university
department. We will perform numerical simulations in CFD on the
cluster. We have chosen that each node (master and compute) of the
cluster will be consisting of 2 AMD Dual Core Opteron Processors on
Tyan Thunder
tion about using
network interconnect technology?
Now i want to clear that parallel (CFD) code development how much
differs in this configuration than the configuration of single
processor on each node.
with best regards:
Amjad Ali.
___
Beowulf mailing lis
use
Gigabit ehternet switch and 1GB of RAM at each node.
Please guide me that how much parallel
programming will differ for the above four choices of processing
nodes.
with best regards:
Amjad Ali.
___
Beowulf mailing list, Beowulf@beowulf.org
To
45 matches
Mail list logo