[Beowulf] Configuring nodes on a scyld cluster
Andre Kerstens
akerstens at penguincomputing.com
Tue Aug 25 11:40:30 PDT 2009
Michael,
On a cluster running Scyld Clusterware (are you running 4 or 5?) there
is no need to install any Ganglia components on the compute nodes: the
compute nodes communicate cluster information incl. ganglia info to the
head node via the beostatus sendstats mechanism. If ganglia is not
enabled yet on your cluster, you can do it as follows:
Edit /etc/xinetd.d/beostat and change 'disable=yes' to 'disable=no'
followed by:
/sbin/chkconfig xinetd on
/sbin/chkconfig httpd on
/sbin/chkconfig gmetad on
and
service xinetd restart
service httpd start
service gemetad start
Then point your web browser to http://localhost/ganglia and off you go.
This information can be found in the release notes document of your
Scyld cluster or in the Scyld admin guide.
Cheers
Andre
------------------------------
Message: 2
Date: Mon, 24 Aug 2009 04:40:22 -0500
From: Michael Muratet <mmuratet at hudsonalpha.org>
Subject: [Beowulf] Configuring nodes on a scyld cluster
To: ganglia-general at lists.sourceforge.net
Cc: beowulf at beowulf.org
Message-ID: <93AC2CF8-3096-487E-BC08-FBC644C5C62C at hudsonalpha.org>
Content-Type: text/plain; charset=US-ASCII; format=flowed; delsp=yes
Greetings
I'm not sure if this is more appropriate for the beowulf or ganglia
list, please forgive a cross-post. I have been trying to get ganglia (v
3.0.7) to record info from the nodes of my scyld cluster. gmond was not
installed on any of the compute nodes nor was gmond.conf in /etc of any
of the compute nodes when we got it from the vendor. I didn't see much
in the documentation about configuring nodes but I did find a 'howto' at
http://www.krazyworks.com/installing-and-configuring-
ganglia/. I have been testing on one of the nodes as follows. I copied
gmond from /usr/sbin on the head node to the subject compute node /usr/
sbin. I ran gmond --default_config and saved the output and changed it
thus:
scyld:etc root$ bpsh 5 cat /etc/gmond.conf
/* This configuration is as close to 2.5.x default behavior as possible
The values closely match ./gmond/metric.h definitions in 2.5.x */
globals {
daemonize = yes
setuid = yes
user = nobody
debug_level = 0
max_udp_msg_len = 1472
mute = no
deaf = no
host_dmax = 0 /*secs */
cleanup_threshold = 300 /*secs */
gexec = no
}
/* If a cluster attribute is specified, then all gmond hosts are wrapped
inside
* of a <CLUSTER> tag. If you do not specify a cluster tag, then all
<HOSTS> will
* NOT be wrapped inside of a <CLUSTER> tag. */ cluster {
name = "mendel"
owner = "unspecified"
latlong = "unspecified"
url = "unspecified"
}
/* The host section describes attributes of the host, like the location
*/ host {
location = "unspecified"
}
/* Feel free to specify as many udp_send_channels as you like. Gmond
used to only support having a single channel */ udp_send_channel {
port = 8649
host = 10.54.50.150 /* head node's IP */ }
/* You can specify as many udp_recv_channels as you like as well. */
/* You can specify as many tcp_accept_channels as you like to share
an xml description of the state of the cluster */ tcp_accept_channel
{
port = 8649
}
I modified gmond on the head node thus:
/* This configuration is as close to 2.5.x default behavior as possible
The values closely match ./gmond/metric.h definitions in 2.5.x */
globals {
daemonize = yes
setuid = yes
user = nobody
debug_level = 0
max_udp_msg_len = 1472
mute = no
deaf = no
host_dmax = 0 /*secs */
cleanup_threshold = 300 /*secs */
gexec = no
}
/* If a cluster attribute is specified, then all gmond hosts are wrapped
inside
* of a <CLUSTER> tag. If you do not specify a cluster tag, then all
<HOSTS> will
* NOT be wrapped inside of a <CLUSTER> tag. */ cluster {
name = "mendel"
owner = "unspecified"
latlong = "unspecified"
url = "unspecified"
}
/* The host section describes attributes of the host, like the location
*/ host {
location = "unspecified"
}
/* Feel free to specify as many udp_send_channels as you like. Gmond
used to only support having a single channel */
/* You can specify as many udp_recv_channels as you like as well. */
udp_recv_channel {
port = 8649
}
/* You can specify as many tcp_accept_channels as you like to share
an xml description of the state of the cluster */ tcp_accept_channel
{
port = 8649
}
I started gmond on the compute node bpsh 5 gmond and restarted gmond and
gmetad. I don't see my node running gmond. ps -elf | grep gmond on the
compute node returns nothing. I tried to add gmond as a service on the
compute node with the script at the krazy site but I get:
scyld:~ root$ bpsh 5 chkconfig --add gmond service gmond does not
support chkconfig
and
scyld:~ root$ bpsh 5 service gmond start
/sbin/service: line 3: /etc/init.d/functions: No such file or directory
I am at a loss over what to try next, it seems this should work. Any and
all suggestions will be appreciated.
Thanks
Mike
Michael Muratet, Ph.D.
Senior Scientist
HudsonAlpha Institute for Biotechnology
mmuratet at hudsonalpha.org
(256) 327-0473 (p)
(256) 327-0966 (f)
Room 4005
601 Genome Way
Huntsville, Alabama 35806
------------------------------
More information about the Beowulf
mailing list