pcp
[Top] [All Lists]

Re: Gathering metrics from other hosts.

To: Jan-Frode Myklebust <janfrode@xxxxxxxxxxxxxxx>
Subject: Re: Gathering metrics from other hosts.
From: "Brent M. Clements" <bclem@xxxxxxxx>
Date: Sat, 27 Mar 2004 15:16:17 -0600 (CST)
Cc: pcp@xxxxxxxxxxx
In-reply-to: <20040327205404.GA11064@xxxxxxxxx>
References: <Pine.GSO.4.58.0403271340290.29724@xxxxxxxxxxx> <20040327205404.GA11064@xxxxxxxxx>
Sender: pcp-bounce@xxxxxxxxxxx
Thanks alot..this is EXACTLY what i needed!

I know I need to run the loggers on the master node, but should I run the
collector daemons and agents on each of the compute nodes?

Ie, should I install the pcp rpm on each of my compute nodes too and run
the pcp service?

-B


On Sat, 27 Mar 2004, Jan-Frode Myklebust wrote:

> On Sat, Mar 27, 2004 at 01:43:21PM -0600, Brent M. Clements wrote:
> >
> > Centralized Monitoring Server(master node in cluster) would gather metrics
> > from multiple hosts(compute nodes in a cluster). How would one actually do
> > this?
>
> I do this in our linux cluster (and wish more cluster-admins would see the
> value in pcp). Go to the /var/pcp/config/pmlogger/ on your frontend
> node, and copy config.sar to config.cluster. Edit the config.cluster
> file, comment out whatever you don't want logged.
>
> Then set up the control-file in the same directory to use this config
> for each of your nodes. I use:
>
>   node1           n   n   /export/home/pmlogger/node1             -c 
> ./config.cluster
>   node2           n   n   /export/home/pmlogger/node2             -c 
> ./config.cluster
>   node3           n   n   /export/home/pmlogger/node3             -c 
> ./config.cluster
>   node4           n   n   /export/home/pmlogger/node4             -c 
> ./config.cluster
>   node5           n   n   /export/home/pmlogger/node5             -c 
> ./config.cluster
>   etc..
>
> Then restart pcp on this frontend node, and pmlogger will begin logging.
>
> Another thing you might want to have a look at is pmie which can be
> used for f.ex. alerting you when file systems are running full. I have
> this in /var/pcp/config/pmie/config.clusternodes
>
>       delta = 4 mins;
>       filesys.filling =
>       some_inst (
>       ( 100 * filesys.used  /
>               filesys.capacity  ) > 80
>       && filesys.used  +
>               20 min * ( rate filesys.used  ) >
>               filesys.capacity
>       ) -> syslog 10 min "File system is filling up" " %v%used[%i]@%h";
>
> and a similar control file /var/pcp/config/pmie/control:
>
>       node1   n   PCP_LOG_DIR/pmie/node1/pmie.log   -c config.clusternodes
>       node2   n   PCP_LOG_DIR/pmie/node2/pmie.log   -c config.clusternodes
>       node3   n   PCP_LOG_DIR/pmie/node3/pmie.log   -c config.clusternodes
>       node4   n   PCP_LOG_DIR/pmie/node4/pmie.log   -c config.clusternodes
>       node5   n   PCP_LOG_DIR/pmie/node5/pmie.log   -c config.clusternodes
>       node6   n   PCP_LOG_DIR/pmie/node6/pmie.log   -c config.clusternodes
>       etc..
>
> Then there will be logged to the syslog on the monitor host if file
> system usage is growing too fast.
>
> BTW: you might also want to add these two entries to root's crontab
> for checking that the loggers are alive, and processing of old logs:
>
>       # daily processing of archive logs
>       10      0       *       *       * /usr/share/pcp/bin/pmlogger_daily -k 
> forever -x 5 -X gzip
>       # every 30 minutes, check pmlogger instances are running
>       25,55   *       *       *       * /usr/share/pcp/bin/pmlogger_check
>       # every 30 minutes, check pmie instances are running
>       24,54   *       *       *       *       /usr/share/pcp/bin/pmie_check
>
>
>    -jf
>

<Prev in Thread] Current Thread [Next in Thread>