Guide to IASTU-HPC2
  • Introduction
  • User's Manual
    • basics
      • basic info
      • connecting
      • storage (must-read)
      • module system (must-read)
      • workflow on jobs
    • softwares
      • python
      • mathematica
      • matlab
      • spark
      • singularity
      • fortran, C, C++
      • java
      • command line tools
      • tensorflow
      • jax
    • FAQ
  • Administrator's Manual
    • Hardwares
    • Toolchains
      • Toolset for DELL servers
      • Ansible
      • Slurm
      • Spack
      • Container
      • BigData
      • Toolset for logging and monitoring
      • Further considerations
    • History
      • ToDo
      • VM Test
      • Real Setup
      • Admin Workflow
      • Softwares for scientific computations
      • Relay Host
Powered by GitBook
On this page
  • rsyslog
  • logrotate
  • Logstash
  • Nagios
  • Installation
  • Ganglia
  • Installation
  • Backend
  • ELK
  • Benchmark
  • Linpack (hpl)
  • MPI benchmarks
  • IO benchmarks

Was this helpful?

  1. Administrator's Manual
  2. Toolchains

Toolset for logging and monitoring

PreviousBigDataNextFurther considerations

Last updated 5 years ago

Was this helpful?

In this section, I will review some aspects on the log and monitoring related softwares such as syslog, nagios and ganglia.

rsyslog

logrotate

Logstash

Merged into elk stack

Pipeline from log to elastic search, see for introduction.

Nagios

Installation

  • See for its installation on ubuntu (one node).

  • Install on ubuntu 18.04 cluster and basic configuration,

Ganglia

Installation

Backend

rrdtool: round robin database

ELK

Elasticsearch + logstash+kibana, to be dployed soon, firstly install on VM environment

initial install finished on the cluster

ganglia {
    port => 28649
    host => "{{ master_ip }}"
    type => "ganglia"

  }

  udp_send_channel {
  host = {{ master_name }}
  port = 28649
  ttl = 1
}

experiment conclusion: ganglia doesn't work well with logstash

Benchmark

Linpack (hpl)

All of the Gold 6000 processors have two AVX512 units, so they are capable of 32 DP FLOPS/cycle. The Gold 5000 processors have one AVX512 unit (except for the Gold 5122, which has two), so they are capable of 16 DP FLOPS/cycle.

Results: 580Gflop on single machine in normal env without fine tuning

MPI benchmarks

IO benchmarks

ganglia input plugin for logstash:

multiple input for logstash and type label:

: combine ELK stack and email alert

Cluster of ES , to be deployed. One also need to configure ssl for es cluster if authetication is enabled. see .

Theorectial Flop value for CPU:

Simply linpack directly from intel parallel studio:

Intel mkl linpack documentation:

An example using spack installed stack for mpi bandwidth benchmark:

An example using spack installed stack for io benchmark:

this blog
this post
post
ubuntu16.04
simple intro
Installation of ELK on ubuntu18.04 by digital ocean
doc
so
elasticalert
post
this doc
post on linpack
post
blog
intel
blog
blog