[hobbit] Re: netapp.pl & xymon 4.2.3 & (no) disk graphs

Peter Welter peter.welter at gmail.com
Tue Apr 7 14:14:33 CEST 2009


Hi Francesco,

Thanks for replying. From hobbitgraph.cfg

[disk]
        FNPATTERN disk(.*).rrd
        TITLE Disk Utilization
        YAXIS % Full
        DEF:p at RRDIDX@=@RRDFN@:pct:AVERAGE
        LINE2:p at RRDIDX@#@COLOR@:@RRDPARAM@
        -u 100
        -l 0
        GPRINT:p at RRDIDX@:LAST: \: %5.1lf (cur)
        GPRINT:p at RRDIDX@:MAX: \: %5.1lf (max)
        GPRINT:p at RRDIDX@:MIN: \: %5.1lf (min)
        GPRINT:p at RRDIDX@:AVERAGE: \: %5.1lf (avg)\n

[disk1]
        FNPATTERN disk(.*).rrd
        TITLE Disk Utilization
        YAXIS Used
        DEF:p at RRDIDX@=@RRDFN@:used:AVERAGE
        CDEF:p at RRDIDX@t=p at RRDIDX@,1024,*
        LINE2:p at RRDIDX@t#@COLOR@:@RRDPARAM@
        -l 0
        GPRINT:p at RRDIDX@:LAST: \: %5.1lf KB (cur)
        GPRINT:p at RRDIDX@:MAX: \: %5.1lf KB (max)
        GPRINT:p at RRDIDX@:MIN: \: %5.1lf KB (min)
        GPRINT:p at RRDIDX@:AVERAGE: \: %5.1lf KB (avg)\n

There is no xstatdisk configuration in hobbitgraph.cfg, or any other
hobbit-config-file.

Regards,
Peter

2009/4/7 Francesco Duranti <fduranti at q8.it>:
> Hi peter,
> can you check the hobbitgraph.cfg file and copy/paste the [disk], [disk1] and [xstatdisk] configuration? Those should be responsible of the file pattern matching the filename to extract the data from the rrd files.
>
> To see the graphs of the xstat data you need to add them to the trend page adding "xstatifnet,xstatdisk,xstatqtree,xstatvolume,xstatlun" to GRAPHS in the hobbitserver.cfg (by default it's disabled because you will get a really high number of graphs).
>
>
> Francesco
>
>
> -----Original Message-----
> From: Peter Welter [mailto:peter.welter at gmail.com]
> Sent: Tuesday, April 07, 2009 11:37 AM
> To: hobbit at hswn.dk
> Subject: [hobbit] Re: netapp.pl & xymon 4.2.3 & (no) disk graphs
>
> I just found out by looking in the source (see below), that these
> xstat-files are new in Xymon. They potentially do deliver useful extra
> information about specific hardware disks and so on. But... when I
> just remove all xstat*-files to a sub-directory, then all my disk
> graphs re-appear again from the mist... so I guess I won't have to go
> back to the previous Hobbit release ;-)
>
> I guess I still need some help here, but now I think Xymon gets
> confused by all the disk-files in the rrd-directory, because we have a
> lot real disks in these NetApp's and that should not get confused with
> the logical disks that are created on them.
>
>
> myhost:/usr/src/packages/BUILD/xymon-4.2.3/hobbitd/rrd # grep xstat *
> do_netapp.c:    do_netapp_extratest_rrd(hostname,"xstatifnet",ifnetstr,tstamp,netapp_ifnet_params,ifnet_test);
> do_netapp.c:    do_netapp_extratest_rrd(hostname,"xstatqtree",qtreestr,tstamp,netapp_qtree_params,qtree_test);
> do_netapp.c:    do_netapp_extratest_rrd(hostname,"xstataggregate",aggregatestr,tstamp,netapp_aggregate_params,aggregate_test);
> do_netapp.c:    do_netapp_extratest_rrd(hostname,"xstatvolume",volumestr,tstamp,netapp_volume_params,volume_test);
> do_netapp.c:    do_netapp_extratest_rrd(hostname,"xstatlun",lunstr,tstamp,netapp_lun_params,lun_test);
> do_netapp.c:    do_netapp_extratest_rrd(hostname,"xstatdisk",diskstr,tstamp,netapp_disk_params,disk_test);
>
>
> 2009/4/7 Peter Welter <peter.welter at gmail.com>:
>> Hello all,
>>
>> In our environment we also monitor 5 NetApp's using (Franscesco
>> Duranti's) netapp.pl and everything worked fine for over 1,5 year or
>> so. Last night I upgraded our environment from Hobbit 4.2.0 to Xymon
>> 4.2.3.
>>
>> From that moment on, two of our main filers (camelot & excalibur) do
>> NOT report any disk graphs anymore, all other graphs just work fine.
>>
>> Our snapvault backup filer (noah) however, is mixed up. It shows 8
>> disk graphs, but most graphs (48...!) can not be seen...
>>
>> I checked the rrd-files (with rrdtool dump ...) and the disk-rrd-files
>> are still updated; they are just not shown anymore in either the disk-
>> and trends-column!?
>>
>> While checking the rrd-files, I discovered there are many
>> xstat*-.rrd-files like xstatdisk*, xstataggregate*, xstatqtree*,
>> xstatvolume* and xstatifnet*? I don't think it has anything to do with
>> the above. I just thought I should mention this.
>>
>> Fortunately I have two other filers who are not used anymore (sort
>> of), so I can think I can use these two filers to pinpoint the main
>> problem. But I really do need your help in this case.
>>
>> -- Thank you, Peter
>>
>>
>> - NO GRAPHS: CAMELOT -
>>
>> Filesystem                            Total        Used   Available
>> %Used Mounted On
>> /VUW                             9274283368  8314448264   959835104   90% /VUW
>> /vol/vol0                          16777216      696860    16080356
>> 4% /vol/vol0
>> /vol/VUW_APPLICATIONS             251658240   159270672    92387568
>> 63% /vol/VUW_APPLICATIONS
>> /vol/VUW_DEPARTMENTS             1417339208  1135870820   281468388
>> 80% /vol/VUW_DEPARTMENTS
>> /vol/VUW_HOMES                   3022918780  2901801520   121117260
>> 96% /vol/VUW_HOMES
>> /vol/VUW_WORKGROUPS              1202087528  1121291188    80796340
>> 93% /vol/VUW_WORKGROUPS
>> /vol/VUW_TEMPORARY                  8388608       12684     8375924
>> 0% /vol/VUW_TEMPORARY
>> /vol/VUW_PROFILES                 112459776    91835428    20624348
>> 82% /vol/VUW_PROFILES
>> /vol/VUW_DEPLOYMENT                83886080      704944    83181136
>> 1% /vol/VUW_DEPLOYMENT
>> /vol/VUW_SOFTWARE                 150994944    65930108    85064836
>> 44% /vol/VUW_SOFTWARE
>> /vol/VUW_HOMES2                   419430400   297936168   121494232
>> 71% /vol/VUW_HOMES2
>> /vol/VUW_HOMES3                   125829120     5290488   120538632
>> 4% /vol/VUW_HOMES3
>> /vol/VUW_PROFILES2                125829120    85202468    40626652
>> 68% /vol/VUW_PROFILES2
>> /vol/VUW_PROFILES3                 41943040      872628    41070412
>> 2% /vol/VUW_PROFILES3
>>
>> netapp.pl version 1.10 - column disk lifetime 60, tested in ~ 00:00:00
>> (max 00:02:00)
>>
>> - NO GRAPHS: EXCALIBUR -
>>
>> Filesystem                            Total        Used   Available
>> %Used Mounted On
>> /UB                              1309580492  1066842844   242737648   81% /UB
>> /SAP                             2976319296  2575401928   400917368   87% /SAP
>> /ULCN_en_VMware                  5238716824  1943827540  3294889284
>> 37% /ULCN_en_VMware
>> /vol/ub_test_aleph2               188743680   177386228    11357452
>> 94% /vol/ub_test_aleph2
>> /vol/vol0                          16777216     3751760    13025456
>> 22% /vol/vol0
>> /vol/ub_prod_arc2                 150994944    93639628    57355316
>> 62% /vol/ub_prod_arc2
>> /vol/sra                          125829120    92536612    33292508
>> 74% /vol/sra
>> /vol/ub_prod_aleph2               330301440   230768244    99533196
>> 70% /vol/ub_prod_aleph2
>> /vol/sld                           52428800    31220604    21208196
>> 60% /vol/sld
>> /vol/eco                          314572800   226719852    87852948
>> 72% /vol/eco
>> /vol/eca                          419430400   304858736   114571664
>> 73% /vol/eca
>> /vol/ecp                          524288000   393638904   130649096
>> 75% /vol/ecp
>> /vol/srp                          136314880    81797720    54517160
>> 60% /vol/srp
>> /vol/sro                          104857600    70797956    34059644
>> 68% /vol/sro
>> /vol/epo                           52428800    33606396    18822404
>> 64% /vol/epo
>> /vol/epa                           52428800    35038276    17390524
>> 67% /vol/epa
>> /vol/epp                           52428800    33691756    18737044
>> 64% /vol/epp
>> /vol/bwo                          167772160   136329488    31442672
>> 81% /vol/bwo
>> /vol/tro                           20971520     3338920    17632600
>> 16% /vol/tro
>> /vol/xio                           94371840    81723648    12648192
>> 87% /vol/xio
>> /vol/bwp                          335544320   242048196    93496124
>> 72% /vol/bwp
>> /vol/solaris                      115343360    27299736    88043624
>> 24% /vol/solaris
>> /vol/tra                           20971520     3041028    17930492
>> 15% /vol/tra
>> /vol/trp                           20971520     3999036    16972484
>> 19% /vol/trp
>> /vol/xip                           78643200    53005168    25638032
>> 67% /vol/xip
>> /vol/dba                           78643200    23371816    55271384
>> 30% /vol/dba
>> /vol/ub_test_aleph3               335544320   154981876   180562444
>> 46% /vol/ub_test_aleph3
>> /vol/vmware_test_marc_hooy        234881024    47024396   187856628
>> 20% /vol/vmware_test_marc_hooy
>> /vol/ulcn_mailstore              1089260752   821614700   267646052
>> 75% /vol/ulcn_mailstore
>> /vol/ulcn_surf                     52428800        2432    52426368
>> 0% /vol/ulcn_surf
>> /vol/sapbwa                       157286400   135781296    21505104
>> 86% /vol/sapbwa
>> /vol/sapxia                       136314880    94309300    42005580
>> 69% /vol/sapxia
>> /vol/ulcn_ota_mailstore            20971520         104    20971416
>> 0% /vol/ulcn_ota_mailstore
>> /vol/saptra                        16777216     3039008    13738208
>> 18% /vol/saptra
>>
>> netapp.pl version 1.10 - column disk lifetime 60, tested in ~ 00:00:00
>> (max 00:02:00)
>>
>> - SOME GRAPHS: NOAH (disks marked with  '        ***' at the end) -
>>
>> Filesystem                            Total        Used   Available
>> %Used Mounted On
>> /backupaggr_2_kog                8162374688  1851811324  6310563364
>> 23% /backupaggr_2_kog
>> /backupaggr_3_kog                8162374688   315328000  7847046688
>> 4% /backupaggr_3_kog
>> /backupaggr_kog                 12885604304 11654758296  1230846008
>> 90% /backupaggr_kog
>> /vol/vol0                          16777216      797048    15980168
>> 5% /vol/vol0
>> /vol/ULCN_MAILSTORE              1717567488  1343487832   374079656
>> 78% /vol/ULCN_MAILSTORE
>> /vol/SAP_PENSITO                  104857600    28819648    76037952
>> 27% /vol/SAP_PENSITO
>> /vol/VUWNAS02_APPLICATIONS        262144000   161875620   100268380
>> 62% /vol/VUWNAS02_APPLICATIONS
>> /vol/SAP_SAPBWA                   419430400   272891352   146539048
>> 65% /vol/SAP_SAPBWA
>> /vol/SAP_SAPXIA                   120586240   114195504     6390736
>> 95% /vol/SAP_SAPXIA
>> /vol/UFB_UFBSVR01                 104857600    14564464    90293136
>> 14% /vol/UFB_UFBSVR01
>> /vol/IBS_NIHIL                     52428800     2303108    50125692
>> 4% /vol/IBS_NIHIL          ***
>> /vol/IGRSFS101                    471859200   141582412   330276788
>> 30% /vol/IGRSFS101
>> /vol/IBS_ORWELL                   104857600    26580588    78277012
>> 25% /vol/IBS_ORWELL
>> /vol/IBS_NEXUS                     52428800    12342608    40086192
>> 24% /vol/IBS_NEXUS          ***
>> /vol/IBS_NUNQUAM                   52428800         120    52428680
>> 0% /vol/IBS_NUNQUAM        ***
>> /vol/VUW_EXCHANGE                 157286400      295364   156991036
>> 0% /vol/VUW_EXCHANGE
>> /vol/IBS_NUSQUAM                   52428800     1892756    50536044
>> 4% /vol/IBS_NUSQUAM        ***
>> /vol/IBS_NAGGER                    31457280     1378740    30078540
>> 4% /vol/IBS_NAGGER         ***
>> /vol/VUW_VUWSDC02                 524288000   491685452    32602548
>> 94% /vol/VUW_VUWSDC02
>> /vol/VUWNAS02_DEPARTMENTS        1287651328  1213016980    74634348
>> 94% /vol/VUWNAS02_DEPARTMENTS
>> /vol/VUWNAS02_HOMES              3328180224  3211611536   116568688
>> 96% /vol/VUWNAS02_HOMES
>> /vol/GAMMA                         68157440      218848    67938592
>> 0% /vol/GAMMA              ***
>> /vol/VUWNAS02_WORKGROUPS         1258291200  1184439936    73851264
>> 94% /vol/VUWNAS02_WORKGROUPS
>> /vol/VUWNAS02_PROFILES            314572800   296196216    18376584
>> 94% /vol/VUWNAS02_PROFILES
>> /vol/VUWNAS02_DEPLOYMENT          419430400    15745112   403685288
>> 4% /vol/VUWNAS02_DEPLOYMENT
>> /vol/VUWNAS02_SOFTWARE            125829120    66229432    59599688
>> 53% /vol/VUWNAS02_SOFTWARE
>> /vol/UB_PROD_ALEPH2               838860800   516906036   321954764
>> 62% /vol/UB_PROD_ALEPH2
>> /vol/VUW_VUWSDC01                 545259520   507335816    37923704
>> 93% /vol/VUW_VUWSDC01
>> /vol/VUW_VUWSDC04                  73400320    62864376    10535944
>> 86% /vol/VUW_VUWSDC04
>> /vol/VUW_VUWSMC02                  62914560    50989600    11924960
>> 81% /vol/VUW_VUWSMC02
>> /vol/VUW_VUWSOM01                  15728640    12395368     3333272
>> 79% /vol/VUW_VUWSOM01
>> /vol/UB_PROD_ARC2                 503316480   454875292    48441188
>> 90% /vol/UB_PROD_ARC2
>> /vol/VUW_VUWSWP02                  47185920    35898380    11287540
>> 76% /vol/VUW_VUWSWP02
>> /vol/VUW_VUWSWP03                 314572800   266760000    47812800
>> 85% /vol/VUW_VUWSWP03
>> /vol/VUW_VUWSDS03                 104857600    24002548    80855052
>> 23% /vol/VUW_VUWSDS03
>> /vol/VUW_VUWSDC03                 676331520   628669724    47661796
>> 93% /vol/VUW_VUWSDC03
>> /vol/VUW_VUWSDC05                 104857600    45930952    58926648
>> 44% /vol/VUW_VUWSDC05
>> /vol/VUW_VUWSMC01                 157286400    84099620    73186780
>> 53% /vol/VUW_VUWSMC01
>> /vol/IBS_INSULA                   104857600    16690432    88167168
>> 16% /vol/IBS_INSULA         ***
>> /vol/VUW_VUWSOM02                 524288000   361859348   162428652
>> 69% /vol/VUW_VUWSOM02
>> /vol/VUW_VUWSOM03                  20971520    10001972    10969548
>> 48% /vol/VUW_VUWSOM03
>> /vol/UB_DELOS                     104857600    47352240    57505360
>> 45% /vol/UB_DELOS
>> /vol/ICLON_ICLAPP01                73400320    53232408    20167912
>> 73% /vol/ICLON_ICLAPP01
>> /vol/ICLON_ICLWEB01                62914560    46468948    16445612
>> 74% /vol/ICLON_ICLWEB01
>> /vol/VMWARE_VIRTUALCENTER         104857600       19512   104838088
>> 0% /vol/VMWARE_VIRTUALCENTER
>> /vol/VUWNAS02_HOMES2              576716800   355405576   221311224
>> 62% /vol/VUWNAS02_HOMES2
>> /vol/VUWNAS02_HOMES3              209715200     6830260   202884940
>> 3% /vol/VUWNAS02_HOMES3
>> /vol/VUWNAS02_PROFILES2           157286400   140940584    16345816
>> 90% /vol/VUWNAS02_PROFILES2
>> /vol/VUWNAS02_PROFILES3            52428800     3244172    49184628
>> 6% /vol/VUWNAS02_PROFILES3
>> /vol/SAP_FERA                     419430400    86547324   332883076
>> 21% /vol/SAP_FERA
>> /vol/SAP_PECUNIA                  146800640    39624548   107176092
>> 27% /vol/SAP_PECUNIA
>> /vol/VOIP_DORIS                   104857600    68756404    36101196
>> 66% /vol/VOIP_DORIS
>> /vol/VOIP_DIONYSUS                 83886080    33812768    50073312
>> 40% /vol/VOIP_DIONYSUS
>> /vol/VOIP_TELESTO                 104857600    65484616    39372984
>> 62% /vol/VOIP_TELESTO
>> /vol/TNB_HESTIA                   209715200       94268   209620932
>> 0% /vol/TNB_HESTIA
>> /vol/UB_TEST_ALEPH                419430400   195480256   223950144
>> 47% /vol/UB_TEST_ALEPH
>> /vol/WEB_KERESIS                  104857600     4617936   100239664
>> 4% /vol/WEB_KERESIS
>> /vol/ERBIS_THOR                   157286400    23998732   133287668
>> 15% /vol/ERBIS_THOR
>> /vol/UB_ARCHIMEDES                 52428800     1580596    50848204
>> 3% /vol/UB_ARCHIMEDES
>> /vol/SAP_SAPTRA                    31457280     2939920    28517360
>> 9% /vol/SAP_SAPTRA
>> /vol/WEB_MIRAMAR                   52428800     1484252    50944548
>> 3% /vol/WEB_MIRAMAR
>> /vol/WEB_PENTHARIAN                52428800     1472288    50956512
>> 3% /vol/WEB_PENTHARIAN
>> /vol/WEB_APOLLYON                  52428800     2899836    49528964
>> 6% /vol/WEB_APOLLYON
>> /vol/IBS_NIRVANA                   62914560     1231284    61683276
>> 2% /vol/IBS_NIRVANA        ***
>> /vol/IGRSFS001_ORG                367001600   136379928   230621672
>> 37% /vol/IGRSFS001_ORG
>> /vol/IGRSFS001                    142606336   116948784    25657552
>> 82% /vol/IGRSFS001
>> /vol/koala                        104857600    14221836    90635764
>> 14% /vol/koala
>>
>> netapp.pl version 1.10 - column disk lifetime 60, tested in ~ 00:00:00
>> (max 00:02:00)
>>
>>
>> Normal behaviour:
>> nas-node1:
>> Filesystem                            Total        Used   Available
>> %Used Mounted On
>> /vol/vuwanas01                    228589568   190858164    37731404
>> 83% /vol/vuwanas01
>> /vol/vol0                          30198992     1039268    29159724
>> 3% /vol/vol0
>>
>> netapp.pl version 1.10 - column disk lifetime 60, tested in ~ 00:00:00
>> (max 00:02:00)
>> nas-node2:
>> Filesystem                            Total        Used   Available
>> %Used Mounted On
>> /vol/vol0                          30198992      548304    29650688
>> 2% /vol/vol0
>>
>> netapp.pl version 1.10 - column disk lifetime 60, tested in ~ 00:00:00
>> (max 00:02:00)
>>
>
> To unsubscribe from the hobbit list, send an e-mail to
> hobbit-unsubscribe at hswn.dk
>
>
>
> To unsubscribe from the hobbit list, send an e-mail to
> hobbit-unsubscribe at hswn.dk
>
>
>



More information about the Xymon mailing list