need help with SNMP Generator

2,954 views
Skip to first unread message

IndGirl6

unread,
Jun 22, 2018, 3:00:06 PM6/22/18
to Prometheus Users
hi,
i am trying to get the snmp exporter to work. 
I downloaded the MIB for my device (cisco UCS) from "https://github.com/librenms/librenms/tree/master/mibs/cisco"

i put the MIB in "/usr/local/share/snmp/mibs"

added the following lines to the generator.yml file:
#CISCO UCS
  cisco_ucs
    walk:
      - sysUpTime
      - interfaces
      - ifXTable
      - 1.3.6.1.4.1.9.9.392.1
      - 1.3.6.1.4.1.9.9.147.1
    lookups:
      - old_index: ifIndex
        new_index: ifDescr
    version: 1


Now i am trying to generate the snmp.yml and am stuck.
My machine is Redhat Linux 6.  The steps for building the generator is for Debian.

Can some one let me know how i can proceed?

THanks 

IndGirl6

unread,
Jun 22, 2018, 4:43:30 PM6/22/18
to Prometheus Users
some update:
after researching this and from this old thread on this forum (https://groups.google.com/forum/#!topic/prometheus-users/8qvob2UjSIs)

i was able to generate the snmp.yml file with no errors for my CISCO UCS

[root]# ./generator generate
INFO[0000] Loading MIBs from $HOME/.snmp/mibs:/usr/share/snmp/mibs  source="net_snmp.go:135"
WARN[0003] NetSNMP reported 1430 parse errors            source="main.go:100"
WARN[0004] Can't find augmenting oid tokenRingMLStatsEntry for tokenRingMLStats2Entry  source="tree.go:71"
WARN[0004] Can't find augmenting oid tokenRingPStatsEntry for tokenRingPStats2Entry  source="tree.go:71"
WARN[0004] Can't find augmenting oid dot1agCfmMepEntry for mefSoamPmMepEntry  source="tree.go:71"
WARN[0004] Can't find augmenting oid rptrPortEntry for ciscoRptrPortEntry  source="tree.go:71"
INFO[0004] Generating config for module if_mib           source="main.go:49"
INFO[0004] Generated 42 metrics for module if_mib        source="main.go:60"
INFO[0004] Generating config for module if_mib_ifalias   source="main.go:49"
INFO[0004] Generated 42 metrics for module if_mib_ifalias  source="main.go:60"
INFO[0004] Generating config for module if_mib_ifdescr   source="main.go:49"
INFO[0004] Generated 42 metrics for module if_mib_ifdescr  source="main.go:60"
INFO[0004] Generating config for module if_mib_ifname    source="main.go:49"
INFO[0004] Generated 42 metrics for module if_mib_ifname  source="main.go:60"
INFO[0004] Generating config for module cisco_ucs        source="main.go:49"
INFO[0004] Generated 5495 metrics for module cisco_ucs   source="main.go:60"
INFO[0005] Config written to /home/prometheus/downloads/snmp_exporter/snmp_exporter-0.11.0.linux-amd64/snmp.yml
[root]#


My prometheus.yml file is as below:
  - job_name: 'snmp'
    static_configs:
      - targets:
        - 172.16.148.20
    metrics_path: /snmp
    params:
      module: [if_mib]
    relabel_configs:
      - source_labels: [__address__]
        target_label: __param_target
      - source_labels: [__param_target]
        target_label: instance
      - target_label: __address__
        replacement: snmp-host:9116  # The SNMP exporter's real hostname:port.


when i go to "http://snmp-host:9116/snmp?target=172.16.148.20" i get the following error after a while:

An error has occurred during metrics gathering:
error collecting metric Desc{fqName: "snmp_error", help: "Error scraping target", constLabels: {}, variableLabels: []}: Error getting target 172.16.148.20: Request timeout (after 3 retries)

in my snmp_exporter log file i see simillar time out errors:
time="2018-06-22T16:42:16-04:00" level=info msg="Error scraping target 172.16.148.20: Error getting target 172.16.148.20: Request timeout (after 3 retries)" source="collector.go:179"

Am i missing something?


IndGirl6

unread,
Jun 22, 2018, 5:19:59 PM6/22/18
to Prometheus Users
one more update, i set the log level to debug for the exporter and it looks like its a time out issue with snmp_exporter.
time="2018-06-22T17:18:46-04:00" level=debug msg="Scrape of target '172.16.148.20' with module 'if_mib' took 60.000773 seconds" source="main.go:97"
time="2018-06-22T17:19:01-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'if_mib'" source="main.go:86"
time="2018-06-22T17:19:01-04:00" level=debug msg="Getting 1 OIDs from target \"172.16.148.20\"" source="collector.go:94"
time="2018-06-22T17:19:01-04:00" level=info msg="Error scraping target 172.16.148.20: Error getting target 172.16.148.20: Request timeout (after 3 retries)" source="collector.go:179"


does any one know how i can increase the timeout for snmp_exporter?

Danny Kulchinsky

unread,
Jun 22, 2018, 9:30:14 PM6/22/18
to Prometheus Users
Just add timeout: <duration> to your generator.yml, for example:

modules:
  <module>:
    walk:
      - sysUpTime
      - interfaces
      - ifXTable
    timeout: 30s
    lookups:
      - old_index: ifIndex
        new_index: ifDescr


Ben Kochie

unread,
Jun 23, 2018, 12:37:00 AM6/23/18
to Kavita, Prometheus Users
You are trying to collect some very large OID trees, with SNMP v1. SNMP v1 does not support bulk fetching, and will be very slow.

Based on the error, your device is not responding to a single SNMP v1 OID request, this is probably means you have an auth problem, or your device doesn't support SNMP v1.

--
You received this message because you are subscribed to the Google Groups "Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to prometheus-use...@googlegroups.com.
To post to this group, send email to promethe...@googlegroups.com.
To view this discussion on the web visit https://groups.google.com/d/msgid/prometheus-users/fd70df4c-c760-4748-be89-57d538670b4a%40googlegroups.com.
For more options, visit https://groups.google.com/d/optout.

Kavita

unread,
Jun 23, 2018, 12:57:19 PM6/23/18
to Danny Kulchinsky, Prometheus Users
Thanks Danny. Did not know of all the options.. will try that on monday.

--
You received this message because you are subscribed to the Google Groups "Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to prometheus-use...@googlegroups.com.
To post to this group, send email to promethe...@googlegroups.com.

Kavita

unread,
Jun 23, 2018, 1:02:04 PM6/23/18
to Ben Kochie, Prometheus Users
Thanks for your response Ben.
My understandibg of SNMP is very limited. 
I have tried this setup with both snmp V1 and V2 and i get the same error.

When i run the snmpwalk command with our community string it is sucessful for both SNMP V1 and V2. Based on this i am thinking it is not an Auth issue.

I will try increaseing the timout in generator.yml and see if that helps too. 

Is there anything else you can think of that i can try? Let me know..

Thanks
Indgirl6



Ben Kochie

unread,
Jun 23, 2018, 1:12:59 PM6/23/18
to Kavita, Prometheus Users
Another thing I just noticed.

You are creating an SNMP module "cisco_ucs", but you have configured Prometheus to request the "if_mib" module.

Danny Kulchinsky

unread,
Jun 23, 2018, 1:18:43 PM6/23/18
to kavita...@gmail.com, Prometheus Users
Sure, you should also look at Ben's comments.


Danny

IndGirl6

unread,
Jun 25, 2018, 9:14:02 AM6/25/18
to Prometheus Users
thanks, i changed the prometheus configs to request cisco_ucs module  but it still appears to time out. Going to try increasing the timeout now..:
Below are the messages from scraping cisco_ucs modeule:

time="2018-06-25T09:09:59-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'c               isco_ucs'" source="main.go:86"
time="2018-06-25T09:09:59-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3               .6.1.2.1.2\"" source="collector.go:124"
time="2018-06-25T09:10:14-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'c               isco_ucs'" source="main.go:86"
time="2018-06-25T09:10:14-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3               .6.1.2.1.2\"" source="collector.go:124"
time="2018-06-25T09:10:29-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'cisco_ucs'" source="main.go:86"
ime="2018-06-25T09:10:29-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3.6.1.2.1.2\"" source="collector.go:124"


time="2018-06-25T09:10:44-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'cisco_ucs'" source="main.go:86"
time="2018-06-25T09:10:44-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3.6.1.2.1.2\"" source="collector.go:124"
time="2018-06-25T09:10:59-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'cisco_ucs'" source="main.go:86"
time="2018-06-25T09:10:59-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3.6.1.2.1.2\"" source="collector.go:124"
time="2018-06-25T09:10:59-04:00" level=info msg="Error scraping target 172.16.148.20: Error walking target 172.16.148.20: Request timeout (after 3 retries)" source="collector.go:179"
time="2018-06-25T09:10:59-04:00" level=debug msg="Scrape of target '172.16.148.20' with module 'cisco_ucs' took 60.012094 seconds" source="main.go:97"

IndGirl6

unread,
Jun 25, 2018, 12:13:23 PM6/25/18
to Prometheus Users
so i tried increasing the timeout and that did not help either.
I ran the snmpwalk command  for both V1 and V2 like so:

snmpwalk -v1 -c <string> 172.16.148.20
snmpwalk -v2c -c <string> 172.16.148.20

both took around 1minute and 12 seconds.

I then increased the timeout to 90s in the generator, and generated a new snmp.yml file.

the exact same thing happens and this time it times out after 90s instead of 60.

any advice on what else i can check or do to narrow this down?

IndGirl6

unread,
Jun 26, 2018, 10:42:49 AM6/26/18
to Prometheus Users
really stuck here and dont know which way to go.. i think its something simple / silly that i am missing somewhere. i used the same configs to looks at a server and it works OK.

Is there any setting changes i need to make on the UCS side?
If you have successfully configured monitoring of the Cisco UCS, would you be able to send my your generator.yml file?
thanks..

Danny Kulchinsky

unread,
Jun 26, 2018, 10:44:32 AM6/26/18
to Kavita, Prometheus Users
Have you double checked the community string? and also it's common to have an acl in the device to restrict which IPs/Networks can access it via SNMP...


Danny

--
You received this message because you are subscribed to a topic in the Google Groups "Prometheus Users" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/prometheus-users/8WlSvHUshGM/unsubscribe.
To unsubscribe from this group and all its topics, send an email to prometheus-use...@googlegroups.com.

To post to this group, send email to promethe...@googlegroups.com.

IndGirl6

unread,
Jun 26, 2018, 10:50:54 AM6/26/18
to Prometheus Users
Yes the community string is correct, both the following commands fail with no response if i give a different community string or set it to public:

snmpwalk -v1 -c <string> 172.16.148.20
snmpwalk -v2c -c <string> 172.16.148.20

Also ensured firewall is off. 

Is there a very very basic snmp.yml file that i can use which does not use any of the cisco OID, just to make sure that the connectivity is happening?
I did try removing all cisco OID and leaving only the IF_MIB module in the yml file, but it still times out

Ben Kochie

unread,
Jun 26, 2018, 11:32:23 AM6/26/18
to Kavita, Prometheus Users
As I noted earlier, you are walking single very large trees of data.

The timeouts work on a walk-by-walk tree basis, so if you split up your walk into several sub-trees that can return more quickly, the overall request will complete.

Also, what is the ping time between the exporter and the network device? SNMP is extremely chatty and latency sensitive. If you have more than a few milliseconds ping between the exporter and the device it can slow down greatly.

--
You received this message because you are subscribed to the Google Groups "Prometheus Users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to prometheus-use...@googlegroups.com.
To post to this group, send email to promethe...@googlegroups.com.

IndGirl6

unread,
Jun 26, 2018, 12:04:06 PM6/26/18
to Prometheus Users
Hi Ben, 
Here is the ping result:

ping 172.16.148.20
PING 172.16.148.20 (172.16.148.20) 56(84) bytes of data.
64 bytes from 172.16.148.20: icmp_seq=1 ttl=62 time=0.856 ms
64 bytes from 172.16.148.20: icmp_seq=2 ttl=62 time=0.685 ms
64 bytes from 172.16.148.20: icmp_seq=3 ttl=62 time=0.707 ms
64 bytes from 172.16.148.20: icmp_seq=4 ttl=62 time=0.679 ms
64 bytes from 172.16.148.20: icmp_seq=5 ttl=62 time=0.768 ms
64 bytes from 172.16.148.20: icmp_seq=6 ttl=62 time=0.671 ms
64 bytes from 172.16.148.20: icmp_seq=7 ttl=62 time=0.733 ms
64 bytes from 172.16.148.20: icmp_seq=8 ttl=62 time=0.724 ms
64 bytes from 172.16.148.20: icmp_seq=9 ttl=62 time=0.705 ms
64 bytes from 172.16.148.20: icmp_seq=10 ttl=62 time=0.671 ms
64 bytes from 172.16.148.20: icmp_seq=11 ttl=62 time=0.714 ms
64 bytes from 172.16.148.20: icmp_seq=12 ttl=62 time=0.642 ms
^C
--- 172.16.148.20 ping statistics ---
12 packets transmitted, 12 received, 0% packet loss, time 11194ms
rtt min/avg/max/mdev = 0.642/0.712/0.856/0.064 ms


would you be able to tell me how i can split the trees. Just for testing even if i can have the most basic config that maybe checks JUST the uptime (or something like that). That would help me confirm this is a time out issue and not something else..

Also on another note, is there any command i can manually run against OID to see how much time a sucessful walk should take? once i have this info i can increase the timeouts accordingly.
i tried the following command to get the time, but i am not sure if i am doingt he correct test here:
time snmpwalk -v1 -c <string> 172.16.148.20 -C i 1.3.6.1.4.1.9.9.719.1.9
the above took 1m12s to complete

Ben Kochie

unread,
Jun 26, 2018, 12:14:39 PM6/26/18
to Kavita, Prometheus Users
Looking at the MIB you're using, 1.3.6.1.4.1.9.9.719.1.9 is huge, and a number of sections look like metrics data.

I would start with one small table, say cucsComputeIOHubEnvStatsTable. (1.3.6.1.4.1.9.9.719.1.9.12)

IndGirl6

unread,
Jun 26, 2018, 2:18:29 PM6/26/18
to Prometheus Users
Ok, so i tried with only that one module and it still times out just like before. i am not sure, but i  have a feeling that this is not related to timeout. could be some other issue (config or other).
Below are the full details of messages and my config file..

Message from snmp_exporter debug logs:
time="2018-06-26T14:09:34-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'cisco_ucs'" source="main.go:86"
time="2018-06-26T14:09:34-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3.6.1.2.1.2\"" source="collector.go:124"
time="2018-06-26T14:11:34-04:00" level=info msg="Error scraping target 172.16.148.20: Error walking target 172.16.148.20: Request timeout (after 1 retries)" source="collector.go:179"
time="2018-06-26T14:11:34-04:00" level=debug msg="Scrape of target '172.16.148.20' with module 'cisco_ucs' took 120.001114 seconds" source="main.go:97"
time="2018-06-26T14:12:21-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'cisco_ucs'" source="main.go:86"
time="2018-06-26T14:12:21-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3.6.1.2.1.2\"" source="collector.go:124"


Messages displayed on metrics web page (after trying for 120sec)
An error has occurred during metrics gathering:
error collecting metric Desc{fqName: "snmp_error", help: "Error scraping target", constLabels: {}, variableLabels: []}: Error walking target 172.16.148.20: Request timeout (after 1 retries)

contents of generator.yml file:
more generator.yml
modules:
  # Default IF-MIB interfaces table with ifIndex.
  if_mib:
    walk: [sysUpTime, interfaces, ifXTable]
  # Interfaces if ifAlias is unique.
  if_mib_ifalias:
    walk: [sysUpTime, interfaces, ifXTable]
    lookups:
      - old_index: ifIndex
        new_index: ifAlias
  # Interfaces if ifDescr is unique.
  if_mib_ifdescr:
    walk: [sysUpTime, interfaces, ifXTable]
    lookups:
      - old_index: ifIndex
        new_index: ifDescr
  # Interfaces if ifName is unique.
  if_mib_ifname:
    walk: [sysUpTime, interfaces, ifXTable]
    lookups:
      - old_index: ifIndex
        # Use OID to avoid conflict with Netscaler NS-ROOT-MIB.
        new_index: 1.3.6.1.2.1.31.1.1.1.1 # ifName

# Cicso UCS
  cisco_ucs:
    walk:
      - interfaces
      - ifXTable
      - 1.3.6.1.4.1.9.9.719.1.9.12  #cucsComputeIOHubEnvStatsTable
#      - 1.3.6.1.4.1.9.9.719.1.16  #CISCO-UNIFIED-COMPUTING-ETHER-MIB
#      - 1.3.6.1.4.1.9.9.719.4  #CISCO-UNIFIED-COMPUTING-TC-MIB
#      - 1.3.6.1.4.1.9.9.719.1.3   #CISCO-UNIFIED-COMPUTING-ADAPTOR-MIB
#      - 1.3.6.1.4.1.9.9.719.1.20  #CISCO-UNIFIED-COMPUTING-FC-MIB
#      - 1.3.6.1.4.1.9.9.719.1.9   #CISCO-UNIFIED-COMPUTING-COMPUTE-MIB
#      - 1.3.6.1.4.1.9.9.719.1.41  #CISCO-UNIFIED-COMPUTING-PROCESSOR-MIB
#      - 1.3.6.1.4.1.9.9.719.1.30  #CISCO-UNIFIED-COMPUTING-MEMORY-MIB
#      - 1.3.6.1.4.1.9.9.719.1.46  #CISCO-UNIFIED-COMPUTING-SW-MIB
    auth:
      community: <string>
    retries: 1
    timeout: 120s



# ./generator generate
INFO[0000] Loading MIBs from $HOME/.snmp/mibs:/usr/share/snmp/mibs  source="net_snmp.go:135"
WARN[0004] NetSNMP reported 8770 parse errors            source="main.go:100"
WARN[0005] Can't find augmenting oid tokenRingMLStatsEntry for tokenRingMLStats2Entry  source="tree.go:71"
WARN[0005] Can't find augmenting oid tokenRingPStatsEntry for tokenRingPStats2Entry  source="tree.go:71"
WARN[0005] Can't find augmenting oid dot1agCfmMepEntry for mefSoamPmMepEntry  source="tree.go:71"
WARN[0005] Can't find augmenting oid cCdmaPcfSoRpRegStatsEntry for ccpCdmaExtPcfSoRpRegStatsEntry  source="tree.go:71"
WARN[0005] Can't find augmenting oid cCdmaPcfSoPppSetupStatsEntry for ccpCdmaExtPcfSoPppSetupStatsEntry  source="tree.go:71"
WARN[0005] Can't find augmenting oid rptrPortEntry for ciscoRptrPortEntry  source="tree.go:71"
WARN[0005] Can't find augmenting oid cSctpAssocEntry for cSctpAssocExtEntry  source="tree.go:71"
WARN[0005] Can't find augmenting oid cSctpAssocRemAddressEntry for cSctpAssocRemAddressExtEntry  source="tree.go:71"
INFO[0005] Generating config for module if_mib           source="main.go:49"
INFO[0005] Generated 42 metrics for module if_mib        source="main.go:60"
INFO[0005] Generating config for module if_mib_ifalias   source="main.go:49"
INFO[0005] Generated 42 metrics for module if_mib_ifalias  source="main.go:60"
INFO[0005] Generating config for module if_mib_ifdescr   source="main.go:49"
INFO[0005] Generated 42 metrics for module if_mib_ifdescr  source="main.go:60"
INFO[0005] Generating config for module if_mib_ifname    source="main.go:49"
INFO[0005] Generated 42 metrics for module if_mib_ifname  source="main.go:60"
INFO[0005] Generating config for module cisco_ucs        source="main.go:49"
INFO[0005] Generated 53 metrics for module cisco_ucs     source="main.go:60"
INFO[0005] Config written to /home/prometheus/downloads/snmp_exporter/snmp_exporter-0.11.0.linux-amd64/snmp.yml  source="main.go:84"
#

Danny Kulchinsky

unread,
Jun 26, 2018, 3:04:45 PM6/26/18
to Kavita, Prometheus Users
Could you try with lower max_repititions (like 10 for example), I had a similar issue once and this helped.

Danny

--
You received this message because you are subscribed to a topic in the Google Groups "Prometheus Users" group.
To unsubscribe from this topic, visit https://groups.google.com/d/topic/prometheus-users/8WlSvHUshGM/unsubscribe.
To unsubscribe from this group and all its topics, send an email to prometheus-use...@googlegroups.com.

To post to this group, send email to promethe...@googlegroups.com.

IndGirl6

unread,
Jun 26, 2018, 5:00:23 PM6/26/18
to Prometheus Users
tried changing Max repetitions to 8 and am sitll having the same issue :(

do you have a working config, if so, can you share that?

nicolai.v...@sap.com

unread,
Jun 29, 2018, 7:57:52 AM6/29/18
to Prometheus Users
Hi IndGirl6,

I also have some of these cisco machines.
Maybe you would like to try with this snmp.yml, as it is smaller and should work?
Just to make it sucessfully run once, to figure out if it is some issue with your specific snmp.yml.

cisco-ucs:
  walk:
  - 1.3.6.1.4.1.9.9.719.1.1.1.1.16
  - 1.3.6.1.4.1.9.9.719.1.1.1.1.2
  metrics:
  - name: cucsFaultOccur
    oid: 1.3.6.1.4.1.9.9.719.1.1.1.1.16
    type: counter
    help: Cisco UCS fault:Inst:occur managed object property - 1.3.6.1.4.1.9.9.719.1.1.1.1.16
    indexes:
    - labelname: cucsFaultDn
      type: gauge
    - labelname: cucsFaultTags
      type: OctetString
    lookups:
    - labels:
      - cucsFaultDn
      labelname: cucsFaultDn
      oid: 1.3.6.1.4.1.9.9.719.1.1.1.1.2
      type: DisplayString
    - labels:
      - cucsFaultTags
      labelname: cucsFaultTags
      oid: 1.3.6.1.4.1.9.9.719.1.1.1.1.21
      type: OctetString
  version: 2
  auth:
    all your personal auth settings.

IndGirl6

unread,
Jun 29, 2018, 9:32:24 AM6/29/18
to Prometheus Users
Hi Nicolai,
Thanks for your response, i tried the snmp.yml you sent, but i have the exact same error as below. What are your timeouts set to? Was there any setup you had to do from the Cisco UCS side?
i fet the folloing errors message on the web browser:

An error has occurred during metrics gathering:
error collecting metric Desc{fqName: "snmp_error", help: "Error scraping target", constLabels: {}, variableLabels: []}: Error walking target 172.16.148.20: Request timeout (after 1 retries)

Following message in my snmp_exporter debug logs:
time="2018-06-29T09:26:35-04:00" level=info msg="Starting snmp exporter (version=0.11.0, branch=HEAD, revision=e4591716c29459cb2a12b1bed129af519ad91d23)" source="main.go:138"
time="2018-06-29T09:26:35-04:00" level=info msg="Build context (go=go1.10.2, user=root@80735d30559d, date=20180530-10:24:52)" source="main.go:139"
time="2018-06-29T09:26:35-04:00" level=info msg="Listening on :9116" source="main.go:218"



time="2018-06-29T09:27:24-04:00" level=debug msg="Scraping target '172.16.148.20' with module 'cisco-ucs'" source="main.go:86"
time="2018-06-29T09:27:24-04:00" level=debug msg="Walking target \"172.16.148.20\" subtree \"1.3.6.1.4.1.9.9.719.1.1.1.1.16\"" source="collector.go:124"
time="2018-06-29T09:29:24-04:00" level=info msg="Error scraping target 172.16.148.20: Error walking target 172.16.148.20: Request timeout (after 1 retries)" source="collector.go:179"
time="2018-06-29T09:29:24-04:00" level=debug msg="Scrape of target '172.16.148.20' with module 'cisco-ucs' took 120.001123 seconds" source="main.go:97"


Below is my snmp.yml config file:
cisco-ucs:
  walk:
  - 1.3.6.1.4.1.9.9.719.1.1.1.1.16
  - 1.3.6.1.4.1.9.9.719.1.1.1.1.2
  metrics:
  - name: cucsFaultOccur
    oid: 1.3.6.1.4.1.9.9.719.1.1.1.1.16
    type: counter
    help: Cisco UCS fault:Inst:occur managed object property - 1.3.6.1.4.1.9.9.719.1.1.1.1.16
    indexes:
    - labelname: cucsFaultDn
      type: gauge
    - labelname: cucsFaultTags
      type: OctetString
    lookups:
    - labels:
      - cucsFaultDn
      labelname: cucsFaultDn
      oid: 1.3.6.1.4.1.9.9.719.1.1.1.1.2
      type: DisplayString
    - labels:
      - cucsFaultTags
      labelname: cucsFaultTags
      oid: 1.3.6.1.4.1.9.9.719.1.1.1.1.21
      type: OctetString
  version: 2
  max_repetitions: 8
  retries: 1
  timeout: 2m0s
  auth:
    community: <secret>
    security_level: noAuthNoPriv
    auth_protocol: MD5
    priv_protocol: DES


IndGirl6

unread,
Jun 29, 2018, 10:13:18 AM6/29/18
to Prometheus Users
Ok, an Update. The issue is RESOLVED!!! phew!
all this time i was polling the Virtual IP of the UCS. It looks like we need to poll the individual ip of each FI. Once i started polling the individual IP, it works!!
I knew it was something simple / silly.

Thanks to all that responded and helped with my issue!

Reply all
Reply to author
Forward
0 new messages