diff config/scrape_main.yaml @ 9:17db5e8e7a2f

big rules and scrape config updates
author drewp@bigasterisk.com
date Sun, 04 Dec 2022 02:08:08 -0800
parents ae41d1d0b567
children 2eab3e6b89f2
line wrap: on
line diff
--- a/config/scrape_main.yaml	Sun Dec 04 02:07:14 2022 -0800
+++ b/config/scrape_main.yaml	Sun Dec 04 02:08:08 2022 -0800
@@ -1,25 +1,21 @@
+# see https://relabeler.promlabs.com/
+
 global:
   scrape_interval: 1m
   scrape_timeout: 10s
 
-scrape_config_files:
-  - scrape_ssl.yaml
+# scrape_config_files:
+#   - build/scrape_ssl.yaml
 # These can even be urls: https://github.com/VictoriaMetrics/VictoriaMetrics/blob/master/app/vmagent/README.md#loading-scrape-configs-from-multiple-files
 
 scrape_configs:
   # some based on https://github.com/prometheus/prometheus/blob/main/documentation/examples/prometheus-kubernetes.yml
 
-  - job_name: "victoriametrics"
-    metrics_path: /m/metrics
-    static_configs:
-      - targets:
-          - victoriametrics.default.svc.cluster.local
-
-  - job_name: "vmalert"
-    metrics_path: /vmalert/metrics
-    static_configs:
-      - targets:
-          - vmalert.default.svc.cluster.local
+  # - job_name: "vmalert"
+  #   metrics_path: /vmalert/metrics
+  #   static_configs:
+  #     - targets:
+  #         - vmalert.default.svc.cluster.local
 
   - job_name: "kubernetes-apiservers"
     scheme: https
@@ -64,35 +60,44 @@
       # To omit a service, add this at pod-level (Deployment.spec.template.metadata.annotations):
       #   annotations: { prometheus.io/scrape: "false" }
       - source_labels: [__meta_kubernetes_pod_annotation_prometheus_io_scrape]
-        regex: false
-        action: drop
-
-      - source_labels: [__meta_kubernetes_service_name]
-        regex: kubernetes
+        regex: "false"
         action: drop
 
       - source_labels: [__meta_kubernetes_namespace]
         regex: default
         action: keep
 
-      - source_labels: [__meta_kubernetes_service_name, __meta_kubernetes_pod_container_port_number]
-        regex: "mitmproxy;1008[01]"
-        action: drop
-
+      # promote these to display
       - source_labels: [__meta_kubernetes_service_name]
         target_label: job
 
       - source_labels: [__meta_kubernetes_pod_node_name]
         target_label: node
 
-      - source_labels: [__meta_kubernetes_pod_label_app, __meta_kubernetes_pod_container_port_number]
-        action: drop
-        regex: jsregistry;4873
+      # for convenience in this config
+      - source_labels: [__meta_kubernetes_pod_container_port_number]
+        target_label: __port_number
+
+      # period tweaks
+      - if: '{job="power-eagle"}'
+        action: replace
+        target_label: __scrape_interval__
+        # from powerEagle/private_config.periodSec
+        replacement: 8s
 
-      - source_labels: [__meta_kubernetes_pod_label_app, __meta_kubernetes_pod_container_port_number]
-        action: drop
-        regex: mosquitto-ext;1883
+      # path tweaks
+      - if: '{job="victoriametrics",__port_number="8428"}'
+        action: replace
+        target_label: "__metrics_path__"
+        replacement: "/m/metrics"
 
+      # discovery is matching extra ports that don't serve metrics- remove these targets
+      - {if: '{job="kubernetes"}', action: drop}
+      - {if: '{job="mongodb",           __port_number="27017"}', action: drop}
+      - {if: '{job="mosquitto-ext",     __port_number="1883"}', action: drop}
+      - {if: '{job="filesync-syncthing",__port_number="8384"}', action: drop}
+      - {if: '{job="jsregistry",        __port_number="4873"}', action: drop}
+  
   # # seems like this would match more stuff, but all I get is coredns
   # - job_name: 'old_coredns'
   #   kubernetes_sd_configs: [{role: pod}]
@@ -116,7 +121,12 @@
   - job_name: "net-routes"
     static_configs:
       - targets:
-          - 10.2.0.3:10001
+          - pipe:9999
+
+  - job_name: "net-traffic"
+    static_configs:
+      - targets:
+          - pipe:8080
 
   - job_name: "ping"
     scrape_interval: 2m
@@ -126,14 +136,14 @@
     static_configs:
       - targets:
           # printer, since it falls out of ntop with no traffic at all. Or, we could poll ink status at http://10.2.0.37/general/status.html?pageid=1
-          - 10.2.0.37
+          - printer014032ED
           # frontbed, for monitoring
           - 10.5.0.17
+          # asher bulb, not sure why it sleeps so long
+          - bulb1
 
     relabel_configs:
-      - source_labels: [__address__]
-        target_label: __param_target
-      - source_labels: [__param_target]
-        target_label: instance
+      - {source_labels: [__address__],    target_label: __param_target}
+      - {source_labels: [__param_target], target_label: instance}
       - target_label: __address__
         replacement: prober