Browse Source

Apply Prettier to yaml files (#8969)

Trask Stalnaker 1 year ago
parent
commit
aabd14fb6e
28 changed files with 482 additions and 487 deletions
  1. 12 12
      .github/dependabot.yml
  2. 4 4
      .github/workflows/native-tests-daily.yml
  3. 5 5
      .github/workflows/overhead-benchmark-daily.yml
  4. 2 2
      .github/workflows/pr-smoke-test-fake-backend-images.yml
  5. 2 2
      .github/workflows/pr-smoke-test-grpc-images.yml
  6. 2 2
      .github/workflows/pr-smoke-test-play-images.yml
  7. 2 2
      .github/workflows/pr-smoke-test-quarkus-images.yml
  8. 2 2
      .github/workflows/pr-smoke-test-security-manager-images.yml
  9. 2 2
      .github/workflows/pr-smoke-test-servlet-images.yml
  10. 2 2
      .github/workflows/pr-smoke-test-spring-boot-images.yml
  11. 1 1
      .github/workflows/publish-petclinic-benchmark-image.yml
  12. 2 3
      .github/workflows/publish-smoke-test-fake-backend-images.yml
  13. 2 2
      .github/workflows/publish-smoke-test-grpc-images.yml
  14. 2 2
      .github/workflows/publish-smoke-test-play-images.yml
  15. 2 2
      .github/workflows/publish-smoke-test-quarkus-images.yml
  16. 2 2
      .github/workflows/publish-smoke-test-security-manager-images.yml
  17. 2 2
      .github/workflows/publish-smoke-test-servlet-images.yml
  18. 2 2
      .github/workflows/publish-smoke-test-spring-boot-images.yml
  19. 10 10
      benchmark-overhead/src/test/resources/collector.yaml
  20. 1 1
      instrumentation/grails-3.0/javaagent/src/test/resources/application.yml
  21. 9 10
      instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/activemq.yaml
  22. 61 61
      instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/hadoop.yaml
  23. 2 3
      instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/jetty.yaml
  24. 200 200
      instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/kafka-broker.yaml
  25. 65 65
      instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/tomcat.yaml
  26. 81 81
      instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/wildfly.yaml
  27. 1 2
      instrumentation/spring/spring-boot-resources/library/src/test/resources/application-multi.yml
  28. 2 3
      instrumentation/spring/spring-boot-resources/library/src/test/resources/application.yml

+ 12 - 12
.github/dependabot.yml

@@ -21,26 +21,26 @@ updates:
     ignore:
       - dependency-name: "io.opentelemetry:*"
         # OpenTelemetry SDK updates are handled by auto-update-otel-sdk.yml
-        versions: [ "(,)" ]
+        versions: ["(,)"]
       - dependency-name: "ch.qos.logback:logback-classic"
         # logback 1.4+ requires Java 11
-        versions: [ "[1.4,)" ]
+        versions: ["[1.4,)"]
       - dependency-name: "com.bmuschko.docker-remote-api"
         # Publishing Servlet images for smoke tests is failing starting from 9.0.0
         # (see https://github.com/open-telemetry/opentelemetry-java-instrumentation/issues/7185)
         # in particular, these commands are failing (reproducible locally):
         # ./gradlew :smoke-tests:images:servlet:buildLinuxTestImages pushMatrix -PsmokeTestServer=jetty
         # ./gradlew :smoke-tests:images:servlet:buildWindowsTestImages pushMatrix -PsmokeTestServer=jetty
-        versions: [ "[9,)" ]
+        versions: ["[9,)"]
       - dependency-name: "net.sf.jt400:jt400"
         # using old version of this obscure artifact to test instrumentation of Java 1.1 bytecode
-        versions: [ "(,)" ]
+        versions: ["(,)"]
       - dependency-name: "org.mockito:*"
         # mockito 5 requires Java 11
-        versions: [ "[5,)" ]
+        versions: ["[5,)"]
       - dependency-name: "org.junit-pioneer:junit-pioneer"
         # junit-pioneer 2.x requires Java 11
-        versions: [ "[1,)" ]
+        versions: ["[1,)"]
     rebase-strategy: "disabled"
     schedule:
       interval: "daily"
@@ -53,15 +53,15 @@ updates:
     ignore:
       - dependency-name: "javax.servlet:javax.servlet-api"
         # examples are intentionally compiled and tested against Servlet 3.0
-        versions: [ "[3.1,)" ]
+        versions: ["[3.1,)"]
       - dependency-name: "org.eclipse.jetty:jetty-server"
         # examples are intentionally compiled and tested against Servlet 3.0
         # "9-alpha" is needed to precede all 9.0.0.v{DATE} versions
-        versions: [ "[9-alpha,)" ]
+        versions: ["[9-alpha,)"]
       - dependency-name: "org.eclipse.jetty:jetty-servlet"
         # examples are intentionally compiled and tested against Servlet 3.0
         # "9-alpha" is needed to precede all 9.0.0.v{DATE} versions
-        versions: [ "[9-alpha,)" ]
+        versions: ["[9-alpha,)"]
     rebase-strategy: "disabled"
     schedule:
       interval: "daily"
@@ -74,15 +74,15 @@ updates:
     ignore:
       - dependency-name: "javax.servlet:javax.servlet-api"
         # examples are intentionally compiled and tested against Servlet 3.0
-        versions: [ "[3.1,)" ]
+        versions: ["[3.1,)"]
       - dependency-name: "org.eclipse.jetty:jetty-server"
         # examples are intentionally compiled and tested against Servlet 3.0
         # "9-alpha" is needed to precede all 9.0.0.v{DATE} versions
-        versions: [ "[9-alpha,)" ]
+        versions: ["[9-alpha,)"]
       - dependency-name: "org.eclipse.jetty:jetty-servlet"
         # examples are intentionally compiled and tested against Servlet 3.0
         # "9-alpha" is needed to precede all 9.0.0.v{DATE} versions
-        versions: [ "[9-alpha,)" ]
+        versions: ["[9-alpha,)"]
     rebase-strategy: "disabled"
     schedule:
       interval: "daily"

+ 4 - 4
.github/workflows/native-tests-daily.yml

@@ -3,7 +3,7 @@ name: GraalVM native tests
 on:
   schedule:
     # daily at 4:00 UTC
-    - cron: '0 4 * * *'
+    - cron: "0 4 * * *"
   workflow_dispatch:
 
 jobs:
@@ -13,9 +13,9 @@ jobs:
       - uses: actions/checkout@v3
       - uses: graalvm/setup-graalvm@v1
         with:
-          version: 'latest'
-          java-version: '17'
-          components: 'native-image'
+          version: "latest"
+          java-version: "17"
+          components: "native-image"
       - name: Running test
         run: |
           echo "GRAALVM_HOME: $GRAALVM_HOME"

+ 5 - 5
.github/workflows/overhead-benchmark-daily.yml

@@ -2,7 +2,7 @@ name: Overhead benchmark (daily)
 on:
   schedule:
     # daily at 5:00 UTC
-    - cron: '0 5 * * *'
+    - cron: "0 5 * * *"
   workflow_dispatch:
 
 jobs:
@@ -36,10 +36,10 @@ jobs:
       - name: Commit updated results
         uses: EndBug/add-and-commit@v9
         with:
-          add: 'benchmark-overhead/results'
-          cwd: './gh-pages'
-          branch: 'gh-pages'
-          message: 'update test result data'
+          add: "benchmark-overhead/results"
+          cwd: "./gh-pages"
+          branch: "gh-pages"
+          message: "update test result data"
           author_name: opentelemetrybot
           author_email: 107717825+opentelemetrybot@users.noreply.github.com
           committer_name: opentelemetrybot

+ 2 - 2
.github/workflows/pr-smoke-test-fake-backend-images.yml

@@ -3,8 +3,8 @@ name: PR build fake backend images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/fake-backend/**'
-      - '.github/workflows/pr-smoke-test-fake-backend-images.yml'
+      - "smoke-tests/images/fake-backend/**"
+      - ".github/workflows/pr-smoke-test-fake-backend-images.yml"
 
 jobs:
   buildLinux:

+ 2 - 2
.github/workflows/pr-smoke-test-grpc-images.yml

@@ -3,8 +3,8 @@ name: PR build gRPC images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/grpc/**'
-      - '.github/workflows/pr-smoke-test-grpc-images.yml'
+      - "smoke-tests/images/grpc/**"
+      - ".github/workflows/pr-smoke-test-grpc-images.yml"
 
 jobs:
   build:

+ 2 - 2
.github/workflows/pr-smoke-test-play-images.yml

@@ -3,8 +3,8 @@ name: PR build Play images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/play/**'
-      - '.github/workflows/pr-smoke-test-play-images.yml'
+      - "smoke-tests/images/play/**"
+      - ".github/workflows/pr-smoke-test-play-images.yml"
 
 jobs:
   build:

+ 2 - 2
.github/workflows/pr-smoke-test-quarkus-images.yml

@@ -3,8 +3,8 @@ name: PR build Quarkus images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/quarkus/**'
-      - '.github/workflows/pr-smoke-test-quarkus-images.yml'
+      - "smoke-tests/images/quarkus/**"
+      - ".github/workflows/pr-smoke-test-quarkus-images.yml"
 
 jobs:
   build:

+ 2 - 2
.github/workflows/pr-smoke-test-security-manager-images.yml

@@ -3,8 +3,8 @@ name: PR build Security Manager test app images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/security-manager/**'
-      - '.github/workflows/pr-smoke-test-security-manager-images.yml'
+      - "smoke-tests/images/security-manager/**"
+      - ".github/workflows/pr-smoke-test-security-manager-images.yml"
 
 jobs:
   build:

+ 2 - 2
.github/workflows/pr-smoke-test-servlet-images.yml

@@ -3,8 +3,8 @@ name: PR build Servlet images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/servlet/**'
-      - '.github/workflows/pr-smoke-test-servlet-images.yml'
+      - "smoke-tests/images/servlet/**"
+      - ".github/workflows/pr-smoke-test-servlet-images.yml"
 
 jobs:
   build:

+ 2 - 2
.github/workflows/pr-smoke-test-spring-boot-images.yml

@@ -3,8 +3,8 @@ name: PR build Spring Boot images for smoke tests
 on:
   pull_request:
     paths:
-      - 'smoke-tests/images/spring-boot/**'
-      - '.github/workflows/pr-smoke-test-spring-boot-images.yml'
+      - "smoke-tests/images/spring-boot/**"
+      - ".github/workflows/pr-smoke-test-spring-boot-images.yml"
 
 jobs:
   build:

+ 1 - 1
.github/workflows/publish-petclinic-benchmark-image.yml

@@ -2,7 +2,7 @@ name: Publish PetClinic benchmark image
 on:
   push:
     paths:
-      - 'benchmark-overhead/Dockerfile-petclinic-base'
+      - "benchmark-overhead/Dockerfile-petclinic-base"
     branches:
       - main
   workflow_dispatch:

+ 2 - 3
.github/workflows/publish-smoke-test-fake-backend-images.yml

@@ -3,8 +3,8 @@ name: Publish fake backend images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/fake-backend/**'
-      - '.github/workflows/publish-smoke-test-fake-backend-images.yml'
+      - "smoke-tests/images/fake-backend/**"
+      - ".github/workflows/publish-smoke-test-fake-backend-images.yml"
     branches:
       - main
   workflow_dispatch:
@@ -80,4 +80,3 @@ jobs:
           needs.publishLinux.result == 'success' &&
           needs.publishWindows.result == 'success'
         }}
-

+ 2 - 2
.github/workflows/publish-smoke-test-grpc-images.yml

@@ -3,8 +3,8 @@ name: Publish gRPC images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/grpc/**'
-      - '.github/workflows/publish-smoke-test-grpc-images.yml'
+      - "smoke-tests/images/grpc/**"
+      - ".github/workflows/publish-smoke-test-grpc-images.yml"
     branches:
       - main
   workflow_dispatch:

+ 2 - 2
.github/workflows/publish-smoke-test-play-images.yml

@@ -3,8 +3,8 @@ name: Publish Play images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/play/**'
-      - '.github/workflows/publish-smoke-test-play-images.yml'
+      - "smoke-tests/images/play/**"
+      - ".github/workflows/publish-smoke-test-play-images.yml"
     branches:
       - main
   workflow_dispatch:

+ 2 - 2
.github/workflows/publish-smoke-test-quarkus-images.yml

@@ -3,8 +3,8 @@ name: Publish Quarkus images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/quarkus/**'
-      - '.github/workflows/publish-smoke-test-quarkus-images.yml'
+      - "smoke-tests/images/quarkus/**"
+      - ".github/workflows/publish-smoke-test-quarkus-images.yml"
     branches:
       - main
   workflow_dispatch:

+ 2 - 2
.github/workflows/publish-smoke-test-security-manager-images.yml

@@ -3,8 +3,8 @@ name: Publish Security Manager test app images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/security-manager/**'
-      - '.github/workflows/publish-smoke-test-security-manager-images.yml'
+      - "smoke-tests/images/security-manager/**"
+      - ".github/workflows/publish-smoke-test-security-manager-images.yml"
     branches:
       - main
   workflow_dispatch:

+ 2 - 2
.github/workflows/publish-smoke-test-servlet-images.yml

@@ -3,8 +3,8 @@ name: Publish Servlet images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/servlet/**'
-      - '.github/workflows/publish-smoke-test-servlet-images.yml'
+      - "smoke-tests/images/servlet/**"
+      - ".github/workflows/publish-smoke-test-servlet-images.yml"
     branches:
       - main
   workflow_dispatch:

+ 2 - 2
.github/workflows/publish-smoke-test-spring-boot-images.yml

@@ -3,8 +3,8 @@ name: Publish Spring Boot images for smoke tests
 on:
   push:
     paths:
-      - 'smoke-tests/images/spring-boot/**'
-      - '.github/workflows/publish-smoke-test-spring-boot-images.yml'
+      - "smoke-tests/images/spring-boot/**"
+      - ".github/workflows/publish-smoke-test-spring-boot-images.yml"
     branches:
       - main
   workflow_dispatch:

+ 10 - 10
benchmark-overhead/src/test/resources/collector.yaml

@@ -18,16 +18,16 @@ exporters:
 service:
   pipelines:
     traces:
-      receivers: [ otlp ]
-      processors: [ batch ]
-      exporters: [ logging/logging_info ]
+      receivers: [otlp]
+      processors: [batch]
+      exporters: [logging/logging_info]
     metrics:
-      receivers: [ otlp ]
-      processors: [ batch ]
-      exporters: [ logging/logging_info ]
+      receivers: [otlp]
+      processors: [batch]
+      exporters: [logging/logging_info]
     logs:
-      receivers: [ otlp ]
-      processors: [ batch ]
-      exporters: [ logging/logging_info  ]
+      receivers: [otlp]
+      processors: [batch]
+      exporters: [logging/logging_info]
 
-  extensions: [ health_check ]
+  extensions: [health_check]

+ 1 - 1
instrumentation/grails-3.0/javaagent/src/test/resources/application.yml

@@ -5,4 +5,4 @@ grails:
 spring:
   groovy:
     template:
-      check-template-location: false
+      check-template-location: false

+ 9 - 10
instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/activemq.yaml

@@ -1,6 +1,5 @@
 ---
 rules:
-
   - beans:
       - org.apache.activemq:type=Broker,brokerName=*,destinationType=Queue,destinationName=*
       - org.apache.activemq:type=Broker,brokerName=*,destinationType=Topic,destinationName=*
@@ -10,36 +9,36 @@ rules:
     prefix: activemq.
     mapping:
       ProducerCount:
-        unit: '{producers}'
+        unit: "{producers}"
         type: updowncounter
         desc: The number of producers attached to this destination
       ConsumerCount:
-        unit: '{consumers}'
+        unit: "{consumers}"
         type: updowncounter
         desc: The number of consumers subscribed to this destination
       MemoryPercentUsage:
         metric: memory.MemoryPercentUsage
-        unit: '%'
+        unit: "%"
         type: gauge
         desc: The percentage of configured memory used
       QueueSize:
         metric: message.QueueSize
-        unit: '{messages}'
+        unit: "{messages}"
         type: updowncounter
         desc: The current number of messages waiting to be consumed
       ExpiredCount:
         metric: message.ExpiredCount
-        unit: '{messages}'
+        unit: "{messages}"
         type: counter
         desc: The number of messages not delivered because they expired
       EnqueueCount:
         metric: message.EnqueueCount
-        unit: '{messages}'
+        unit: "{messages}"
         type: counter
         desc: The number of messages sent to this destination
       DequeueCount:
         metric: message.DequeueCount
-        unit: '{messages}'
+        unit: "{messages}"
         type: counter
         desc: The number of messages acknowledged and removed from this destination
       AverageEnqueueTime:
@@ -52,13 +51,13 @@ rules:
     metricAttribute:
       broker: param(brokerName)
     prefix: activemq.
-    unit: '%'
+    unit: "%"
     type: gauge
     mapping:
       CurrentConnectionsCount:
         metric: connections.CurrentConnectionsCount
         type: updowncounter
-        unit: '{connections}'
+        unit: "{connections}"
         desc: The total number of current connections
       StorePercentUsage:
         metric: disc.StorePercentUsage

+ 61 - 61
instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/hadoop.yaml

@@ -1,63 +1,63 @@
 ---
 rules:
- - bean: Hadoop:service=NameNode,name=FSNamesystem
-   unit: "1"
-   prefix: hadoop.
-   metricAttribute:
-     node_name: param(tag.Hostname)
-   mapping:
-     CapacityUsed:
-       metric: capacity.CapacityUsed
-       type: updowncounter
-       unit: By
-       desc: Current used capacity across all data nodes
-     CapacityTotal:
-       metric: capacity.CapacityTotal
-       type: updowncounter
-       unit: By
-     BlocksTotal:
-       metric: block.BlocksTotal
-       type: updowncounter
-       unit: '{blocks}'
-       desc: Current number of allocated blocks in the system
-     MissingBlocks:
-       metric: block.MissingBlocks
-       type: updowncounter
-       unit: '{blocks}'
-       desc: Current number of missing blocks
-     CorruptBlocks:
-       metric: block.CorruptBlocks
-       type: updowncounter
-       unit: '{blocks}'
-       desc: Current number of blocks with corrupt replicas
-     VolumeFailuresTotal:
-       metric: volume.VolumeFailuresTotal
-       type: updowncounter
-       unit: '{volumes}'
-       desc: Total number of volume failures across all data nodes
-       metricAttribute:
-         direction: const(sent)
-     FilesTotal:
-       metric: file.FilesTotal
-       type: updowncounter
-       unit: '{files}'
-       desc: Current number of files and directories
-     TotalLoad:
-       metric: file.TotalLoad
-       type: updowncounter
-       unit: '{operations}'
-       desc: Current number of connections
-     NumLiveDataNodes:
-       metric: datenode.Count
-       type: updowncounter
-       unit: '{nodes}'
-       desc: The Number of data nodes
-       metricAttribute:
-         state: const(live)
-     NumDeadDataNodes:
-       metric: datenode.Count
-       type: updowncounter
-       unit: '{nodes}'
-       desc: The Number of data nodes
-       metricAttribute:
-         state: const(dead)
+  - bean: Hadoop:service=NameNode,name=FSNamesystem
+    unit: "1"
+    prefix: hadoop.
+    metricAttribute:
+      node_name: param(tag.Hostname)
+    mapping:
+      CapacityUsed:
+        metric: capacity.CapacityUsed
+        type: updowncounter
+        unit: By
+        desc: Current used capacity across all data nodes
+      CapacityTotal:
+        metric: capacity.CapacityTotal
+        type: updowncounter
+        unit: By
+      BlocksTotal:
+        metric: block.BlocksTotal
+        type: updowncounter
+        unit: "{blocks}"
+        desc: Current number of allocated blocks in the system
+      MissingBlocks:
+        metric: block.MissingBlocks
+        type: updowncounter
+        unit: "{blocks}"
+        desc: Current number of missing blocks
+      CorruptBlocks:
+        metric: block.CorruptBlocks
+        type: updowncounter
+        unit: "{blocks}"
+        desc: Current number of blocks with corrupt replicas
+      VolumeFailuresTotal:
+        metric: volume.VolumeFailuresTotal
+        type: updowncounter
+        unit: "{volumes}"
+        desc: Total number of volume failures across all data nodes
+        metricAttribute:
+          direction: const(sent)
+      FilesTotal:
+        metric: file.FilesTotal
+        type: updowncounter
+        unit: "{files}"
+        desc: Current number of files and directories
+      TotalLoad:
+        metric: file.TotalLoad
+        type: updowncounter
+        unit: "{operations}"
+        desc: Current number of connections
+      NumLiveDataNodes:
+        metric: datenode.Count
+        type: updowncounter
+        unit: "{nodes}"
+        desc: The Number of data nodes
+        metricAttribute:
+          state: const(live)
+      NumDeadDataNodes:
+        metric: datenode.Count
+        type: updowncounter
+        unit: "{nodes}"
+        desc: The Number of data nodes
+        metricAttribute:
+          state: const(dead)

+ 2 - 3
instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/jetty.yaml

@@ -1,6 +1,5 @@
 ---
 rules:
-
   - bean: org.eclipse.jetty.server.session:context=*,type=sessionhandler,id=*
     unit: s
     prefix: jetty.session.
@@ -9,7 +8,7 @@ rules:
       resource: param(context)
     mapping:
       sessionsCreated:
-        unit: '{sessions}'
+        unit: "{sessions}"
         type: counter
         desc: The number of sessions established in total
       sessionTimeTotal:
@@ -24,7 +23,7 @@ rules:
 
   - bean: org.eclipse.jetty.util.thread:type=queuedthreadpool,id=*
     prefix: jetty.threads.
-    unit: '{threads}'
+    unit: "{threads}"
     type: updowncounter
     mapping:
       busyThreads:

+ 200 - 200
instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/kafka-broker.yaml

@@ -2,203 +2,203 @@
 rules:
   # Broker metrics
 
- - bean: kafka.server:type=BrokerTopicMetrics,name=MessagesInPerSec
-   mapping:
-     Count:
-       metric: kafka.message.count
-       type: counter
-       desc: The number of messages received by the broker
-       unit: '{messages}'
-
- - bean: kafka.server:type=BrokerTopicMetrics,name=TotalFetchRequestsPerSec
-   metricAttribute:
-     type: const(fetch)
-   mapping:
-     Count:
-       metric: kafka.request.count
-       type: counter
-       desc: The number of requests received by the broker
-       unit: '{requests}'
-
- - bean: kafka.server:type=BrokerTopicMetrics,name=TotalProduceRequestsPerSec
-   metricAttribute:
-     type: const(produce)
-   mapping:
-     Count:
-       metric: kafka.request.count
-       type: counter
-       desc: The number of requests received by the broker
-       unit: '{requests}'
-
- - bean: kafka.server:type=BrokerTopicMetrics,name=FailedFetchRequestsPerSec
-   metricAttribute:
-     type: const(fetch)
-   mapping:
-     Count:
-       metric: kafka.request.failed
-       type: counter
-       desc: The number of requests to the broker resulting in a failure
-       unit: '{requests}'
-
- - bean: kafka.server:type=BrokerTopicMetrics,name=FailedProduceRequestsPerSec
-   metricAttribute:
-     type: const(produce)
-   mapping:
-     Count:
-       metric: kafka.request.failed
-       type: counter
-       desc: The number of requests to the broker resulting in a failure
-       unit: '{requests}'
-
- - beans:
-     - kafka.network:type=RequestMetrics,name=TotalTimeMs,request=Produce
-     - kafka.network:type=RequestMetrics,name=TotalTimeMs,request=FetchConsumer
-     - kafka.network:type=RequestMetrics,name=TotalTimeMs,request=FetchFollower
-   metricAttribute:
-     type: param(request)
-   unit: ms
-   mapping:
-     Count:
-       metric: kafka.request.time.total
-       type: counter
-       desc: The total time the broker has taken to service requests
-     50thPercentile:
-       metric: kafka.request.time.50p
-       type: gauge
-       desc: The 50th percentile time the broker has taken to service requests
-     99thPercentile:
-       metric: kafka.request.time.99p
-       type: gauge
-       desc: The 99th percentile time the broker has taken to service requests
-
- - bean: kafka.network:type=RequestChannel,name=RequestQueueSize
-   mapping:
-     Value:
-       metric: kafka.request.queue
-       type: updowncounter
-       desc: Size of the request queue
-       unit: '{requests}'
-
- - bean: kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec
-   metricAttribute:
-     direction: const(in)
-   mapping:
-     Count:
-       metric: kafka.network.io
-       type: counter
-       desc: The bytes received or sent by the broker
-       unit: By
-
- - bean: kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec
-   metricAttribute:
-     direction: const(out)
-   mapping:
-     Count:
-       metric: kafka.network.io
-       type: counter
-       desc: The bytes received or sent by the broker
-       unit: By
-
- - beans:
-     - kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Produce
-     - kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Fetch
-   metricAttribute:
-     type: param(delayedOperation)
-   mapping:
-     Value:
-       metric: kafka.purgatory.size
-       type: updowncounter
-       desc: The number of requests waiting in purgatory
-       unit: '{requests}'
-
- - bean: kafka.server:type=ReplicaManager,name=PartitionCount
-   mapping:
-     Value:
-       metric: kafka.partition.count
-       type: updowncounter
-       desc: The number of partitions on the broker
-       unit: '{partitions}'
-
- - bean: kafka.controller:type=KafkaController,name=OfflinePartitionsCount
-   mapping:
-     Value:
-       metric: kafka.partition.offline
-       type: updowncounter
-       desc: The number of partitions offline
-       unit: '{partitions}'
-
- - bean: kafka.server:type=ReplicaManager,name=UnderReplicatedPartitions
-   mapping:
-     Value:
-       metric: kafka.partition.underReplicated
-       type: updowncounter
-       desc: The number of under replicated partitions
-       unit: '{partitions}'
-
- - bean: kafka.server:type=ReplicaManager,name=IsrShrinksPerSec
-   metricAttribute:
-     operation: const(shrink)
-   mapping:
-     Count:
-       metric: kafka.isr.operation.count
-       type: updowncounter
-       desc: The number of in-sync replica shrink and expand operations
-       unit: '{operations}'
-
- - bean: kafka.server:type=ReplicaManager,name=IsrExpandsPerSec
-   metricAttribute:
-     operation: const(expand)
-   mapping:
-     Count:
-       metric: kafka.isr.operation.count
-       type: updowncounter
-       desc: The number of in-sync replica shrink and expand operations
-       unit: '{operations}'
-
- - bean: kafka.server:type=ReplicaFetcherManager,name=MaxLag,clientId=Replica
-   mapping:
-     Value:
-       metric: kafka.lag.max
-       desc: The max lag in messages between follower and leader replicas
-       unit: '{messages}'
-
- - bean: kafka.controller:type=KafkaController,name=ActiveControllerCount
-   mapping:
-     Value:
-       metric: kafka.controller.active.count
-       type: updowncounter
-       desc: The number of controllers active on the broker
-       unit: '{controllers}'
-
- - bean: kafka.controller:type=ControllerStats,name=LeaderElectionRateAndTimeMs
-   mapping:
-     Count:
-       metric: kafka.leaderElection.count
-       type: counter
-       desc: The leader election count
-       unit: '{elections}'
-
- - bean: kafka.controller:type=ControllerStats,name=UncleanLeaderElectionsPerSec
-   mapping:
-     Count:
-       metric: kafka.leaderElection.unclean.count
-       type: counter
-       desc: Unclean leader election count - increasing indicates broker failures
-       unit: '{elections}'
-
- # Log metrics
-
- - bean: kafka.log:type=LogFlushStats,name=LogFlushRateAndTimeMs
-   unit: ms
-   type: gauge
-   prefix: kafka.logs.flush.
-   mapping:
-     Count:
-       type: counter
-       desc: Log flush count
-     50thPercentile:
-       metric: time.50p
-       desc: Log flush time - 50th percentile
-     99thPercentile:
-       metric: time.99p
-       desc: Log flush time - 99th percentile
+  - bean: kafka.server:type=BrokerTopicMetrics,name=MessagesInPerSec
+    mapping:
+      Count:
+        metric: kafka.message.count
+        type: counter
+        desc: The number of messages received by the broker
+        unit: "{messages}"
+
+  - bean: kafka.server:type=BrokerTopicMetrics,name=TotalFetchRequestsPerSec
+    metricAttribute:
+      type: const(fetch)
+    mapping:
+      Count:
+        metric: kafka.request.count
+        type: counter
+        desc: The number of requests received by the broker
+        unit: "{requests}"
+
+  - bean: kafka.server:type=BrokerTopicMetrics,name=TotalProduceRequestsPerSec
+    metricAttribute:
+      type: const(produce)
+    mapping:
+      Count:
+        metric: kafka.request.count
+        type: counter
+        desc: The number of requests received by the broker
+        unit: "{requests}"
+
+  - bean: kafka.server:type=BrokerTopicMetrics,name=FailedFetchRequestsPerSec
+    metricAttribute:
+      type: const(fetch)
+    mapping:
+      Count:
+        metric: kafka.request.failed
+        type: counter
+        desc: The number of requests to the broker resulting in a failure
+        unit: "{requests}"
+
+  - bean: kafka.server:type=BrokerTopicMetrics,name=FailedProduceRequestsPerSec
+    metricAttribute:
+      type: const(produce)
+    mapping:
+      Count:
+        metric: kafka.request.failed
+        type: counter
+        desc: The number of requests to the broker resulting in a failure
+        unit: "{requests}"
+
+  - beans:
+      - kafka.network:type=RequestMetrics,name=TotalTimeMs,request=Produce
+      - kafka.network:type=RequestMetrics,name=TotalTimeMs,request=FetchConsumer
+      - kafka.network:type=RequestMetrics,name=TotalTimeMs,request=FetchFollower
+    metricAttribute:
+      type: param(request)
+    unit: ms
+    mapping:
+      Count:
+        metric: kafka.request.time.total
+        type: counter
+        desc: The total time the broker has taken to service requests
+      50thPercentile:
+        metric: kafka.request.time.50p
+        type: gauge
+        desc: The 50th percentile time the broker has taken to service requests
+      99thPercentile:
+        metric: kafka.request.time.99p
+        type: gauge
+        desc: The 99th percentile time the broker has taken to service requests
+
+  - bean: kafka.network:type=RequestChannel,name=RequestQueueSize
+    mapping:
+      Value:
+        metric: kafka.request.queue
+        type: updowncounter
+        desc: Size of the request queue
+        unit: "{requests}"
+
+  - bean: kafka.server:type=BrokerTopicMetrics,name=BytesInPerSec
+    metricAttribute:
+      direction: const(in)
+    mapping:
+      Count:
+        metric: kafka.network.io
+        type: counter
+        desc: The bytes received or sent by the broker
+        unit: By
+
+  - bean: kafka.server:type=BrokerTopicMetrics,name=BytesOutPerSec
+    metricAttribute:
+      direction: const(out)
+    mapping:
+      Count:
+        metric: kafka.network.io
+        type: counter
+        desc: The bytes received or sent by the broker
+        unit: By
+
+  - beans:
+      - kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Produce
+      - kafka.server:type=DelayedOperationPurgatory,name=PurgatorySize,delayedOperation=Fetch
+    metricAttribute:
+      type: param(delayedOperation)
+    mapping:
+      Value:
+        metric: kafka.purgatory.size
+        type: updowncounter
+        desc: The number of requests waiting in purgatory
+        unit: "{requests}"
+
+  - bean: kafka.server:type=ReplicaManager,name=PartitionCount
+    mapping:
+      Value:
+        metric: kafka.partition.count
+        type: updowncounter
+        desc: The number of partitions on the broker
+        unit: "{partitions}"
+
+  - bean: kafka.controller:type=KafkaController,name=OfflinePartitionsCount
+    mapping:
+      Value:
+        metric: kafka.partition.offline
+        type: updowncounter
+        desc: The number of partitions offline
+        unit: "{partitions}"
+
+  - bean: kafka.server:type=ReplicaManager,name=UnderReplicatedPartitions
+    mapping:
+      Value:
+        metric: kafka.partition.underReplicated
+        type: updowncounter
+        desc: The number of under replicated partitions
+        unit: "{partitions}"
+
+  - bean: kafka.server:type=ReplicaManager,name=IsrShrinksPerSec
+    metricAttribute:
+      operation: const(shrink)
+    mapping:
+      Count:
+        metric: kafka.isr.operation.count
+        type: updowncounter
+        desc: The number of in-sync replica shrink and expand operations
+        unit: "{operations}"
+
+  - bean: kafka.server:type=ReplicaManager,name=IsrExpandsPerSec
+    metricAttribute:
+      operation: const(expand)
+    mapping:
+      Count:
+        metric: kafka.isr.operation.count
+        type: updowncounter
+        desc: The number of in-sync replica shrink and expand operations
+        unit: "{operations}"
+
+  - bean: kafka.server:type=ReplicaFetcherManager,name=MaxLag,clientId=Replica
+    mapping:
+      Value:
+        metric: kafka.lag.max
+        desc: The max lag in messages between follower and leader replicas
+        unit: "{messages}"
+
+  - bean: kafka.controller:type=KafkaController,name=ActiveControllerCount
+    mapping:
+      Value:
+        metric: kafka.controller.active.count
+        type: updowncounter
+        desc: The number of controllers active on the broker
+        unit: "{controllers}"
+
+  - bean: kafka.controller:type=ControllerStats,name=LeaderElectionRateAndTimeMs
+    mapping:
+      Count:
+        metric: kafka.leaderElection.count
+        type: counter
+        desc: The leader election count
+        unit: "{elections}"
+
+  - bean: kafka.controller:type=ControllerStats,name=UncleanLeaderElectionsPerSec
+    mapping:
+      Count:
+        metric: kafka.leaderElection.unclean.count
+        type: counter
+        desc: Unclean leader election count - increasing indicates broker failures
+        unit: "{elections}"
+
+  # Log metrics
+
+  - bean: kafka.log:type=LogFlushStats,name=LogFlushRateAndTimeMs
+    unit: ms
+    type: gauge
+    prefix: kafka.logs.flush.
+    mapping:
+      Count:
+        type: counter
+        desc: Log flush count
+      50thPercentile:
+        metric: time.50p
+        desc: Log flush time - 50th percentile
+      99thPercentile:
+        metric: time.99p
+        desc: Log flush time - 99th percentile

+ 65 - 65
instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/tomcat.yaml

@@ -1,67 +1,67 @@
 ---
 rules:
- - bean: Catalina:type=GlobalRequestProcessor,name=*
-   unit: "1"
-   prefix: http.server.tomcat.
-   metricAttribute:
-     name: param(name)
-   mapping:
-     errorCount:
-       metric: errorCount
-       type: gauge
-       desc: The number of errors per second on all request processors
-     requestCount:
-       metric: requestCount
-       type: gauge
-       desc: The number of requests per second across all request processors
-     maxTime:
-       metric: maxTime
-       type: gauge
-       unit: ms
-       desc: The longest request processing time
-     processingTime:
-       metric: processingTime
-       type: counter
-       unit: ms
-       desc: Total time for processing all requests
-     bytesReceived:
-       metric: traffic
-       type: counter
-       unit: By
-       desc: The number of bytes transmitted
-       metricAttribute:
-         direction: const(received)
-     bytesSent:
-       metric: traffic
-       type: counter
-       unit: By
-       desc: The number of bytes transmitted
-       metricAttribute:
-         direction: const(sent)
- - bean: Catalina:type=Manager,host=localhost,context=*
-   unit: "1"
-   prefix: http.server.tomcat.
-   type: updowncounter
-   metricAttribute:
-     context: param(context)
-   mapping:
-     activeSessions:
-       metric: sessions.activeSessions
-       desc: The number of active sessions
- - bean: Catalina:type=ThreadPool,name=*
-   unit: '{threads}'
-   prefix: http.server.tomcat.
-   type: updowncounter
-   metricAttribute:
-     name: param(name)
-   mapping:
-     currentThreadCount:
-       metric: threads
-       desc: Thread Count of the Thread Pool
-       metricAttribute:
-         state: const(idle)
-     currentThreadsBusy:
-       metric: threads
-       desc: Thread Count of the Thread Pool
-       metricAttribute:
-         state: const(busy)
+  - bean: Catalina:type=GlobalRequestProcessor,name=*
+    unit: "1"
+    prefix: http.server.tomcat.
+    metricAttribute:
+      name: param(name)
+    mapping:
+      errorCount:
+        metric: errorCount
+        type: gauge
+        desc: The number of errors per second on all request processors
+      requestCount:
+        metric: requestCount
+        type: gauge
+        desc: The number of requests per second across all request processors
+      maxTime:
+        metric: maxTime
+        type: gauge
+        unit: ms
+        desc: The longest request processing time
+      processingTime:
+        metric: processingTime
+        type: counter
+        unit: ms
+        desc: Total time for processing all requests
+      bytesReceived:
+        metric: traffic
+        type: counter
+        unit: By
+        desc: The number of bytes transmitted
+        metricAttribute:
+          direction: const(received)
+      bytesSent:
+        metric: traffic
+        type: counter
+        unit: By
+        desc: The number of bytes transmitted
+        metricAttribute:
+          direction: const(sent)
+  - bean: Catalina:type=Manager,host=localhost,context=*
+    unit: "1"
+    prefix: http.server.tomcat.
+    type: updowncounter
+    metricAttribute:
+      context: param(context)
+    mapping:
+      activeSessions:
+        metric: sessions.activeSessions
+        desc: The number of active sessions
+  - bean: Catalina:type=ThreadPool,name=*
+    unit: "{threads}"
+    prefix: http.server.tomcat.
+    type: updowncounter
+    metricAttribute:
+      name: param(name)
+    mapping:
+      currentThreadCount:
+        metric: threads
+        desc: Thread Count of the Thread Pool
+        metricAttribute:
+          state: const(idle)
+      currentThreadsBusy:
+        metric: threads
+        desc: Thread Count of the Thread Pool
+        metricAttribute:
+          state: const(busy)

+ 81 - 81
instrumentation/jmx-metrics/javaagent/src/main/resources/jmx/rules/wildfly.yaml

@@ -1,83 +1,83 @@
 ---
 rules:
- - bean: jboss.as:deployment=*,subsystem=undertow
-   metricAttribute:
-     deployment: param(deployment)
-   prefix: wildfly.session.
-   type: counter
-   unit: "1"
-   mapping:
-     sessionsCreated:
-     activeSessions:
-       type: updowncounter
-     expiredSessions:
-     rejectedSessions:
- - bean: jboss.as:subsystem=undertow,server=*,http-listener=*
-   metricAttribute:
-     server: param(server)
-     listener: param(http-listener)
-   prefix: wildfly.request.
-   type: counter
-   unit: "1"
-   mapping:
-     requestCount:
-     processingTime:
-       unit: ns
-     errorCount:
- - bean: jboss.as:subsystem=undertow,server=*,http-listener=*
-   metricAttribute:
-     server: param(server)
-     listener: param(http-listener)
-   type: counter
-   unit: By
-   mapping:
-     bytesSent:
-       metric: wildfly.network.io
-       desc: Total number of bytes transferred
-       metricAttribute:
-         direction: const(out)
-     bytesReceived:
-       metric: wildfly.network.io
-       desc: Total number of bytes transferred
-       metricAttribute:
-         direction: const(in)
- - bean: jboss.as:subsystem=datasources,data-source=*,statistics=pool
-   unit: "1"
-   metricAttribute:
-     data_source: param(data-source)
-   mapping:
-     ActiveCount:
-       metric: wildfly.db.client.connections.usage
-       metricAttribute:
-         state: const(used)
-       desc: The number of open jdbc connections
-     IdleCount:
-       metric: wildfly.db.client.connections.usage
-       metricAttribute:
-         state: const(idle)
-       desc: The number of open jdbc connections
-     WaitCount:
-       metric: wildfly.db.client.connections.WaitCount
-       type: counter
- - bean: jboss.as:subsystem=transactions
-   type: counter
-   prefix: wildfly.db.client.
-   unit: "{transactions}"
-   mapping:
-     numberOfTransactions:
-       metric: transaction.NumberOfTransactions
-     numberOfApplicationRollbacks:
-       metric: rollback.count
-       metricAttribute:
-         cause: const(application)
-       desc: The total number of transactions rolled back
-     numberOfResourceRollbacks:
-       metric: rollback.count
-       metricAttribute:
-         cause: const(resource)
-       desc: The total number of transactions rolled back
-     numberOfSystemRollbacks:
-       metric: rollback.count
-       metricAttribute:
-         cause: const(system)
-       desc: The total number of transactions rolled back
+  - bean: jboss.as:deployment=*,subsystem=undertow
+    metricAttribute:
+      deployment: param(deployment)
+    prefix: wildfly.session.
+    type: counter
+    unit: "1"
+    mapping:
+      sessionsCreated:
+      activeSessions:
+        type: updowncounter
+      expiredSessions:
+      rejectedSessions:
+  - bean: jboss.as:subsystem=undertow,server=*,http-listener=*
+    metricAttribute:
+      server: param(server)
+      listener: param(http-listener)
+    prefix: wildfly.request.
+    type: counter
+    unit: "1"
+    mapping:
+      requestCount:
+      processingTime:
+        unit: ns
+      errorCount:
+  - bean: jboss.as:subsystem=undertow,server=*,http-listener=*
+    metricAttribute:
+      server: param(server)
+      listener: param(http-listener)
+    type: counter
+    unit: By
+    mapping:
+      bytesSent:
+        metric: wildfly.network.io
+        desc: Total number of bytes transferred
+        metricAttribute:
+          direction: const(out)
+      bytesReceived:
+        metric: wildfly.network.io
+        desc: Total number of bytes transferred
+        metricAttribute:
+          direction: const(in)
+  - bean: jboss.as:subsystem=datasources,data-source=*,statistics=pool
+    unit: "1"
+    metricAttribute:
+      data_source: param(data-source)
+    mapping:
+      ActiveCount:
+        metric: wildfly.db.client.connections.usage
+        metricAttribute:
+          state: const(used)
+        desc: The number of open jdbc connections
+      IdleCount:
+        metric: wildfly.db.client.connections.usage
+        metricAttribute:
+          state: const(idle)
+        desc: The number of open jdbc connections
+      WaitCount:
+        metric: wildfly.db.client.connections.WaitCount
+        type: counter
+  - bean: jboss.as:subsystem=transactions
+    type: counter
+    prefix: wildfly.db.client.
+    unit: "{transactions}"
+    mapping:
+      numberOfTransactions:
+        metric: transaction.NumberOfTransactions
+      numberOfApplicationRollbacks:
+        metric: rollback.count
+        metricAttribute:
+          cause: const(application)
+        desc: The total number of transactions rolled back
+      numberOfResourceRollbacks:
+        metric: rollback.count
+        metricAttribute:
+          cause: const(resource)
+        desc: The total number of transactions rolled back
+      numberOfSystemRollbacks:
+        metric: rollback.count
+        metricAttribute:
+          cause: const(system)
+        desc: The total number of transactions rolled back

+ 1 - 2
instrumentation/spring/spring-boot-resources/library/src/test/resources/application-multi.yml

@@ -1,6 +1,5 @@
 flib:
-  something:
-    12
+  something: 12
 
 section:
   two: 2

+ 2 - 3
instrumentation/spring/spring-boot-resources/library/src/test/resources/application.yml

@@ -1,6 +1,5 @@
 flib:
-  something:
-    12
+  something: 12
 
 section:
   two: 2
@@ -11,4 +10,4 @@ server:
 
 spring:
   application:
-    name: cat-store
+    name: cat-store