diff --git a/otel-integration/CHANGELOG.md b/otel-integration/CHANGELOG.md index 1589af99..0e3466a4 100644 --- a/otel-integration/CHANGELOG.md +++ b/otel-integration/CHANGELOG.md @@ -2,6 +2,11 @@ ## OpenTelemtry-Integration +### v0.0.43 / 2023-12-12 + +- [FIX] Use correct labels for target allocator components. +- [CHORE] Specify replica parameter for target allocator. + ### v0.0.42 / 2023-12-11 - [CHORE] Update collector to `v0.90.1`. diff --git a/otel-integration/k8s-helm/Chart.yaml b/otel-integration/k8s-helm/Chart.yaml index c95d978b..db4cc0da 100644 --- a/otel-integration/k8s-helm/Chart.yaml +++ b/otel-integration/k8s-helm/Chart.yaml @@ -1,7 +1,7 @@ apiVersion: v2 name: otel-integration description: OpenTelemetry Integration -version: 0.0.42 +version: 0.0.43 keywords: - OpenTelemetry Collector - OpenTelemetry Agent @@ -11,17 +11,17 @@ keywords: dependencies: - name: opentelemetry-collector alias: opentelemetry-agent - version: "0.77.7" + version: "0.77.8" repository: https://cgx.jfrog.io/artifactory/coralogix-charts-virtual condition: opentelemetry-agent.enabled - name: opentelemetry-collector alias: opentelemetry-agent-windows - version: "0.77.7" + version: "0.77.8" repository: https://cgx.jfrog.io/artifactory/coralogix-charts-virtual condition: opentelemetry-agent-windows.enabled - name: opentelemetry-collector alias: opentelemetry-cluster-collector - version: "0.77.7" + version: "0.77.8" repository: https://cgx.jfrog.io/artifactory/coralogix-charts-virtual condition: opentelemetry-cluster-collector.enabled sources: diff --git a/otel-integration/k8s-helm/README.md b/otel-integration/k8s-helm/README.md index 1f7bb676..14a75ccd 100644 --- a/otel-integration/k8s-helm/README.md +++ b/otel-integration/k8s-helm/README.md @@ -139,7 +139,7 @@ helm upgrade --install otel-coralogix-integration coralogix-charts-virtual/otel- If you're leveraging the Prometheus Operator custom resources (`ServiceMonitor` and `PodMonitor`) and you would like to keep using them with the OpenTelemetry collector, you can enable the scraping of these resources by a special, optional component called target allocator. This feature is disabled by default and can be enabled by setting the `opentelemetry-agent.targetAllocator.enabled` value to `true` in the `values.yaml` file. -If enabled, the target allocator will be deployed as a separate deployment in the same namespace as the collector. It will be responsible for allocating targets for the agent collector on each node, to scrape targets that reside on the given node (a form of simple sharding). +If enabled, the target allocator will be deployed as a separate deployment in the same namespace as the collector. It will be responsible for allocating targets for the agent collector on each node, to scrape targets that reside on the given node (a form of simple sharding). If needed, you can run multiple instances of the target allocator for high availability. This can be achieved by setting the `opentelemetry-agent.targetAllocator.replicas` value to a number greater than 1. For more details on Prometheus custom resources and target allocator see the documentation [here](https://github.com/open-telemetry/opentelemetry-operator/tree/main/cmd/otel-allocator#discovery-of-prometheus-custom-resources). diff --git a/otel-integration/k8s-helm/values.yaml b/otel-integration/k8s-helm/values.yaml index 07822ad2..680bfee3 100644 --- a/otel-integration/k8s-helm/values.yaml +++ b/otel-integration/k8s-helm/values.yaml @@ -48,6 +48,7 @@ opentelemetry-agent: targetAllocator: enabled: false + replicas: 1 allocationStrategy: "per-node" prometheusCR: enabled: true