Skip to content
This repository has been archived by the owner on Jul 3, 2024. It is now read-only.

Commit

Permalink
fixes #85
Browse files Browse the repository at this point in the history
  • Loading branch information
Jochen Kappel committed May 4, 2020
1 parent 82f0e28 commit 455541a
Show file tree
Hide file tree
Showing 2 changed files with 146 additions and 0 deletions.
5 changes: 5 additions & 0 deletions helm/osslm-ansible-resource-manager/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,6 +40,11 @@ NOTE: You can skip the need for persistent volumes by settings cassandra.persist

Install Helm Chart

To install standalone on a k8s cluster ( 1 ansible rm instance, 1 cassandra instance, remote kafka ) run:
```
helm install osslm-ansible-resource-manager-1.3.8.tgz --name osslm-ansible-rm --namespace default --values osslm-ansible-rm.standalone_k8s_values.yaml --debug
```
To install into a ALM node ( 1 ansible rm , shared ALM cassandra and kafka) run:
```
helm install osslm-ansible-resource-manager-1.3.8.tgz --name osslm-ansible-rm --namespace default --values osslm-ansible-rm.values.yaml --debug
```
Expand Down
141 changes: 141 additions & 0 deletions helm/osslm-ansible-resource-manager/standalone_k8s_values.yaml
Original file line number Diff line number Diff line change
@@ -0,0 +1,141 @@
## Docker Image for the osslm-ansible-rm application
docker:
## Make this the full path, including registry host and port if using one
image: accanto/osslm-ansible-rm
version: 1.3.9
imagePullPolicy: IfNotPresent

## Configuration for the application deployment
app:
## Number of pods to deploy
replicas: 1
## osslm-ansible-rm specific configuration
config:
kafka:
## Assumes the responseKafkaConnectionUrl value provided as part of the clientConfigFile is still kafka:9092 - which means we must have a "kafka" host available
## There 2 options for this

## Option 1 (enabled by default)
## Access an existing Kafka installed by the Lifecycle Manager in the same Kubernetes namespace
## This creates an additional service that forwards traffic to the existing Kafka pods
## Expects a Kafka installed using helm-foundation provided with the Lifecycle Manager
useExisting: true
existingSelectors:
app: kafka
release: foundation

## Option 2
## Fixed IP. This is added as a host entry to the container.
hostEnabled: false
host:

## Option 3
## Install kafka as part of this helm chart (see kafka.enabled) then disable useExisting and hostEnabled

cassandra:
## Option 1 (enabled by default)
## Access an existing Cassandra installed by the Lifecycle Manager in the same Kubernetes namespace
## This creates an additional service that forwards traffic to the existing Cassandra pods
## Expects a Cassandra installed using helm-foundation provided with the Lifecycle Manager
useExisting: false
existingSelectors:
app: cassandra
release: foundation

## Alternative IP for connecting to cassandra
hostEnabled: false
host:
## Configure logging
log:
level: DEBUG
dir: /var/alm_ansible_rm/logs
## Sets the OS_CLIENT_CONFIG_FILE environment variable
clientConfigFile: /var/alm_ansible_rm/clouds/clouds.yaml
## Pass additional environment variables to the application containers
env:
# cassandra_ttl:
cassandra_uri: alm-ansible-rm-db
# to support indexing of logs in Elasticsearch using Filebeat, we use logstash format. This allows
# us to bundle the log message and other metadata in a json log message and let Fielbeat extract them
# as top level fields in the Elasticsearch index.
LOG_TYPE: logstash
ssl_enabled: "false"
driver_responseKafkaConnectionUrl: "foundation-kafka:9092"
kafka_replicationFactor: "1"
KAFKA_URL: "add kafak IP address and port"
security:
ssl:
enabled: false
secretName: osslm-ansible-rm-tls

persistence:
size: "1Gi"
## If enabled, will create the PVs needed. Only creates 4 volumes, so there must only be 1 replica of Cassandra, Kafka and Zookeeper
enabled: true
storageClass: "manual"
hostPath: /tmp/osslm

service:
# Using NodePort allows access to the IPs through http://k8s-host:nodePort/
type: NodePort
nodePort: 31080
sslNodePort: 31081

cassandra:
enabled: true
## Required as the ansible-rm application always connects to Cassandra on this hostname
fullnameOverride: alm-ansible-rm-db
## Requires setting up the LocalStorage class and creating a Persistent Volume
persistence:
enabled: true
storageClass: "manual"
size: "1Gi"
config:
cluster_name: alm-ansible-rm-cluster
cluster_size: 1
seed_size: 1
max_heap_size: 1024M
heap_new_size: 512M
resources:
requests:
cpu: 500m
memory: "1Gi"
limits:
cpu: 2
memory: "4Gi"

kafka:
enabled: false
## Required as the ansible-rm application connects to Kafka on this hostname (from the default config file: OS_CLIENT_CONFIG_FILE=/var/alm_ansible_rm/clouds/clouds.yaml)
fullnameOverride: kafka
replicas: 1
storageClass: "manual"
## Requires setting up the LocalStorage class and creating a Persistent Volume
persistence:
enabled: true
size: "1Gi"
storageClass: "manual"

## The following allows Kafka to be accessed outside of K8s at kafka.osslm:31090, if you add kafka.osslm to your hosts file with the k8s IP
external:
enabled: true
domain: osslm
rbac:
enabled: true
configurationOverrides:
offsets.topic.replication.factor: 1
advertised.listeners: |-
EXTERNAL://kafka.osslm:$((31090 + ${KAFKA_BROKER_ID}))
listener.security.protocol.map: |-
PLAINTEXT:PLAINTEXT,EXTERNAL:PLAINTEXT
topics:
- name: ansible-rm
partitions: 1
replicationFactor: 1
zookeeper:
replicaCount: 1
persistence:
enabled: true
storageClass: "manual"
size: "1Gi"

0 comments on commit 455541a

Please sign in to comment.