openimsdk / helm-charts

helm charts repository for openim
https://openimsdk.github.io/helm-charts/
Apache License 2.0
14 stars 10 forks source link

build(deps): Bump tj-actions/changed-files from 41.0.1 to 42.0.2 #78

Closed dependabot[bot] closed 5 months ago

dependabot[bot] commented 5 months ago

Bumps tj-actions/changed-files from 41.0.1 to 42.0.2.

Release notes

Sourced from tj-actions/changed-files's releases.

v42.0.2

What's Changed

Full Changelog: https://github.com/tj-actions/changed-files/compare/v42...v42.0.2

v42.0.1

What's Changed

Full Changelog: https://github.com/tj-actions/changed-files/compare/v42...v42.0.1

v42.0.0

🔥🔥 BREAKING CHANGE 🔥🔥

  • Input file patterns that end with a / would now match all sub-files within the directory without requiring you to specify the globstar pattern.

...
      - name: Get changed files
        id: changed-files
        uses: tj-actions/changed-files@v42
        with:
          files: 'dir/'  # Would also be the same as dir/** 

What's Changed

Full Changelog: https://github.com/tj-actions/changed-files/compare/v41...v42.0.0

v42

Changes in v42.0.2

... (truncated)

Changelog

Sourced from tj-actions/changed-files's changelog.

Changelog

42.0.2 - (2024-01-25)

🚀 Features

  • Enhance error handling for non-git directories (#1885) (90a06d6) - (Tonye Jack)

⬆️ Upgrades

  • Upgraded to v42.0.1 (#1884)

Co-authored-by: jackton1 jackton1@users.noreply.github.com (2cb2c92) - (tj-actions[bot])

42.0.1 - (2024-01-24)

➕ Add

  • Added missing changes and modified dist assets. (ea024b2) - (GitHub Action)
  • Added missing changes and modified dist assets. (3af07c2) - (GitHub Action)

🔄 Update

  • Update env.ts (3680129) - (Tonye Jack)

⚙️ Miscellaneous Tasks

  • deps: Update dependency @​types/node to v20.11.6 (ac21d93) - (renovate[bot])
  • deps: Update typescript-eslint monorepo to v6.19.1 (a4637ea) - (renovate[bot])
  • deps: Update dependency ts-jest to v29.1.2 (fd9998c) - (renovate[bot])
  • deps: Lock file maintenance (db4e584) - (renovate[bot])
  • Rename example worflows from test to example (#1878) (c6543c4) - (Tonye Jack)
  • Update README.md (#1877) (88f9f3e) - (Tonye Jack)
  • deps: Lock file maintenance (5d866cb) - (renovate[bot])
  • deps: Update tj-actions/eslint-changed-files action to v23 (#1875) (346f237) - (renovate[bot])

⬆️ Upgrades

  • Upgraded to v42 (#1874)

Co-authored-by: jackton1 jackton1@users.noreply.github.com (c037f1e) - (tj-actions[bot])

42.0.0 - (2024-01-18)

🚀 Features

  • Add support for include/exclude all nested files when a directory is specified and ends with a slash (#1873) (ae82ed4) - (Tonye Jack)

... (truncated)

Commits
  • 90a06d6 feat: enhance error handling for non-git directories (#1885)
  • 2cb2c92 Upgraded to v42.0.1 (#1884)
  • ac21d93 chore(deps): update dependency @​types/node to v20.11.6
  • a4637ea chore(deps): update typescript-eslint monorepo to v6.19.1
  • fd9998c chore(deps): update dependency ts-jest to v29.1.2
  • ea024b2 Added missing changes and modified dist assets.
  • db4e584 chore(deps): lock file maintenance
  • c6543c4 chore: rename example worflows from test to example (#1878)
  • 88f9f3e chore: update README.md (#1877)
  • 3af07c2 Added missing changes and modified dist assets.
  • Additional commits viewable in compare view


Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR: - `@dependabot rebase` will rebase this PR - `@dependabot recreate` will recreate this PR, overwriting any edits that have been made to it - `@dependabot merge` will merge this PR after your CI passes on it - `@dependabot squash and merge` will squash and merge this PR after your CI passes on it - `@dependabot cancel merge` will cancel a previously requested merge and block automerging - `@dependabot reopen` will reopen this PR if it is closed - `@dependabot close` will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually - `@dependabot show ignore conditions` will show all of the ignore conditions of the specified dependency - `@dependabot ignore this major version` will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this minor version` will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself) - `@dependabot ignore this dependency` will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)
kubbot commented 5 months ago

Kubernetes Templates in openim Namespace

openim templates get ./charts/openim-server -f k8s-open-im-server-config.yaml -f config-imserver.yaml ```markdown --- # Source: openim-api/templates/app-cm.yaml apiVersion: v1 kind: ConfigMap metadata: name: openim-cm data: config.yaml: |+ api: listenIP: 0.0.0.0 openImApiPort: - 80 callback: afterSendGroupMsg: enable: false timeout: 5 afterSendSingleMsg: enable: false timeout: 5 beforeAddFriend: enable: false failedContinue: true timeout: 5 beforeCreateGroup: enable: false failedContinue: true timeout: 5 beforeMemberJoinGroup: enable: false failedContinue: true timeout: 5 beforeSendGroupMsg: enable: false failedContinue: true timeout: 5 beforeSendSingleMsg: enable: false failedContinue: true timeout: 5 beforeSetGroupMemberInfo: enable: false failedContinue: true timeout: 5 msgModify: enable: false failedContinue: true timeout: 5 offlinePush: enable: false failedContinue: true timeout: 5 onlinePush: enable: false failedContinue: true timeout: 5 setMessageReactionExtensions: enable: false failedContinue: true timeout: 5 superGroupOnlinePush: enable: false failedContinue: true timeout: 5 url: null userKickOff: enable: false timeout: 5 userOffline: enable: false timeout: 5 userOnline: enable: false timeout: 5 chatPersistenceMysql: true chatRecordsClearTime: 0 2 * * 3 envs: discovery: k8s groupMessageHasReadReceiptEnable: true iosPush: badgeCount: true production: false pushSound: xxx kafka: addr: - im-kafka:9092 consumerGroupID: msgToMongo: mongo msgToMySql: mysql msgToPush: push msgToRedis: redis latestMsgToRedis: topic: latestMsgToRedis msgToPush: topic: msgToPush offlineMsgToMongo: topic: offlineMsgToMongoMysql password: proot username: root log: isJson: false isStdout: true remainLogLevel: 6 remainRotationCount: 2 rotationTime: 24 storageLocation: ../logs/ withStack: false longConnSvr: openImMessageGatewayPort: - 88 openImWsPort: - 80 websocketMaxConnNum: 100000 websocketMaxMsgLen: 4096 websocketTimeout: 10 manager: nickname: - system1 - system2 - system3 userID: - openIM123456 - openIM654321 - openIMAdmin messageVerify: friendVerify: false mongo: address: - im-mongodb:27017 database: openIM_v3 maxPoolSize: 100 password: openIM123 uri: "" username: root msgCacheTimeout: 86400 msgDestructTime: 0 2 * * * multiLoginPolicy: 1 mysql: address: - im-mysql:3306 database: openIM_v3 logLevel: 4 maxIdleConn: 100 maxLifeTime: 60 maxOpenConn: 1000 password: openIM123 slowThreshold: 500 username: root object: apiURL: https://openim1.server.top/api cos: bucketURL: https://temp-1252357374.cos.ap-chengdu.myqcloud.com secretID: "" secretKey: "" sessionToken: "" enable: minio minio: accessKeyID: root bucket: openim endpoint: http://im-minio:9000 secretAccessKey: openIM123 sessionToken: "" signEndpoint: https://openim1.server.top/im-minio-api oss: accessKeyID: "" accessKeySecret: "" bucket: demo-9999999 bucketURL: https://demo-9999999.oss-cn-chengdu.aliyuncs.com endpoint: https://oss-cn-chengdu.aliyuncs.com sessionToken: "" prometheus: apiPrometheusPort: - 90 authPrometheusPort: - 90 conversationPrometheusPort: - 90 enable: false friendPrometheusPort: - 90 grafanaUrl: https://openim2.server.top/ groupPrometheusPort: - 90 messageGatewayPrometheusPort: - 90 messagePrometheusPort: - 90 messageTransferPrometheusPort: - 90 - 90 - 90 - 90 pushPrometheusPort: - 90 rtcPrometheusPort: - 90 thirdPrometheusPort: - 90 userPrometheusPort: - 90 push: enable: getui fcm: serviceAccount: x.json geTui: appKey: "" channelID: "" channelName: "" intent: "" masterSecret: "" pushUrl: https://restapi.getui.com/v2/$appId jpns: appKey: null masterSecret: null pushIntent: null pushUrl: null redis: address: - im-redis-master:6379 password: openIM123 username: "" retainChatRecords: 365 rpc: listenIP: 0.0.0.0 registerIP: "" rpcPort: openImAuthPort: - 80 openImConversationPort: - 80 openImFriendPort: - 80 openImGroupPort: - 80 openImMessageGatewayPort: - 88 openImMessagePort: - 80 openImPushPort: - 80 openImThirdPort: - 80 openImUserPort: - 80 rpcRegisterName: openImAuthName: openimserver-openim-rpc-auth:80 openImConversationName: openimserver-openim-rpc-conversation:80 openImFriendName: openimserver-openim-rpc-friend:80 openImGroupName: openimserver-openim-rpc-group:80 openImMessageGatewayName: openimserver-openim-msggateway:88 openImMsgName: openimserver-openim-rpc-msg:80 openImPushName: openimserver-openim-push:80 openImThirdName: openimserver-openim-rpc-third:80 openImUserName: openimserver-openim-rpc-user:80 secret: openIM123 singleMessageHasReadReceiptEnable: true tokenPolicy: expire: 90 zookeeper: address: - 172.28.0.1:12181 password: "" schema: openim username: "" notification.yaml: |+ --- # Source: openim-api/charts/openim-msggateway-proxy/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-msggateway-proxy labels: helm.sh/chart: openim-msggateway-proxy-0.1.0 app.kubernetes.io/name: openim-msggateway-proxy app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 88 targetPort: rpc protocol: TCP name: rpc - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-msggateway-proxy app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-msggateway/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-msggateway labels: helm.sh/chart: openim-msggateway-0.1.0 app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 88 targetPort: rpc protocol: TCP name: rpc - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-msggateway/templates/serviceheadless.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-msggateway-headless labels: helm.sh/chart: openim-msggateway-0.1.0 app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: ports: - port: 80 targetPort: http protocol: TCP name: http - port: 88 targetPort: rpc protocol: TCP name: rpc - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name clusterIP: None --- # Source: openim-api/charts/openim-msgtransfer/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-msgtransfer labels: helm.sh/chart: openim-msgtransfer-0.1.0 app.kubernetes.io/name: openim-msgtransfer app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-msgtransfer app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-push/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-push labels: helm.sh/chart: openim-push-0.1.0 app.kubernetes.io/name: openim-push app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-push app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-auth/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-auth labels: helm.sh/chart: openim-rpc-auth-0.1.0 app.kubernetes.io/name: openim-rpc-auth app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-auth app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-conversation/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-conversation labels: helm.sh/chart: openim-rpc-conversation-0.1.0 app.kubernetes.io/name: openim-rpc-conversation app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-conversation app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-friend/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-friend labels: helm.sh/chart: openim-rpc-friend-0.1.0 app.kubernetes.io/name: openim-rpc-friend app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-friend app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-group/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-group labels: helm.sh/chart: openim-rpc-group-0.1.0 app.kubernetes.io/name: openim-rpc-group app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-group app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-msg/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-msg labels: helm.sh/chart: openim-rpc-msg-0.1.0 app.kubernetes.io/name: openim-rpc-msg app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-msg app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-third/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-third labels: helm.sh/chart: openim-rpc-third-0.1.0 app.kubernetes.io/name: openim-rpc-third app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-third app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-rpc-user/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-rpc-user labels: helm.sh/chart: openim-rpc-user-0.1.0 app.kubernetes.io/name: openim-rpc-user app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-rpc-user app.kubernetes.io/instance: release-name --- # Source: openim-api/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-openim-api labels: helm.sh/chart: openim-api-0.1.16 app.kubernetes.io/name: openim-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http - port: 90 targetPort: 90 protocol: TCP name: metrics-port selector: app.kubernetes.io/name: openim-api app.kubernetes.io/instance: release-name --- # Source: openim-api/charts/openim-msggateway-proxy/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-msggateway-proxy labels: helm.sh/chart: openim-msggateway-proxy-0.1.0 app.kubernetes.io/name: openim-msggateway-proxy app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-msggateway-proxy app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-msggateway-proxy app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-msggateway-proxy securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-msggateway-proxy:v3.5.0" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP - name: rpc containerPort: 88 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-msgtransfer/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-msgtransfer labels: helm.sh/chart: openim-msgtransfer-0.1.0 app.kubernetes.io/name: openim-msgtransfer app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-msgtransfer app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-msgtransfer app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-msgtransfer securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-msgtransfer:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-push/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-push labels: helm.sh/chart: openim-push-0.1.0 app.kubernetes.io/name: openim-push app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-push app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-push app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-push securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-push:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-auth/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-auth labels: helm.sh/chart: openim-rpc-auth-0.1.0 app.kubernetes.io/name: openim-rpc-auth app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-auth app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-auth app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-auth securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-auth:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-conversation/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-conversation labels: helm.sh/chart: openim-rpc-conversation-0.1.0 app.kubernetes.io/name: openim-rpc-conversation app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-conversation app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-conversation app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-conversation securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-conversation:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-friend/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-friend labels: helm.sh/chart: openim-rpc-friend-0.1.0 app.kubernetes.io/name: openim-rpc-friend app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-friend app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-friend app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-friend securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-friend:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-group/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-group labels: helm.sh/chart: openim-rpc-group-0.1.0 app.kubernetes.io/name: openim-rpc-group app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-group app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-group app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-group securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-group:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-msg/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-msg labels: helm.sh/chart: openim-rpc-msg-0.1.0 app.kubernetes.io/name: openim-rpc-msg app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-msg app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-msg app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-msg securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-msg:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-third/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-third labels: helm.sh/chart: openim-rpc-third-0.1.0 app.kubernetes.io/name: openim-rpc-third app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-third app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-third app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-third securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-third:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-rpc-user/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-rpc-user labels: helm.sh/chart: openim-rpc-user-0.1.0 app.kubernetes.io/name: openim-rpc-user app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-rpc-user app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-rpc-user app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-rpc-user securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-rpc-user:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-openim-api labels: helm.sh/chart: openim-api-0.1.16 app.kubernetes.io/name: openim-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-api app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-api app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-api securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-api:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-msggateway/templates/deployment.yaml apiVersion: apps/v1 kind: StatefulSet metadata: name: release-name-openim-msggateway labels: helm.sh/chart: openim-msggateway-0.1.0 app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: serviceName: release-name-openim-msggateway-headless replicas: 1 selector: matchLabels: app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: openim-msggateway app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: openim-msggateway securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/openim-msggateway:release-v3.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP - name: rpc containerPort: 88 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http env: - name: MY_MSGGATEWAY_REPLICACOUNT value: "1" - name: MY_POD_NAME valueFrom: fieldRef: fieldPath: metadata.name - name: MY_POD_NAMESPACE valueFrom: fieldRef: fieldPath: metadata.namespace resources: {} volumeMounts: - mountPath: /openim/openim-server/config/config.yaml name: config subPath: config.yaml - mountPath: /openim/openim-server/config/notification.yaml name: config subPath: notification.yaml volumes: - name: config configMap: name: openim-cm --- # Source: openim-api/charts/openim-msggateway-proxy/templates/ingress.yaml apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: release-name-openim-msggateway-proxy labels: helm.sh/chart: openim-msggateway-proxy-0.1.0 app.kubernetes.io/name: openim-msggateway-proxy app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm annotations: nginx.ingress.kubernetes.io/rewrite-target: /$2 nginx.ingress.kubernetes.io/use-regex: "true" spec: ingressClassName: nginx tls: - hosts: - "openim1.server.top" secretName: webapitls rules: - host: "openim1.server.top" http: paths: - path: /msg_gateway(/|$)(.*) pathType: ImplementationSpecific backend: service: name: release-name-openim-msggateway-proxy port: number: 80 --- # Source: openim-api/templates/ingress.yaml apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: release-name-openim-api labels: helm.sh/chart: openim-api-0.1.16 app.kubernetes.io/name: openim-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm annotations: nginx.ingress.kubernetes.io/rewrite-target: /$2 nginx.ingress.kubernetes.io/use-regex: "true" spec: ingressClassName: nginx tls: - hosts: - "openim1.server.top" secretName: webapitls rules: - host: "openim1.server.top" http: paths: - path: /api(/|$)(.*) pathType: ImplementationSpecific backend: service: name: release-name-openim-api port: number: 80 ```
openim templates get ./charts/openim-chat -f k8s-chat-server-config.yaml -f config-chatserver.yaml ```markdown --- # Source: admin-api/templates/app-cm.yaml apiVersion: v1 kind: ConfigMap metadata: name: imchat-cm data: config.yaml: |+ adminApi: listenIP: null openImAdminApiPort: - 80 adminList: - adminID: admin1 imAdmin: openIM123456 nickname: chat1 - adminID: admin2 imAdmin: openIM654321 nickname: chat2 - adminID: admin3 imAdmin: openIMAdmin nickname: chat3 chatApi: listenIP: null openImChatApiPort: - 80 envs: discovery: k8s log: isJson: false isStdout: true remainLogLevel: 6 remainRotationCount: 2 rotationTime: 24 storageLocation: ../logs/ withStack: false mysql: address: - im-mysql:3306 database: openim_enterprise logLevel: 4 maxIdleConn: 100 maxLifeTime: 60 maxOpenConn: 1000 password: openIM123 slowThreshold: 500 username: root openIMUrl: http://openimserver-openim-api redis: address: - im-redis-master:6379 password: openIM123 username: "" rpc: listenIP: null registerIP: null rpcPort: openImAdminPort: - 80 openImChatPort: - 80 rpcRegisterName: openImAdminName: openimchat-admin-rpc:80 openImChatName: openimchat-chat-rpc:80 secret: openIM123 tokenPolicy: expire: 86400 verifyCode: ali: accessKeyId: "" accessKeySecret: "" endpoint: dysmsapi.aliyuncs.com signName: "" verificationCodeTemplateCode: "" len: 6 maxCount: 10 superCode: "666666" uintTime: 86400 use: "" validCount: 5 validTime: 300 zookeeper: password: "" schema: openim username: "" zkAddr: - 127.0.0.1:12181 --- # Source: admin-api/charts/admin-rpc/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-admin-rpc labels: helm.sh/chart: admin-rpc-0.1.0 app.kubernetes.io/name: admin-rpc app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http selector: app.kubernetes.io/name: admin-rpc app.kubernetes.io/instance: release-name --- # Source: admin-api/charts/chat-api/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-chat-api labels: helm.sh/chart: chat-api-0.1.0 app.kubernetes.io/name: chat-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http selector: app.kubernetes.io/name: chat-api app.kubernetes.io/instance: release-name --- # Source: admin-api/charts/chat-rpc/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-chat-rpc labels: helm.sh/chart: chat-rpc-0.1.0 app.kubernetes.io/name: chat-rpc app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http selector: app.kubernetes.io/name: chat-rpc app.kubernetes.io/instance: release-name --- # Source: admin-api/templates/service.yaml apiVersion: v1 kind: Service metadata: name: release-name-admin-api labels: helm.sh/chart: admin-api-0.1.16 app.kubernetes.io/name: admin-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: type: ClusterIP ports: - port: 80 targetPort: http protocol: TCP name: http selector: app.kubernetes.io/name: admin-api app.kubernetes.io/instance: release-name --- # Source: admin-api/charts/admin-rpc/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-admin-rpc labels: helm.sh/chart: admin-rpc-0.1.0 app.kubernetes.io/name: admin-rpc app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: admin-rpc app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: admin-rpc app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: admin-rpc securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/chat-rpc-admin:release-v1.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http resources: {} volumeMounts: - mountPath: /openim/openim-chat/config/config.yaml name: config subPath: config.yaml volumes: - name: config configMap: name: imchat-cm --- # Source: admin-api/charts/chat-api/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-chat-api labels: helm.sh/chart: chat-api-0.1.0 app.kubernetes.io/name: chat-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: chat-api app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: chat-api app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: chat-api securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/chat-api-chat:release-v1.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http resources: {} volumeMounts: - mountPath: /openim/openim-chat/config/config.yaml name: config subPath: config.yaml volumes: - name: config configMap: name: imchat-cm --- # Source: admin-api/charts/chat-rpc/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-chat-rpc labels: helm.sh/chart: chat-rpc-0.1.0 app.kubernetes.io/name: chat-rpc app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: chat-rpc app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: chat-rpc app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: chat-rpc securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/chat-rpc-chat:release-v1.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http resources: {} volumeMounts: - mountPath: /openim/openim-chat/config/config.yaml name: config subPath: config.yaml volumes: - name: config configMap: name: imchat-cm --- # Source: admin-api/templates/deployment.yaml apiVersion: apps/v1 kind: Deployment metadata: name: release-name-admin-api labels: helm.sh/chart: admin-api-0.1.16 app.kubernetes.io/name: admin-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm spec: replicas: 1 selector: matchLabels: app.kubernetes.io/name: admin-api app.kubernetes.io/instance: release-name template: metadata: labels: app.kubernetes.io/name: admin-api app.kubernetes.io/instance: release-name spec: serviceAccountName: default securityContext: {} containers: - name: admin-api securityContext: {} image: "registry.cn-hangzhou.aliyuncs.com/openimsdk/chat-api-admin:release-v1.5" imagePullPolicy: Always ports: - name: http containerPort: 80 protocol: TCP #livenessProbe: # httpGet: # path: / # port: http #readinessProbe: # httpGet: # path: / # port: http resources: {} volumeMounts: - mountPath: /openim/openim-chat/config/config.yaml name: config subPath: config.yaml volumes: - name: config configMap: name: imchat-cm --- # Source: admin-api/charts/chat-api/templates/ingress.yaml apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: release-name-chat-api labels: helm.sh/chart: chat-api-0.1.0 app.kubernetes.io/name: chat-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm annotations: nginx.ingress.kubernetes.io/rewrite-target: /$2 nginx.ingress.kubernetes.io/use-regex: "true" spec: ingressClassName: nginx tls: - hosts: - "openim1.server.top" secretName: webapitls rules: - host: "openim1.server.top" http: paths: - path: /chat(/|$)(.*) pathType: ImplementationSpecific backend: service: name: release-name-chat-api port: number: 80 --- # Source: admin-api/templates/ingress.yaml apiVersion: networking.k8s.io/v1 kind: Ingress metadata: name: release-name-admin-api labels: helm.sh/chart: admin-api-0.1.16 app.kubernetes.io/name: admin-api app.kubernetes.io/instance: release-name app.kubernetes.io/version: "1.16.0" app.kubernetes.io/managed-by: Helm annotations: nginx.ingress.kubernetes.io/rewrite-target: /$2 nginx.ingress.kubernetes.io/use-regex: "true" spec: ingressClassName: nginx tls: - hosts: - "openim1.server.top" secretName: webapitls rules: - host: "openim1.server.top" http: paths: - path: /complete_admin(/|$)(.*) pathType: ImplementationSpecific backend: service: name: release-name-admin-api port: number: 80 ```
dependabot[bot] commented 5 months ago

Superseded by #82.