Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.

...

  • We need EMCO to work with a logging stack such as ELKElastic Search, Fluentd and Kibana. (e.g. see this EFK article).
  • We need a good way to persist the logs.

These don't need any code changes. We can do a PoC for a deployment of EMCO with log persistence and a log stack, and document the ingredients and the recipe. Perhaps, the needed YAML files can be checked in as well.

We also want to add user id to logs - this requires configuring Istio Ingress to pass the user id, and making EMCO code changes to include user id in the logs.

We also need to investigate an events framework. This is TBD.

...

Database persistence

Today, db persistence is not enabled by default. We defaults to local storage. We need cluster-wide storage for different use cases. So, we need to validate with persistence enabledcluster-wide persistence and provide a reference solution ('experience kit') for easily deploying mongo/etcd/etc. with cluster-wide storage. The experience kit will ideally have documentation in EMCO repo and perhaps a set of YAML files that can be applied to configure NFS with appropriate PVs.

  • We have tested with NFS-based PV in the past, and we still have the NFS-related YAMLs. If So there is consensus on around using NFS as the default storage, we need to set up NFS environmentcluster-wide storage..
  • With persistence enabled, we cannot rely anymore upon developer-oriented troubleshooting and workarounds based on re-installing EMCO to blow away the db. Developers should also
    test with persistence enabled.

...

Without the GitOps model, rsync should apply configurable retry/timeout policies to handle cluster connectivity loss. We have the
/projects/.../{dig}/stop API but that is a workaround -- the user needs to invoke that API manually. We need to validate rsync has retries/timeout for cluster connectivity., and we considered validating it.

However, the non-GitOps approach has many issues. So, it is preferable to make the GitOps approach fully functional (ideally in 22.09), and deprecate other apparoaches later.

So, there is no need to validate rsync retries/timeout for cluster connectivity. Question: can we recommend the GitOps approach and leave things as is? If not, we need to fix this.

Storage Considerations

We need storage in the cluster where EMCO runs for:

  • mongo db
  • etcd
  • logs
  • metrics? (assuming we aggregate cluster metrics in the central cluster via federated Prometheus, Thanos, Cortex, ....)
  • ?

TBD: Enable db replication and/or sharding. This requires further thought.

Upgrades

It should be possible to upgrade in-place from one released version to the next released version.  The primary concern is any database schema changes between versions.

A smaller concern is a deprecation schedule for removing APIs and/or components.

  • It would be ideal to make schema changes without breaking backwards compatibility - add new fields, don't rename/delete fields, etc.
  • For schema migration, the standard way is to provide a (set of) script(s) that are run to update the db.

This needs to be taken up for consideration.

22.09 Release Tasks for Productizability

  • Tracing for core components (orchestrator, rsync, clm and dcm)
  • Metrics for core components (orchestrator, rsync, clm and dcm) - HTTP request rate etc., error counters, etc. based on this discussion.
  • Experience kit for Logging (EFK stack): documentation with YAMLs
  • Experience kit for cluster-wide persistence with NFS (for mongo, etcd, logs and m3db): documentation + YAMLs
  • Validation that each microservice can restart with integrity after a crash
  • Investigation of which microservices make multiple db writes for a single API call

Also, start planning for:

  • Considerations for acquiring db lock for multiple db writes in a single API call
  • Upgrades