docs/proposals/004-scalability-benchmarking.md
Users of Argo CD are interested to know how to scale Argo CD, what configuration tweaks and deployment options they have, and how far they can push resources (in terms of the number of supported applications, Git repositories, managing clusters, etc.).
While the Argo CD documentation discusses options to scale up, the actual process is not clear and, as articulated in this thread, oftentimes a point of confusion for users.
By running large-scale benchmarking, we aim at helping the Argo CD community with the following:
argo-cd-benchmarking repo under argoproj-labs so that anyone can easily replicate it and the development of the procedures can happen outside of the lifecycle for Argo CD (unlike the current gen-resources hack in the main project).The initial members for this proposal and their affiliations are:
| Name | Company |
|---|---|
| Andrew Lee | AWS |
| Carlos Santana | AWS |
| Nicholas Morey | Akuity |
| Nima Kaviani | AWS |
With the introduction of the proposed Scalability SIG, the members participating in the proposal may change.
Any community member is welcome to participate in the work for this proposal. Either by joining the Scalability SIG or through contributing to the proposed argoproj-labs/argo-cd-benchmarking repository containing the tooling.
argo-cd-benchmarking repo under argoproj-labs and add the authors of this proposal as maintainers.Each use case will cover a specific topology with N permutations based on the key scalability factors. They will be measured using the metrics given in the proposal.
We intend to focus on two key topologies: one Argo CD per cluster and one Argo CD for all remote clusters. All other variations are an extension of these.
The exact key scalability factors used in each permutation will be determined once we get to the testing.
This will capture the impact of network throughput on performance.
This instance will only manage namespace-level resources to determine the impact of monitoring cluster-scoped resources (related to the effect of resource churn in the cluster).
There is already some tooling in the Argo CD repository for scalability testing. We plan to build on the existing effort and further push the boundaries of testing it.
Once we have the benchmarking tooling, we can determine if the simulated components skew the results compared to the real world.
AWS intends to provide the infrastructure required to benchmark large-scale scenarios.
There is no intention to change the security model of Argo CD and therefore this project has no direct security considerations.