Fortsa


Keep Istio's data-plane up-to-date automatically.
Name
The name comes from the Greek phrase "Όρτσα τα πανιά!" meaning roughly "Full Sail Ahead"
Description
When updating Istio, one particular task is pretty much impossible to do with a GitOps-style
workflow: Restarting all the pods running the previous versions of Istio's sidecar proxy. This
is an onerous task even when doing everything via scripts or using something like Ansible. It
can take hours - or even longer - and it's easy to make mistakes. An that's just for a single
medium-sized cluster. If you have many clusters, keeping Istio's data-plane up-to-date can be
a full-time job!
This project is a Kubernetes Operator that automates away the whole task, as much as possible.
In the spirit of being as simple as possible, there are currently no CRDs and no dependencies
other than running in a cluster that uses Istio. Note that we've only tested this when using
Istio's default envoy-based proxy.
Inspiration
This project was inspired by my experience with the toil of keeping the istio data-plane
up-to-date and then discovering the solution Google had come up with, described in this
YouTube video: https://youtu.be/R86ZsYH7Ka4
Architecture
Fortsa is a relatively simple Kubernetes Operator with limited ability to interact with
the cluster. Primarily, it does the following:
- Watches the configuration of Istio’s MutationWebhookConfiguration objects as well as the
configuration of certain Istio-related namespace labels, annotations, and ConfigMaps.
- Compares the Istio configuration with the configuration of pods running in Istio-enabled
namespaces
- Updates the objects controlling the pods to cause them to gracefully restart.
flowchart TB
subgraph watches [Fortsa Watches]
ConfigMaps["ConfigMaps
istio-sidecar-injector*"]
MWCs["MutatingWebhookConfigurations
istio-revision-tag-*"]
Namespaces["Namespaces
istio.io/rev, istio-injection"]
Periodic["Periodic Reconcile"]
end
subgraph fortsa [Fortsa Operator]
Reconciler[IstioChangeReconciler]
Scanner[PodScanner]
Annotator[WorkloadAnnotator]
end
subgraph k8s [Kubernetes Cluster]
Pods["Pods with Istio sidecar"]
Workloads["Deployments / DaemonSets / StatefulSets"]
end
subgraph istio [Istio]
Webhook["Istio MutatingWebhook
injects sidecar"]
end
ConfigMaps --> Reconciler
MWCs --> Reconciler
Namespaces --> Reconciler
Periodic --> Reconciler
Reconciler -->|"parse expected image"| Scanner
Scanner -->|"compare pod vs config"| Pods
Scanner -->|"outdated workloads"| Annotator
Annotator -->|"patch with restartedAt"| Workloads
Workloads -->|"rolling restart"| Pods
Pods -->|"new pod creation"| Webhook
Webhook -->|"inject updated sidecar"| Pods
When a workload pod is restarted in this way, it automatically gets re-configured with an updated
proxy sidecar container. This works because Istio-enabled pods are mutated by Istio’s own
MutatingWebhooks in order to inject a sidecar proxy container whose version matches that of the
control-plane configuration. These webhooks are not a part of Fortsa, they are part of Istio.
When deployed via its Helm chart, Fortsa is configured with RBAC permissions to only perform the
actions necessary for its operation, on only the objects it needs to act upon.
The only changes Fortsa makes within a running Kubernetes cluster are to update pods’ controllers
(typically Deployments, DaemonSets, and StatefulSets) with a well-known annotation. Adding or
updating this annotation is what causes the controller to initiate a controlled restart of the
pods. This is in fact the same exact mechanism used by the command-line tool kubectl when issuing
a rollout restart.
Fortsa currently has no CRDs and there are no plans to introduce any.
Fortsa is written in Go, and compiles to a single binary that is deployed via a bare container with
only the binary (FROM scratch, in Docker parlance)
Fortsa does not require access to anything outside the cluster, nor access to any applications other
than the k8s API and istiod. It does not listen on any ports except for liveness/readiness checks
and metrics.
Getting Started
Prerequisites
- go version v1.26.0+
- docker version 17.03+.
- kubectl version v1.11.3+.
- Access to a Kubernetes v1.11.3+ cluster running Istio v1.21.0+
To Deploy on the cluster
This project is packaged for deployment using Helm or OLM. See the "Packages" for the
various docker images available. The helm repo is here:
https://istio-ecosystem.github.io/fortsa/
Installation should be like any other app packeged for Helm or OLM depending on the
method you want to use.
Build and push your image to the location specified by IMG:
make docker-build docker-push IMG=<some-registry>/fortsa:tag
NOTE: This image ought to be published in the personal registry you specified.
And it is required to have access to pull the image from the working environment.
Make sure you have the proper permission to the registry if the above commands don’t work.
Install the CRDs into the cluster:
make install
Deploy the Manager to the cluster with the image specified by IMG:
make deploy IMG=<some-registry>/fortsa:tag
NOTE: If you encounter RBAC errors, you may need to grant yourself cluster-admin
privileges or be logged in as admin.
Create instances of your solution
You can apply the samples (examples) from the config/sample:
kubectl apply -k config/samples/
NOTE: Ensure that the samples has default values to test it out.
To Uninstall
Delete the instances (CRs) from the cluster:
kubectl delete -k config/samples/
Delete the APIs(CRDs) from the cluster:
make uninstall
UnDeploy the controller from the cluster:
make undeploy
Project Distribution
When a release is published, the project uses a Github Action to build several artifacts,
including Docker images of the Operator application, OLM bundle, OLM catalog, and the Helm
chart. The helm chart tarball is also attached to the release alongside the source code
archives.
Because this project was created using the operator-sdk, which itself is a wrapper around
kubebuilder, please look at the documentation for those frameworks for more info about
build, test, distribution, and deployment details.
Following the options to release and provide this solution to the users.
By providing a bundle with all YAML files
- Build the installer for the image built and published in the registry:
make build-installer IMG=<some-registry>/fortsa:tag
NOTE: The makefile target mentioned above generates an 'install.yaml'
file in the dist directory. This file contains all the resources built
with Kustomize, which are necessary to install this project without its
dependencies.
- Using the installer
Users can just run 'kubectl apply -f ' to install
the project, i.e.:
kubectl apply -f https://raw.githubusercontent.com/<org>/fortsa/<tag or branch>/dist/install.yaml
By providing a Helm Chart
- Build the chart using the optional helm plugin
operator-sdk edit --plugins=helm/v2-alpha
- See that a chart was generated under 'dist/chart', and users
can obtain this solution from there.
NOTE: If you change the project, you need to update the Helm Chart
using the same command above to sync the latest changes. Furthermore,
if you create webhooks, you need to use the above command with
the '--force' flag and manually ensure that any custom configuration
previously added to 'dist/chart/values.yaml' or 'dist/chart/manager/manager.yaml'
is manually re-applied afterwards.
TODO
There's lots to do! If you want to help, see the Contributing section, below.
Origin
Fortsa originated within Cloudera in late 2024 as an SRE's
hackathon project. Realizing the potential benefit to the Istio Community, it was
subsequently open-sourced.
Contributing
This is a young project and we can use all the help we can get! Feature requests, bug
reports, patches, documentation, testing, CI/CD, helm charts, you name it!
NOTE: Run make help for more information on all potential make targets
More information can be found via the Kubebuilder Documentation
License
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.