The motivations for the new layout are related to bringing more flexibility to users and part of the process to Integrating Kubebuilder and Operator SDK. Because of this integration you may be referred to the Kubebuilder documentation for more information about certain topics. When using this document just remember to replace $ kubebuilder <command>
with $ operator-sdk <command>
.
Note: It is recommended that you have your project upgraded to the latest SDK v1.y release version before following the steps in this guide to migrate to the new layout. However, the steps might work from previous versions as well. In this case, if you find an issue which is not covered here then check the previous Migration Guides which might help out.
The
deploy
directory was replaced with theconfig
directory including a new layout of Kubernetes manifests files:- CRD manifests in
deploy/crds/
are now inconfig/crd/bases
- CR manifests in
deploy/crds/
are now inconfig/samples
- Controller manifest
deploy/operator.yaml
is now inconfig/manager/manager.yaml
- RBAC manifests in
deploy
are now inconfig/rbac/
- CRD manifests in
What is new
Scaffolded projects now use:
- kustomize to manage Kubernetes resources needed to deploy your operator
- A
Makefile
with helpful targets for build, test, and deployment, and to give you flexibility to tailor things to your project’s needs - Updated metrics configuration using , a
--metrics-bind-address
flag, and kustomize-based deployment of a KubernetesService
and prometheus operatorServiceMonitor
- Preliminary support for CLI plugins. For more info see the [plugins design document][plugins-phase1-design-doc]
- A
PROJECT
configuration file to store information about GVKs, plugins, and help the CLI make decisions.
Generated files with the default API versions:
apiextensions/v1
for generated CRDs (apiextensions/v1beta1
was deprecated in Kubernetes1.16
and will be removed in1.22
)admissionregistration.k8s.io/v1
for webhooks (admissionregistration.k8s.io/v1beta1
was deprecated in Kubernetes1.16
and will be removed in1.22
)
How to migrate
The easy migration path is to initialize a new project, re-recreate APIs, then copy pre-v1.0.0 configuration files into the new project.
Prerequisites
- Go through the .
- User authorized with
cluster-admin
permissions. - An accessible image registry for various operator images (ex. hub.docker.com, ) and be logged in in your command line environment.
example.com
is used as the registry Docker Hub namespace in these examples. Replace it with another value if using a different registry or namespace.- Authentication and certificates if the registry is private or uses a custom CA.
In Kubebuilder-style projects, CRD groups are defined using two different flags (--group
and ).
When we initialize a new project, we need to specify the domain that all APIs in our project will share, so before creating the new project, we need to determine which domain we’re using for the APIs in our existing project.
To determine the domain, look at the spec.group
field in your CRDs in the deploy/crds
directory.
So let’s create a new project with the same domain (example.com
):
Now that we have our new project initialized, we need to re-create each of our APIs. Using our API example from earlier (demo.example.com
), we’ll use demo
for the --group
flag.
For --version
and --kind
, we use spec.versions[0].name
and spec.names.kind
, respectively.
For each API in the existing project, run:
operator-sdk create api \
--group=demo \
--version=<version> \
--kind=<Kind> \
--helm-chart=<path_to_existing_project>/helm-charts/<chart>
Migrating your Custom Resource samples
Update the CR manifests in config/samples
with the values of the CRs in your existing project which are in deploy/crds/<group>_<version>_<kind>_cr.yaml
Migrating watches.yaml
Check if you have custom options in the watches.yaml
file of your existing project. If so, update the new watches.yaml
file to match. In our example, it will look like:
NOTE: Do not remove the +kubebuilder:scaffold:watch
. It allows the tool to update the watches file when new APIs are created.
In your new project, roles are automatically generated in config/rbac/role.yaml
. If you modified these permissions manually in deploy/role.yaml
in your existing project, you need to re-apply them in config/rbac/role.yaml
.
New projects are configured to watch all namespaces by default, so they need a ClusterRole
to have the necessary permissions. Ensure that config/rbac/role.yaml
remains a ClusterRole
if you want to retain the default behavior of the new project conventions.
- monitoring.coreos.com
resources:
- servicemonitors
verbs:
- get
- create
- apiGroups:
- apps
resourceNames:
- nginx-operator
resources:
- deployments/finalizers
verbs:
- update
Updating your ServiceAccount
New Helm projects come with a ServiceAccount controller-manager
in config/rbac/service_account.yaml
. Your project’s RoleBinding and ClusterRoleBinding subjects, and Deployment’s spec.template.spec.serviceAccountName
that reference a ServiceAccount already refer to this new name. When you run make deploy
, your project’s name will be prepended to controller-manager
, making it unique within a namespace, much like your old deploy/service_account.yaml
. If you wish to use the old ServiceAccount, make sure to update all RBAC bindings and your manager Deployment.
Configuring your Operator
If your existing project has customizations in deploy/operator.yaml
, they need to be ported to config/manager/manager.yaml
. If you are passing custom arguments in your deployment, make sure to also update config/default/auth_proxy_patch.yaml
.
Note that the following environment variables are no longer used.
OPERATOR_NAME
is deprecated. It is used to define the name for a leader election config map. Operator authors should begin using--leader-election-id
instead.POD_NAME
was used to enable a particular pod to hold the leader election lock when the Helm operator used the leader for life mechanism. Helm operator now uses controller-runtime’s leader with lease mechanism, andPOD_NAME
is no longer necessary.
Exporting metrics
If you are using metrics and would like to keep them exported you will need to configure it in the config/default/kustomization.yaml
. Please see the doc to know how you can perform this setup.
The default port used by the metric endpoint binds to was changed from :8383
to :8080
. To continue using port 8383
, specify when you start the operator.
The project can now be deployed on cluster by running the command:
You can troubleshoot your deployment by checking container logs:
kubectl logs deployment.apps/nginx-operator-controller-manager -n nginx-operator-system -c manager
For further steps regarding the deployment of the operator, creation of custom resources, and cleaning up of resources, see the tutorial.