Understanding TargetRef policies
What do targetRef
policies look like?
There are two parts in a policy:
- The metadata
- The spec
Metadata identifies the policies by its name
, type
and what mesh
it is part of.
This is how it looks:
A policy metadata looks like:
In Kubernetes all our policies are implemented as in the group kuma.io/v1alpha1
.
A policy metadata looks like:
apiVersion: kuma.io/v1alpha1
kind: ExamplePolicy
metadata:
name: my-policy-name
namespace: kuma-system
spec:
... # spec data specific to the policy kind
By default the policy is created in the default
mesh. You can specify the mesh by using the kuma.io/mesh
label.
For example:
apiVersion: kuma.io/v1alpha1
kind: ExamplePolicy
metadata:
name: my-policy-name
namespace: kuma-system
labels:
kuma.io/mesh: "my-mesh"
spec:
... # spec data specific to the policy kind
Policies are namespaced scope and currently the namespace must be the one the control-plane is running in kuma-system
by default.
The spec contains the actual configuration of the policy.
All specs have a top level targetRef
which identifies which proxies this policy applies to. In particular, it defines which proxies have their Envoy configuration modified.
Some policies also support further narrowing.
The actual configuration is defined in a default
map.
For example:
type: ExamplePolicy
name: my-example
mesh: default
spec:
targetRef:
kind: Mesh
to:
- targetRef:
kind: Mesh
default: # Configuration that applies to outgoing traffic
key: value
from:
- targetRef:
default: # Configuration that applies to incoming traffic
key: value
Some policies are not directional and will not have to
and from
. For example
One of the benefits of targetRef
policies is that the spec is always the same between Kubernetes and Universal.
This means that converting policies between Universal and Kubernetes only means rewriting the metadata.
Writing a targetRef
targetRef
is a concept borrowed from Kubernetes Gateway API its usage is fully defined in . Its goal is to select subsets of proxies with maximum flexibility.
It looks like:
targetRef:
kind: Mesh | MeshSubset | MeshService | MeshServiceSubset | MeshGatewayRoute
name: "my-name" # For kinds MeshService, MeshServiceSubset and MeshGatewayRoute a name can be defined
tags:
key: value # For kinds MeshServiceSubset and MeshSubset a list of matching tags can be used
Here’s an explanation of each kinds and their scope:
- Mesh: applies to all proxies running in the mesh
- MeshSubset: same as Mesh but filters only proxies who have matching
targetRef.tags
- MeshService: all proxies with a tag
kuma.io/service
equal totargetRef.name
- MeshServiceSubset: same as but further refine to proxies that have matching
targetRef.tags
- MeshGatewayRoute: gateway using
MeshGatewayRoute
that have a name equal totargetRef.name
Consider the example below:
apiVersion: kuma.io/v1alpha1
kind: MeshAccessLog
metadata:
name: example
namespace: kuma-system
labels:
kuma.io/mesh: default
spec:
targetRef: # top level targetRef
kind: MeshService
name: web-frontend
to:
- targetRef: # to level targetRef
kind: MeshService
name: web-backend
default:
backends:
- file:
format:
plain: '{"start_time": "%START_TIME%"}'
path: "/tmp/logs.txt"
from:
- targetRef: # from level targetRef
kind: Mesh
default:
backends:
- file:
format:
plain: '{"start_time": "%START_TIME%"}'
path: "/tmp/logs.txt"
Using spec.targetRef
, this policy targets all proxies that implement the service web-frontend
. It defines the scope of this policy as applying to traffic either from or to web-frontend
services.
The spec.to.targetRef
section enables logging for any traffic going to web-backend
. The spec.from.targetRef
section enables logging for any traffic coming from any service in the Mesh
.
Not every policy supports to
and from
levels. Additionally, not every resource can appear at every supported level. The specified top level resource can also affect which resources can appear in to
or from
.
This table looks like:
Here it indicates that the top level can use any targetRef kinds. But in targetRef.to
only kind Mesh
can be used and in targetRef.from
only kind MeshService
.
It is necessary to define a policy for merging configuration, because a proxy can be targeted by multiple targetRef
’s.
We define a total order of policies:
- Mesh > MeshSubset > MeshService > MeshServiceSubset > MeshGatewayRoute (the more a
targetRef
is focused the higher priority it has) - If levels are equal the lexicographic order of policy names is used
For to
and from
policies we concatenate the array for each matching policies. We then build configuration by merging each level using JSON patch merge.
For example if I have 2 default
ordered this way:
sub:
array: [1, 2, 3]
other: 50
other-array: [3, 4, 5]
---
default:
sub:
array: []
other: null
other-array: [5, 6]
extra: 2
The merge result is:
Applying a global default
type: ExamplePolicy
name: example
mesh: default
spec:
targetRef:
kind: Mesh
to:
- targetRef:
kind: Mesh
default:
key: value
All traffic from any proxy (top level targetRef
) going to any proxy (to targetRef
) will have this policy applied with value key=value
.
Recommending to users
type: ExamplePolicy
name: example
mesh: default
spec:
targetRef:
kind: Mesh
to:
- targetRef:
kind: MeshService
name: my-service
default:
key: value
All traffic from any proxy (top level targetRef
) going to the service “my-service” (to targetRef
) will have this policy applied with value key=value
.
This is useful when a service owner wants to suggest its clients as set of configuration.
Configuring all proxies of a team
type: ExamplePolicy
name: example
mesh: default
spec:
targetRef:
kind: MeshSubset
tags:
team: "my-team"
from:
- targetRef:
kind: Mesh
key: value
All traffic from any proxies (from targetRef
) going to any proxy that has the tag team=my-team
(top level targetRef
) will have this policy applied with value key=value
.
This is a useful way to define coarse grain rules for example.
Configuring all proxies in a zone
This can be very useful when observability stores are different for each zone for example.