How to enable ports pool support¶
To enable the utilization of the ports pool feature, the selected pool driver needs to be included at the kuryr.conf at the kubernetes section. So, for the baremetal deployment:
vif_pool_driver = neutron
And for the nested (VLAN+Trunk) case:
vif_pool_driver = nested
On the other hand, there are a few extra (optional) configuration options regarding the maximum and minimum desired sizes of the pools, where the maximum size can be disabled by setting it to 0:
ports_pool_max = 10
ports_pool_min = 5
In addition the size of the bulk operation, e.g., the number of ports created in a bulk request upon pool population, can be modified:
ports_pool_batch = 5
Note this value should be smaller than the ports_pool_max (if the ports_pool_max is enabled).
Finally, to define the frequency (in seconds) of ports recycle to allow them to be reused by future pods, configure the following option:
ports_pool_update_frequency = 20
After these configurations, the final step is to restart the kuryr-k8s-controller. At devstack deployment:
$ sudo systemctl restart email@example.com
And for RDO packaging based installations:
$ sudo systemctl restart kuryr-controller
Note that for the containerized deployment, you need to edit the associated ConfigMap to change the kuryr.conf files with:
$ kubectl -n kube-system edit cm kuryr-config
Then modify the kuryr.conf to modify the controller configuration regarding the pools. After that, to have the new configuration applied you need to restart the kuryr-controller just by killing the existing pod:
$ kubectl -n kube-system get pod | grep kuryr-controller
$ kubectl -n kube-system delete pod KURYR_CONTROLLER_POD_NAME
Ports loading into pools¶
Pre-created ports for the pools will be loaded and put back into their respective pools upon controller restart. This allows the pre-creation of neutron ports (or subports for the nested case) with a script or any other preferred tool (e.g., heat templates) and load them into their respective pools just by restarting the kuryr-controller (or even before installing it). To do that you just need to ensure the ports are created with the right device_owner:
For neutron pod driver: compute:kuryr (of the value at kuryr.lib.constants.py)
For nested-vlan pod driver: trunk:subport or compute:kuryr (or the value at kuryr.lib.constants.py). But in this case they also need to be attached to an active neutron trunk port, i.e., they need to be subports of an existing trunk
Subports pools management tool¶
Note there is a developers tool available at contrib/pools-management to create/delete ports in the desired pool(s) as well as to control the amount of existing ports loaded into each pool. For more details on this read the readme file on that folder.
Multi pod-vif drivers support with pools¶
There is a multi pool driver that supports hybrid environments where some nodes are Bare Metal while others are running inside VMs, therefore having different VIF drivers (e.g., neutron and nested-vlan).
This new multi pool driver is the default pool driver used even if a different vif_pool_driver is set at the config option. However if the configuration about the mappings between the different pod vif and pools drivers is not provided at the vif_pool_mapping config option of vif_pool configuration section only one pool driver will be loaded – using the standard pod_vif_driver and vif_pool_driver config options, i.e., using the one selected at kuryr.conf options.
To enable the option of having different pools depending on the node’s pod vif types, you need to state the type of pool that you want for each pod vif driver, e.g.:
This will use a pool driver nested to handle the pods whose vif driver is nested-vlan, and a pool driver neutron to handle the pods whose vif driver is neutron-vif. When the controller is requesting a vif for a pod in node X, it will first read the node’s annotation about pod_vif driver to use, e.g., pod_vif: nested-vlan, and then use the corresponding pool driver – which has the right pod-vif driver set.
Previously, pools_vif_drivers configuration option provided similar functionality, but is now deprecated and not recommended. It stored a mapping from pool_driver => pod_vif_driver instead, disallowing the use of a single pool driver as keys for multiple pod_vif_drivers.
Note that if no annotation is set on a node, the default pod_vif_driver is used.
Populate pools on subnets creation for namespace subnet driver¶
When the namespace subnet driver is used (either for namespace isolation or for network policies) a new subnet is created for each namespace. The ports associated to each namespace will therefore be on different pools. In order to prepopulate the pools associated to a newly created namespace (i.e., subnet), the next handler needs to be enabled:
This can be enabled at devstack deployment time to by adding the next to the local.conf: