Source code for ironic.drivers.modules.inspector.interface

# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
#      http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.

"""
Modules required to work with ironic_inspector:
    https://pypi.org/project/ironic-inspector
"""

from urllib import parse as urlparse

import eventlet
from oslo_log import log as logging

from ironic.common import exception
from ironic.common.i18n import _
from ironic.common import states
from ironic.common import utils
from ironic.conductor import periodics
from ironic.conductor import task_manager
from ironic.conductor import utils as cond_utils
from ironic.conf import CONF
from ironic.drivers import base
from ironic.drivers.modules import deploy_utils
from ironic.drivers.modules import inspect_utils
from ironic.drivers.modules.inspector import client

LOG = logging.getLogger(__name__)

# Internal field to mark whether ironic or inspector manages boot for the node
_IRONIC_MANAGES_BOOT = 'inspector_manage_boot'


def _get_callback_endpoint(client):
    root = CONF.inspector.callback_endpoint_override or client.get_endpoint()
    if root == 'mdns':
        return root

    parts = urlparse.urlsplit(root)

    if utils.is_loopback(parts.hostname):
        raise exception.InvalidParameterValue(
            _('Loopback address %s cannot be used as an introspection '
              'callback URL') % parts.hostname)

    # NOTE(dtantsur): the IPA side is quite picky about the exact format.
    if parts.path.endswith('/v1'):
        add = '/continue'
    else:
        add = '/v1/continue'

    return urlparse.urlunsplit((parts.scheme, parts.netloc,
                                parts.path.rstrip('/') + add,
                                parts.query, parts.fragment))


[docs] def tear_down_managed_boot(task): errors = [] ironic_manages_boot = utils.pop_node_nested_field( task.node, 'driver_internal_info', _IRONIC_MANAGES_BOOT) if not ironic_manages_boot: return errors try: task.driver.boot.clean_up_ramdisk(task) except Exception as exc: errors.append(_('unable to clean up ramdisk boot: %s') % exc) LOG.exception('Unable to clean up ramdisk boot for node %s', task.node.uuid) try: with cond_utils.power_state_for_network_configuration(task): task.driver.network.remove_inspection_network(task) except Exception as exc: errors.append(_('unable to remove inspection ports: %s') % exc) LOG.exception('Unable to remove inspection network for node %s', task.node.uuid) if CONF.inspector.power_off and not utils.fast_track_enabled(task.node): try: cond_utils.node_power_action(task, states.POWER_OFF) except Exception as exc: errors.append(_('unable to power off the node: %s') % exc) LOG.exception('Unable to power off node %s', task.node.uuid) return errors
[docs] def inspection_error_handler(task, error, raise_exc=False, clean_up=True): if clean_up: tear_down_managed_boot(task) task.node.last_error = error if raise_exc: task.node.save() raise exception.HardwareInspectionFailure(error=error) else: task.process_event('fail')
[docs] def ironic_manages_boot(task, raise_exc=False): """Whether ironic should manage boot for this node.""" try: task.driver.boot.validate_inspection(task) except exception.UnsupportedDriverExtension as e: LOG.debug('The boot interface %(iface)s of the node %(node)s does ' 'not support managed boot for in-band inspection or ' 'the required options are not populated: %(exc)s', {'node': task.node.uuid, 'iface': task.node.get_interface('boot'), 'exc': e}) if raise_exc: raise return False try: task.driver.network.validate_inspection(task) except exception.UnsupportedDriverExtension as e: LOG.debug('The network interface %(iface)s of the node %(node)s does ' 'not support managed boot for in-band inspection or ' 'the required options are not populated: %(exc)s', {'node': task.node.uuid, 'iface': task.node.get_interface('network'), 'exc': e}) if raise_exc: raise return False return True
[docs] def prepare_managed_inspection(task, endpoint): """Prepare the boot interface for managed inspection.""" params = dict( utils.parse_kernel_params(CONF.inspector.extra_kernel_params), **{'ipa-inspection-callback-url': endpoint}) if utils.fast_track_enabled(task.node): params['ipa-api-url'] = deploy_utils.get_ironic_api_url() cond_utils.node_power_action(task, states.POWER_OFF) with cond_utils.power_state_for_network_configuration(task): task.driver.network.add_inspection_network(task) task.driver.boot.prepare_ramdisk(task, ramdisk_params=params)
[docs] class Common(base.InspectInterface): default_require_managed_boot = False def __init__(self): super().__init__() if CONF.inspector.require_managed_boot is None: LOG.warning("The option [inspector]require_managed_boot will " "change its default value to True in the future. " "Set it to an explicit boolean value to avoid a " "potential breakage.") def _require_managed_boot(self): return (CONF.inspector.require_managed_boot if CONF.inspector.require_managed_boot is not None else self.default_require_managed_boot)
[docs] def get_properties(self): """Return the properties of the interface. :returns: dictionary of <property name>:<property description> entries. """ return {} # no properties
[docs] def validate(self, task): """Validate the driver-specific inspection information. If invalid, raises an exception; otherwise returns None. :param task: a task from TaskManager. :raises: UnsupportedDriverExtension """ utils.parse_kernel_params(CONF.inspector.extra_kernel_params) if self._require_managed_boot(): ironic_manages_boot(task, raise_exc=True)
[docs] def inspect_hardware(self, task): """Inspect hardware to obtain the hardware properties. This particular implementation only starts inspection using ironic-inspector. Results will be checked in a periodic task. :param task: a task from TaskManager. :returns: states.INSPECTWAIT :raises: HardwareInspectionFailure on failure """ try: inspect_utils.create_ports_if_not_exist(task) except exception.UnsupportedDriverExtension: LOG.debug('Pre-creating ports prior to inspection not supported' ' on node %s.', task.node.uuid) manage_boot = ironic_manages_boot( task, raise_exc=self._require_managed_boot()) utils.set_node_nested_field(task.node, 'driver_internal_info', _IRONIC_MANAGES_BOOT, manage_boot) # Make this interface work with the Ironic own /continue_inspection # endpoint to simplify migration to the new in-band inspection # implementation. inspect_utils.cache_lookup_addresses(task.node) task.node.save() LOG.debug('Starting inspection for node %(uuid)s using ' 'ironic-inspector, booting is managed by %(project)s', {'uuid': task.node.uuid, 'project': 'ironic' if manage_boot else 'ironic-inspector'}) if manage_boot: try: self._start_managed_inspection(task) except Exception as exc: LOG.exception('Unable to start managed inspection for node ' '%(uuid)s: %(err)s', {'uuid': task.node.uuid, 'err': exc}) error = _('unable to start inspection: %s') % exc inspection_error_handler(task, error, raise_exc=True) else: self._start_unmanaged_inspection(task) return states.INSPECTWAIT
[docs] class Inspector(Common): """In-band inspection via ironic-inspector project.""" def _start_managed_inspection(self, task): """Start inspection managed by ironic.""" cli = client.get_client(task.context) endpoint = _get_callback_endpoint(cli) prepare_managed_inspection(task, endpoint) cli.start_introspection(task.node.uuid, manage_boot=False) cond_utils.node_power_action(task, states.POWER_ON) def _start_unmanaged_inspection(self, task): """Call to inspector to start inspection.""" # NOTE(dtantsur): spawning a short-living green thread so that # we can release a lock as soon as possible and allow # ironic-inspector to operate on the node. eventlet.spawn_n(_start_inspection, task.node.uuid, task.context)
[docs] def abort(self, task): """Abort hardware inspection. :param task: a task from TaskManager. """ node_uuid = task.node.uuid LOG.debug('Aborting inspection for node %(uuid)s using ' 'ironic-inspector', {'uuid': node_uuid}) client.get_client(task.context).abort_introspection(node_uuid) if inspect_utils.clear_lookup_addresses(task.node): task.node.save()
@periodics.node_periodic( purpose='checking hardware inspection status', spacing=CONF.inspector.status_check_period, filters={'provision_state': states.INSPECTWAIT}, ) def _periodic_check_result(self, task, manager, context): """Periodic task checking results of inspection.""" if isinstance(task.driver.inspect, self.__class__): _check_status(task)
[docs] def continue_inspection(self, task, inventory, plugin_data=None): """Continue in-band hardware inspection. This implementation simply defers to ironic-inspector. It only exists to simplify the transition to Ironic-native in-band inspection. :param task: a task from TaskManager. :param inventory: hardware inventory from the node. :param plugin_data: optional plugin-specific data. """ cli = client.get_client(task.context) endpoint = _get_callback_endpoint(cli) data = dict(plugin_data, inventory=inventory) # older format task.process_event('wait') task.downgrade_lock() cli.post(endpoint, json=data) return states.INSPECTWAIT
def _start_inspection(node_uuid, context): """Call to inspector to start inspection.""" try: client.get_client(context).start_introspection(node_uuid) except Exception as exc: LOG.error('Error contacting ironic-inspector for inspection of node ' '%(node)s: %(cls)s: %(err)s', {'node': node_uuid, 'cls': type(exc).__name__, 'err': exc}) # NOTE(dtantsur): if acquire fails our last option is to rely on # timeout lock_purpose = 'recording hardware inspection error' with task_manager.acquire(context, node_uuid, purpose=lock_purpose) as task: error = _('Failed to start inspection: %s') % exc inspection_error_handler(task, error) else: LOG.info('Node %s was sent to inspection to ironic-inspector', node_uuid) def _check_status(task): """Check inspection status for node given by a task.""" node = task.node if node.provision_state != states.INSPECTWAIT: return if not isinstance(task.driver.inspect, Inspector): return LOG.debug('Calling to inspector to check status of node %s', task.node.uuid) try: inspector_client = client.get_client(task.context) status = inspector_client.get_introspection(node.uuid) except Exception: # NOTE(dtantsur): get_status should not normally raise # let's assume it's a transient failure and retry later LOG.exception('Unexpected exception while getting ' 'inspection status for node %s, will retry later', node.uuid) return if not status.error and not status.is_finished: return # If the inspection has finished or failed, we need to update the node, so # upgrade our lock to an exclusive one. task.upgrade_lock() node = task.node inspect_utils.clear_lookup_addresses(node) if status.error: LOG.error('Inspection failed for node %(uuid)s with error: %(err)s', {'uuid': node.uuid, 'err': status.error}) error = _('ironic-inspector inspection failed: %s') % status.error inspection_error_handler(task, error) elif status.is_finished: clean_up(task) if CONF.inventory.data_backend == 'none': LOG.debug('Inspection data storage is disabled, the data will ' 'not be saved for node %s', node.uuid) return introspection_data = inspector_client.get_introspection_data( node.uuid, processed=True) # TODO(dtantsur): having no inventory is an abnormal state, handle it. inventory = introspection_data.pop('inventory', {}) inspect_utils.store_inspection_data(node, inventory, introspection_data, task.context)
[docs] def clean_up(task, finish=True): errors = tear_down_managed_boot(task) if errors: errors = ', '.join(errors) LOG.error('Inspection clean up failed for node %(uuid)s: %(err)s', {'uuid': task.node.uuid, 'err': errors}) msg = _('Inspection clean up failed: %s') % errors inspection_error_handler(task, msg, raise_exc=False, clean_up=False) elif finish: LOG.info('Inspection finished successfully for node %s', task.node.uuid) task.process_event('done')