Configuration and hardening

Configuration and hardening

There are several configuration options and deployment strategies that can improve security in the Data processing service. The service controller is configured through a main configuration file and one or more policy files. Installations that are using the data-locality features will also have two additional files to specify the physical location of Compute and Object Storage nodes.


The Data processing service controller, like many other OpenStack controllers, can be configured to require TLS connections.

Pre-Kilo releases will require a TLS proxy as the controller does not allow direct TLS connections. Configuring TLS proxies is covered in TLS proxies and HTTP services, and we recommend following the advice there to create this type of installation.

From the Kilo release onward the data processing controller allows direct TLS connections, which we recommend. Enabling this behavior requires some small adjustments to the controller configuration file.

Example. Configuring TLS access to the controller

ca_file = cafile.pem
cert_file = certfile.crt
key_file = keyfile.key

Role-based access control policies

The Data processing service uses a policy file, as described in Policies, to configure role-based access control. Using the policy file an operator can restrict a group’s access to specific data processing functionality.

The reasons for doing this will change depending on the organizational requirements of the installation. In general, these fine grained controls are used in situations where an operator needs to restrict the creation, deletion, and retrieval of the Data processing service resources. Operators who need to restrict access within a project should be fully aware that there will need to be alternative means for users to gain access to the core functionality of the service (for example, provisioning clusters).

Example. Allow all methods to all users (default policy)

    "default": ""

Example. Disallow image registry manipulations to non-admin users

    "default": "",

    "data-processing:images:register": "role:admin",
    "data-processing:images:unregister": "role:admin",
    "data-processing:images:add_tags": "role:admin",
    "data-processing:images:remove_tags": "role:admin"

Security groups

The Data processing service allows for the association of security groups with instances provisioned for its clusters. With no additional configuration the service will use the default security group for any project that provisions clusters. A different security group may be used if requested, or an automated option exists which instructs the service to create a security group based on ports specified by the framework being accessed.

For production environments we recommend controlling the security groups manually and creating a set of group rules that are appropriate for the installation. In this manner the operator can ensure that the default security group will contain all the appropriate rules. For an expanded discussion of security groups please see Security groups.

Proxy domains

When using the Object Storage service in conjunction with data processing it is necessary to add credentials for the store access. With proxy domains the Data processing service can instead use a delegated trust from the Identity service to allow store access via a temporary user created in the domain. For this delegation mechanism to work the Data processing service must be configured to use proxy domains and the operator must configure an identity domain for the proxy users.

The data processing controller retains temporary storage of the username and password provided for object store access. When using proxy domains the controller will generate this pair for the proxy user, and the access of this user will be limited to that of the identity trust. We recommend using proxy domains in any installation where the controller or its database have routes to or from public networks.

Example. Configuring for a proxy domain named “dp_proxy”

use_domain_for_proxy_users = true
proxy_user_domain_name = dp_proxy
proxy_user_role_names = Member

Custom network topologies

The data processing controller can be configured to use proxy commands for accessing its cluster instances. In this manner custom network topologies can be created for installations which will not use the networks provided directly by the Networking service. We recommend using this option for installations which require limiting access between the controller and the instances.

Example. Access instances through a specified relay machine

proxy_command='ssh relay-machine-{tenant_id} nc {host} {port}'

Example. Access instances through a custom network namespace

proxy_command='ip netns exec ns_for_{network_id} nc {host} {port}'

Indirect access

For installations in which the controller will have limited access to all the instances of a cluster, due to limits on floating IP addresses or security rules, indirect access may be configured. This allows some instances to be designated as proxy gateways to the other instances of the cluster.

This configuration can only be enabled while defining the node group templates that will make up the data processing clusters. It is provided as a run time option to be enabled during the cluster provisioning process.


When creating custom topologies for network access it can be necessary to allow non-root users the ability to run the proxy commands. For these situations the oslo rootwrap package is used to provide a facility for non-root users to run privileged commands. This configuration requires the user associated with the data processing controller application to be in the sudoers list and for the option to be enabled in the configuration file. Optionally, an alternative rootwrap command can be provided.

Example. Enabling rootwrap usage and showing the default command

rootwrap_command=’sudo sahara-rootwrap /etc/sahara/rootwrap.conf’

For more information on the rootwrap project, please see the official documentation:


Monitoring the output of the service controller is a powerful forensic tool, as described more thoroughly in Monitoring and logging. The Data processing service controller offers a few options for setting the location and level of logging.

Example. Setting the log level higher than warning and specifying an output file.

verbose = true
log_file = /var/log/data-processing.log

Bibliography, Welcome to Sahara!. 2016. Sahara project documentation

The Apache Software Foundation, Welcome to Apache Hadoop!. 2016. Apache Hadoop project

The Apache Software Foundation, Hadoop in Secure Mode. 2016. Hadoop secure mode docs

The Apache Software Foundation, HDFS User Guide. 2016. Hadoop HDFS documentation

The Apache Software Foundation, Spark. 2016. Spark project

The Apache Software Foundation, Spark Security. 2016. Spark security documentation

The Apache Software Foundation, Apache Storm. 2016. Storm project

The Apache Software Foundation, Apache Zookeeper. 2016. Zookeeper project

The Apache Software Foundation, Apache Oozie Workflow Scheduler for Hadoop. 2016. Oozie project

The Apache Software Foundation, Apache Hive. 2016. Hive

The Apache Software Foundation, Welcome to Apache Pig. 2016. Pig

The Apache Software Foundation, Cloudera Product Documentation. 2016. Cloudera CDH documentation

Hortonworks, Hortonworks. 2016. Hortonworks Data Platform documentation

MapR Technologies, Apache Hadoop for the MapR Converged Data Platform. 2016. MapR project

Creative Commons Attribution 3.0 License

Except where otherwise noted, this document is licensed under Creative Commons Attribution 3.0 License. See all OpenStack Legal Documents.