|  | #!/bin/bash | 
|  | # | 
|  | # lib/nova | 
|  | # Functions to control the configuration and operation of the **Nova** service | 
|  |  | 
|  | # Dependencies: | 
|  | # | 
|  | # - ``functions`` file | 
|  | # - ``DEST``, ``DATA_DIR``, ``STACK_USER`` must be defined | 
|  | # - ``FILES`` | 
|  | # - ``SERVICE_{TENANT_NAME|PASSWORD}`` must be defined | 
|  | # - ``LIBVIRT_TYPE`` must be defined | 
|  | # - ``INSTANCE_NAME_PREFIX``, ``VOLUME_NAME_PREFIX`` must be defined | 
|  | # - ``KEYSTONE_TOKEN_FORMAT`` must be defined | 
|  |  | 
|  | # ``stack.sh`` calls the entry points in this order: | 
|  | # | 
|  | # - install_nova | 
|  | # - configure_nova | 
|  | # - create_nova_conf | 
|  | # - init_nova | 
|  | # - start_nova | 
|  | # - stop_nova | 
|  | # - cleanup_nova | 
|  |  | 
|  | # Save trace setting | 
|  | _XTRACE_LIB_NOVA=$(set +o | grep xtrace) | 
|  | set +o xtrace | 
|  |  | 
|  | # Defaults | 
|  | # -------- | 
|  |  | 
|  | # Set up default directories | 
|  | GITDIR["python-novaclient"]=$DEST/python-novaclient | 
|  | GITDIR["os-vif"]=$DEST/os-vif | 
|  | NOVA_DIR=$DEST/nova | 
|  |  | 
|  | # Nova virtual environment | 
|  | if [[ ${USE_VENV} = True ]]; then | 
|  | PROJECT_VENV["nova"]=${NOVA_DIR}.venv | 
|  | NOVA_BIN_DIR=${PROJECT_VENV["nova"]}/bin | 
|  | else | 
|  | NOVA_BIN_DIR=$(get_python_exec_prefix) | 
|  | fi | 
|  |  | 
|  | NOVA_STATE_PATH=${NOVA_STATE_PATH:=$DATA_DIR/nova} | 
|  | # INSTANCES_PATH is the previous name for this | 
|  | NOVA_INSTANCES_PATH=${NOVA_INSTANCES_PATH:=${INSTANCES_PATH:=$NOVA_STATE_PATH/instances}} | 
|  |  | 
|  | NOVA_CONF_DIR=/etc/nova | 
|  | NOVA_CONF=$NOVA_CONF_DIR/nova.conf | 
|  | NOVA_COND_CONF=$NOVA_CONF_DIR/nova.conf | 
|  | NOVA_CPU_CONF=$NOVA_CONF_DIR/nova-cpu.conf | 
|  | NOVA_FAKE_CONF=$NOVA_CONF_DIR/nova-fake.conf | 
|  | NOVA_API_DB=${NOVA_API_DB:-nova_api} | 
|  | NOVA_UWSGI=nova.wsgi.osapi_compute:application | 
|  | NOVA_METADATA_UWSGI=nova.wsgi.metadata:application | 
|  | NOVA_UWSGI_CONF=$NOVA_CONF_DIR/nova-api-uwsgi.ini | 
|  | NOVA_METADATA_UWSGI_CONF=$NOVA_CONF_DIR/nova-metadata-uwsgi.ini | 
|  |  | 
|  | # Allow forcing the stable compute uuid to something specific. This would be | 
|  | # done by deployment tools that pre-allocate the UUIDs, but it is also handy | 
|  | # for developers that need to re-stack a compute-only deployment multiple | 
|  | # times. Since the DB is non-local and not erased on an unstack, making it | 
|  | # stay the same each time is what developers want. Set to a uuid here or | 
|  | # leave it blank for default allocate-on-start behavior. | 
|  | NOVA_CPU_UUID="" | 
|  |  | 
|  | # The total number of cells we expect. Must be greater than one and doesn't | 
|  | # count cell0. | 
|  | NOVA_NUM_CELLS=${NOVA_NUM_CELLS:-1} | 
|  | # Our cell index, so we know what rabbit vhost to connect to. | 
|  | # This should be in the range of 1-$NOVA_NUM_CELLS | 
|  | NOVA_CPU_CELL=${NOVA_CPU_CELL:-1} | 
|  |  | 
|  | NOVA_API_PASTE_INI=${NOVA_API_PASTE_INI:-$NOVA_CONF_DIR/api-paste.ini} | 
|  |  | 
|  | # We do not need to report service status every 10s for devstack-like | 
|  | # deployments. In the gate this generates extra work for the services and the | 
|  | # database which are already taxed. | 
|  | NOVA_SERVICE_REPORT_INTERVAL=${NOVA_SERVICE_REPORT_INTERVAL:-120} | 
|  |  | 
|  | if is_service_enabled tls-proxy; then | 
|  | NOVA_SERVICE_PROTOCOL="https" | 
|  | fi | 
|  |  | 
|  | # Whether to use TLS for comms between the VNC/SPICE/serial proxy | 
|  | # services and the compute node | 
|  | NOVA_CONSOLE_PROXY_COMPUTE_TLS=${NOVA_CONSOLE_PROXY_COMPUTE_TLS:-False} | 
|  |  | 
|  | # Validate configuration | 
|  | if ! is_service_enabled tls-proxy && [ "$NOVA_CONSOLE_PROXY_COMPUTE_TLS" == "True" ]; then | 
|  | die $LINENO "enabling TLS for the console proxy requires the tls-proxy service" | 
|  | fi | 
|  |  | 
|  | # Public facing bits | 
|  | NOVA_SERVICE_HOST=${NOVA_SERVICE_HOST:-$SERVICE_HOST} | 
|  | NOVA_SERVICE_PORT=${NOVA_SERVICE_PORT:-8774} | 
|  | NOVA_SERVICE_PORT_INT=${NOVA_SERVICE_PORT_INT:-18774} | 
|  | NOVA_SERVICE_PROTOCOL=${NOVA_SERVICE_PROTOCOL:-$SERVICE_PROTOCOL} | 
|  | NOVA_SERVICE_LISTEN_ADDRESS=${NOVA_SERVICE_LISTEN_ADDRESS:-$(ipv6_unquote $SERVICE_LISTEN_ADDRESS)} | 
|  | METADATA_SERVICE_PORT=${METADATA_SERVICE_PORT:-8775} | 
|  | NOVA_ENABLE_CACHE=${NOVA_ENABLE_CACHE:-True} | 
|  |  | 
|  | # Flag to set the oslo_policy.enforce_scope and oslo_policy.enforce_new_defaults. | 
|  | # This is used to disable the compute API policies scope and new defaults. | 
|  | # By Default, it is True. | 
|  | # For more detail: https://docs.openstack.org/oslo.policy/latest/configuration/index.html#oslo_policy.enforce_scope | 
|  | NOVA_ENFORCE_SCOPE=$(trueorfalse True NOVA_ENFORCE_SCOPE) | 
|  |  | 
|  | if [[ $SERVICE_IP_VERSION == 6 ]]; then | 
|  | NOVA_MY_IP="$HOST_IPV6" | 
|  | else | 
|  | NOVA_MY_IP="$HOST_IP" | 
|  | fi | 
|  |  | 
|  | # Option to enable/disable config drive | 
|  | # NOTE: Set ``FORCE_CONFIG_DRIVE="False"`` to turn OFF config drive | 
|  | FORCE_CONFIG_DRIVE=${FORCE_CONFIG_DRIVE:-"False"} | 
|  |  | 
|  | # The following NOVA_FILTERS contains SameHostFilter and DifferentHostFilter with | 
|  | # the default filters. | 
|  | NOVA_FILTERS="ComputeFilter,ComputeCapabilitiesFilter,ImagePropertiesFilter,ServerGroupAntiAffinityFilter,ServerGroupAffinityFilter,SameHostFilter,DifferentHostFilter" | 
|  |  | 
|  | QEMU_CONF=/etc/libvirt/qemu.conf | 
|  |  | 
|  | # ``NOVA_VNC_ENABLED`` can be used to forcibly enable VNC configuration. | 
|  | # In multi-node setups allows compute hosts to not run ``n-novnc``. | 
|  | NOVA_VNC_ENABLED=$(trueorfalse False NOVA_VNC_ENABLED) | 
|  |  | 
|  | # Get hypervisor configuration | 
|  | # ---------------------------- | 
|  |  | 
|  | NOVA_PLUGINS=$TOP_DIR/lib/nova_plugins | 
|  | if is_service_enabled nova && [[ -r $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER ]]; then | 
|  | # Load plugin | 
|  | source $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER | 
|  | fi | 
|  |  | 
|  | # Other Nova configurations | 
|  | # ---------------------------- | 
|  |  | 
|  | # ``NOVA_USE_SERVICE_TOKEN`` is a mode where service token is passed along with | 
|  | # user token while communicating to external RESP API's like Neutron, Cinder | 
|  | # and Glance. | 
|  | NOVA_USE_SERVICE_TOKEN=$(trueorfalse True NOVA_USE_SERVICE_TOKEN) | 
|  |  | 
|  | # ``NOVA_ALLOW_MOVE_TO_SAME_HOST`` can be set to False in multi node DevStack, | 
|  | # where there are at least two nova-computes. | 
|  | NOVA_ALLOW_MOVE_TO_SAME_HOST=$(trueorfalse True NOVA_ALLOW_MOVE_TO_SAME_HOST) | 
|  |  | 
|  | # Enable debugging levels for iscsid service (goes from 0-8) | 
|  | ISCSID_DEBUG=$(trueorfalse False ISCSID_DEBUG) | 
|  | ISCSID_DEBUG_LEVEL=${ISCSID_DEBUG_LEVEL:-4} | 
|  |  | 
|  | # Format for notifications. Nova defaults to "unversioned" since Train. | 
|  | # Other options include "versioned" and "both". | 
|  | NOVA_NOTIFICATION_FORMAT=${NOVA_NOTIFICATION_FORMAT:-unversioned} | 
|  |  | 
|  | # Timeout for servers to gracefully shutdown the OS during operations | 
|  | # like shelve, rescue, stop, rebuild. Defaults to 0 since the default | 
|  | # image in devstack is CirrOS. | 
|  | NOVA_SHUTDOWN_TIMEOUT=${NOVA_SHUTDOWN_TIMEOUT:-0} | 
|  |  | 
|  | # Whether to use Keystone unified limits instead of legacy quota limits. | 
|  | NOVA_USE_UNIFIED_LIMITS=$(trueorfalse False NOVA_USE_UNIFIED_LIMITS) | 
|  |  | 
|  | # TB Cache Size in MiB for qemu guests | 
|  | NOVA_LIBVIRT_TB_CACHE_SIZE=${NOVA_LIBVIRT_TB_CACHE_SIZE:-0} | 
|  |  | 
|  | # Functions | 
|  | # --------- | 
|  |  | 
|  | # Test if any Nova services are enabled | 
|  | # is_nova_enabled | 
|  | function is_nova_enabled { | 
|  | [[ ,${DISABLED_SERVICES} =~ ,"nova" ]] && return 1 | 
|  | [[ ,${ENABLED_SERVICES} =~ ,"n-" ]] && return 0 | 
|  | return 1 | 
|  | } | 
|  |  | 
|  | # is_nova_console_proxy_compute_tls_enabled() - Test if the Nova Console Proxy | 
|  | # service has TLS enabled | 
|  | function is_nova_console_proxy_compute_tls_enabled { | 
|  | [[ ${NOVA_CONSOLE_PROXY_COMPUTE_TLS} = "True" ]] && return 0 | 
|  | return 1 | 
|  | } | 
|  |  | 
|  | # Helper to clean iptables rules | 
|  | function clean_iptables { | 
|  | # Delete rules | 
|  | sudo iptables -S -v | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-A" |  sed "s/-A/-D/g" | awk '{print "sudo iptables",$0}' | bash | 
|  | # Delete nat rules | 
|  | sudo iptables -S -v -t nat | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" |  grep "\-A" | sed "s/-A/-D/g" | awk '{print "sudo iptables -t nat",$0}' | bash | 
|  | # Delete chains | 
|  | sudo iptables -S -v | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" | grep "\-N" |  sed "s/-N/-X/g" | awk '{print "sudo iptables",$0}' | bash | 
|  | # Delete nat chains | 
|  | sudo iptables -S -v -t nat | sed "s/-c [0-9]* [0-9]* //g" | grep "nova" |  grep "\-N" | sed "s/-N/-X/g" | awk '{print "sudo iptables -t nat",$0}' | bash | 
|  | } | 
|  |  | 
|  | # cleanup_nova() - Remove residual data files, anything left over from previous | 
|  | # runs that a clean run would need to clean up | 
|  | function cleanup_nova { | 
|  | if is_service_enabled n-cpu; then | 
|  | # Clean iptables from previous runs | 
|  | clean_iptables | 
|  |  | 
|  | # Destroy old instances | 
|  | local instances | 
|  | instances=`sudo virsh list --all | grep $INSTANCE_NAME_PREFIX | sed "s/.*\($INSTANCE_NAME_PREFIX[0-9a-fA-F]*\).*/\1/g"` | 
|  | if [ ! "$instances" = "" ]; then | 
|  | echo $instances | xargs -n1 sudo virsh destroy || true | 
|  | if ! xargs -n1 sudo virsh undefine --managed-save --nvram <<< $instances; then | 
|  | # Can't delete with nvram flags, then just try without this flag | 
|  | xargs -n1 sudo virsh undefine --managed-save <<< $instances | 
|  | fi | 
|  | fi | 
|  |  | 
|  | # Logout and delete iscsi sessions | 
|  | local tgts | 
|  | tgts=$(sudo iscsiadm --mode node | grep $VOLUME_NAME_PREFIX | cut -d ' ' -f2) | 
|  | local target | 
|  | for target in $tgts; do | 
|  | sudo iscsiadm --mode node -T $target --logout || true | 
|  | done | 
|  | sudo iscsiadm --mode node --op delete || true | 
|  |  | 
|  | # Disconnect all nvmeof connections | 
|  | sudo nvme disconnect-all || true | 
|  |  | 
|  | # Clean out the instances directory. | 
|  | sudo rm -rf $NOVA_INSTANCES_PATH/* | 
|  | fi | 
|  |  | 
|  | sudo rm -rf $NOVA_STATE_PATH | 
|  |  | 
|  | # NOTE(dtroyer): This really should be called from here but due to the way | 
|  | #                nova abuses the _cleanup() function we're moving it | 
|  | #                directly into cleanup.sh until this can be fixed. | 
|  | #if is_service_enabled n-cpu && [[ -r $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER ]]; then | 
|  | #    cleanup_nova_hypervisor | 
|  | #fi | 
|  |  | 
|  | stop_process "n-api" | 
|  | stop_process "n-api-meta" | 
|  | remove_uwsgi_config "$NOVA_UWSGI_CONF" "nova-api" | 
|  | remove_uwsgi_config "$NOVA_METADATA_UWSGI_CONF" "nova-metadata" | 
|  |  | 
|  | if [[ "$NOVA_BACKEND" == "LVM" ]]; then | 
|  | clean_lvm_volume_group $DEFAULT_VOLUME_GROUP_NAME | 
|  | fi | 
|  | } | 
|  |  | 
|  | # configure_nova() - Set config files, create data dirs, etc | 
|  | function configure_nova { | 
|  | # Put config files in ``/etc/nova`` for everyone to find | 
|  | sudo install -d -o $STACK_USER $NOVA_CONF_DIR | 
|  |  | 
|  | configure_rootwrap nova | 
|  |  | 
|  | if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then | 
|  | # Get the sample configuration file in place | 
|  | cp $NOVA_DIR/etc/nova/api-paste.ini $NOVA_CONF_DIR | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-cpu; then | 
|  | # Force IP forwarding on, just on case | 
|  | sudo sysctl -w net.ipv4.ip_forward=1 | 
|  |  | 
|  | if [[ "$VIRT_DRIVER" = 'libvirt' ]]; then | 
|  | # Check for kvm (hardware based virtualization).  If unable to initialize | 
|  | # kvm, we drop back to the slower emulation mode (qemu).  Note: many systems | 
|  | # come with hardware virtualization disabled in BIOS. | 
|  | if [[ "$LIBVIRT_TYPE" == "kvm" ]]; then | 
|  | sudo modprobe kvm || true | 
|  | if [ ! -e /dev/kvm ]; then | 
|  | echo "WARNING: Switching to QEMU" | 
|  | LIBVIRT_TYPE=qemu | 
|  | LIBVIRT_CPU_MODE=custom | 
|  | LIBVIRT_CPU_MODEL=Nehalem | 
|  | if which selinuxenabled >/dev/null 2>&1 && selinuxenabled; then | 
|  | # https://bugzilla.redhat.com/show_bug.cgi?id=753589 | 
|  | sudo setsebool virt_use_execmem on | 
|  | fi | 
|  | fi | 
|  | fi | 
|  |  | 
|  | # Install and configure **LXC** if specified.  LXC is another approach to | 
|  | # splitting a system into many smaller parts.  LXC uses cgroups and chroot | 
|  | # to simulate multiple systems. | 
|  | if [[ "$LIBVIRT_TYPE" == "lxc" ]]; then | 
|  | if is_ubuntu; then | 
|  | # enable nbd for lxc unless you're using an lvm backend | 
|  | # otherwise you can't boot instances | 
|  | if [[ "$NOVA_BACKEND" != "LVM" ]]; then | 
|  | sudo modprobe nbd | 
|  | fi | 
|  | fi | 
|  | fi | 
|  | fi | 
|  |  | 
|  | # Instance Storage | 
|  | # ---------------- | 
|  |  | 
|  | # Nova stores each instance in its own directory. | 
|  | sudo install -d -o $STACK_USER $NOVA_INSTANCES_PATH | 
|  |  | 
|  | # You can specify a different disk to be mounted and used for backing the | 
|  | # virtual machines.  If there is a partition labeled nova-instances we | 
|  | # mount it (ext filesystems can be labeled via e2label). | 
|  | if [ -L /dev/disk/by-label/nova-instances ]; then | 
|  | if ! mount -n | grep -q $NOVA_INSTANCES_PATH; then | 
|  | sudo mount -L nova-instances $NOVA_INSTANCES_PATH | 
|  | sudo chown -R $STACK_USER $NOVA_INSTANCES_PATH | 
|  | fi | 
|  | fi | 
|  |  | 
|  | # Due to cinder bug #1966513 we ALWAYS need an initiator name for LVM | 
|  | # Ensure each compute host uses a unique iSCSI initiator | 
|  | echo InitiatorName=$(iscsi-iname) | sudo tee /etc/iscsi/initiatorname.iscsi | 
|  |  | 
|  | if [[ ${ISCSID_DEBUG} == "True" ]]; then | 
|  | # Install an override that starts iscsid with debugging | 
|  | # enabled. | 
|  | cat > /tmp/iscsid.override <<EOF | 
|  | [Service] | 
|  | ExecStart= | 
|  | ExecStart=/usr/sbin/iscsid -d${ISCSID_DEBUG_LEVEL} | 
|  | EOF | 
|  | sudo mkdir -p /etc/systemd/system/iscsid.service.d | 
|  | sudo mv /tmp/iscsid.override /etc/systemd/system/iscsid.service.d/override.conf | 
|  | sudo systemctl daemon-reload | 
|  | fi | 
|  |  | 
|  | # set chap algorithms.  The default chap_algorithm is md5 which will | 
|  | # not work under FIPS. | 
|  | iniset -sudo /etc/iscsi/iscsid.conf DEFAULT "node.session.auth.chap_algs" "SHA3-256,SHA256" | 
|  |  | 
|  | if [[ $CINDER_TARGET_HELPER != 'nvmet' ]]; then | 
|  | # ensure that iscsid is started, even when disabled by default | 
|  | restart_service iscsid | 
|  |  | 
|  | # For NVMe-oF we need different packages that many not be present | 
|  | else | 
|  | install_package nvme-cli | 
|  | sudo modprobe nvme-fabrics | 
|  |  | 
|  | # Ensure NVMe is ready and create the Soft-RoCE device over the networking interface | 
|  | if [[ $CINDER_TARGET_PROTOCOL == 'nvmet_rdma' ]]; then | 
|  | sudo modprobe nvme-rdma | 
|  | iface=${HOST_IP_IFACE:-`ip -br -$SERVICE_IP_VERSION a | grep $NOVA_MY_IP | awk '{print $1}'`} | 
|  | if ! sudo rdma link | grep $iface ; then | 
|  | sudo rdma link add rxe_$iface type rxe netdev $iface | 
|  | fi | 
|  | elif [[ $CINDER_TARGET_PROTOCOL == 'nvmet_tcp' ]]; then | 
|  | sudo modprobe nvme-tcp | 
|  | else  # 'nvmet_fc' | 
|  | sudo modprobe nvme-fc | 
|  | fi | 
|  | fi | 
|  | fi | 
|  |  | 
|  | # Rebuild the config file from scratch | 
|  | create_nova_conf | 
|  |  | 
|  | if is_service_enabled n-cpu && [[ -r $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER ]]; then | 
|  | # Configure hypervisor plugin | 
|  | configure_nova_hypervisor | 
|  | fi | 
|  | } | 
|  |  | 
|  | # create_nova_accounts() - Set up common required nova accounts | 
|  | # | 
|  | # Project              User         Roles | 
|  | # ------------------------------------------------------------------ | 
|  | # SERVICE_PROJECT_NAME  nova         admin | 
|  | # SERVICE_PROJECT_NAME  nova         ResellerAdmin (if Swift is enabled) | 
|  | function create_nova_accounts { | 
|  |  | 
|  | # Nova | 
|  | if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then | 
|  |  | 
|  | # NOTE(jamielennox): Nova doesn't need the admin role here, however neutron uses | 
|  | # this service user when notifying nova of changes and that requires the admin role. | 
|  | create_service_user "nova" "admin" | 
|  |  | 
|  | local nova_api_url | 
|  | nova_api_url="$NOVA_SERVICE_PROTOCOL://$NOVA_SERVICE_HOST/compute" | 
|  |  | 
|  | get_or_create_service "nova_legacy" "compute_legacy" "Nova Compute Service (Legacy 2.0)" | 
|  | get_or_create_endpoint \ | 
|  | "compute_legacy" \ | 
|  | "$REGION_NAME" \ | 
|  | "$nova_api_url/v2/\$(project_id)s" | 
|  |  | 
|  | get_or_create_service "nova" "compute" "Nova Compute Service" | 
|  | get_or_create_endpoint \ | 
|  | "compute" \ | 
|  | "$REGION_NAME" \ | 
|  | "$nova_api_url/v2.1" | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-api; then | 
|  | # Swift | 
|  | if is_service_enabled swift; then | 
|  | # Nova needs ResellerAdmin role to download images when accessing | 
|  | # swift through the s3 api. | 
|  | get_or_add_user_project_role ResellerAdmin nova $SERVICE_PROJECT_NAME $SERVICE_DOMAIN_NAME $SERVICE_DOMAIN_NAME | 
|  | fi | 
|  | fi | 
|  |  | 
|  | # S3 | 
|  | if is_service_enabled s3api; then | 
|  | get_or_create_service "s3" "s3" "S3" | 
|  | get_or_create_endpoint \ | 
|  | "s3" \ | 
|  | "$REGION_NAME" \ | 
|  | "http://$SERVICE_HOST:$S3_SERVICE_PORT" \ | 
|  | "http://$SERVICE_HOST:$S3_SERVICE_PORT" \ | 
|  | "http://$SERVICE_HOST:$S3_SERVICE_PORT" | 
|  | fi | 
|  |  | 
|  | # Unified limits | 
|  | if is_service_enabled n-api; then | 
|  | if [[ "$NOVA_USE_UNIFIED_LIMITS" = True ]]; then | 
|  | configure_nova_unified_limits | 
|  | fi | 
|  | fi | 
|  | } | 
|  |  | 
|  | # create_nova_conf() - Create a new nova.conf file | 
|  | function create_nova_conf { | 
|  | # Remove legacy ``nova.conf`` | 
|  | rm -f $NOVA_DIR/bin/nova.conf | 
|  |  | 
|  | # (Re)create ``nova.conf`` | 
|  | rm -f $NOVA_CONF | 
|  | iniset $NOVA_CONF DEFAULT debug "$ENABLE_DEBUG_LOG_LEVEL" | 
|  | if [ "$NOVA_ALLOW_MOVE_TO_SAME_HOST" == "True" ]; then | 
|  | iniset $NOVA_CONF DEFAULT allow_resize_to_same_host "True" | 
|  | fi | 
|  | iniset $NOVA_CONF wsgi api_paste_config "$NOVA_API_PASTE_INI" | 
|  | iniset $NOVA_CONF DEFAULT rootwrap_config "$NOVA_CONF_DIR/rootwrap.conf" | 
|  | iniset $NOVA_CONF filter_scheduler enabled_filters "$NOVA_FILTERS" | 
|  | iniset $NOVA_CONF scheduler workers "$API_WORKERS" | 
|  | iniset $NOVA_CONF neutron default_floating_pool "$PUBLIC_NETWORK_NAME" | 
|  | iniset $NOVA_CONF DEFAULT my_ip "$NOVA_MY_IP" | 
|  | iniset $NOVA_CONF DEFAULT instance_name_template "${INSTANCE_NAME_PREFIX}%08x" | 
|  | iniset $NOVA_CONF DEFAULT osapi_compute_listen "$NOVA_SERVICE_LISTEN_ADDRESS" | 
|  | iniset $NOVA_CONF DEFAULT metadata_listen "$NOVA_SERVICE_LISTEN_ADDRESS" | 
|  | iniset $NOVA_CONF DEFAULT shutdown_timeout $NOVA_SHUTDOWN_TIMEOUT | 
|  |  | 
|  | iniset $NOVA_CONF key_manager backend nova.keymgr.conf_key_mgr.ConfKeyManager | 
|  |  | 
|  | iniset $NOVA_CONF DEFAULT report_interval $NOVA_SERVICE_REPORT_INTERVAL | 
|  | iniset $NOVA_CONF DEFAULT service_down_time $(($NOVA_SERVICE_REPORT_INTERVAL * 6)) | 
|  |  | 
|  | if is_fedora; then | 
|  | # nova defaults to /usr/local/bin, but fedora pip like to | 
|  | # install things in /usr/bin | 
|  | iniset $NOVA_CONF DEFAULT bindir "/usr/bin" | 
|  | fi | 
|  |  | 
|  | # only setup database connections and cache backend if there are services | 
|  | # that require them running on the host. The ensures that n-cpu doesn't | 
|  | # leak a need to use the db in a multinode scenario. | 
|  | if is_service_enabled n-api n-cond n-sched; then | 
|  | # If we're in multi-tier cells mode, we want our control services pointing | 
|  | # at cell0 instead of cell1 to ensure isolation. If not, we point everything | 
|  | # at the main database like normal. | 
|  | if [[ "$CELLSV2_SETUP" == "singleconductor" ]]; then | 
|  | local db="nova_cell1" | 
|  | else | 
|  | local db="nova_cell0" | 
|  | # When in superconductor mode, nova-compute can't send instance | 
|  | # info updates to the scheduler, so just disable it. | 
|  | iniset $NOVA_CONF filter_scheduler track_instance_changes False | 
|  | fi | 
|  |  | 
|  | iniset $NOVA_CONF database connection `database_connection_url $db` | 
|  | iniset $NOVA_CONF api_database connection `database_connection_url nova_api` | 
|  |  | 
|  | # Cache related settings | 
|  | # Those settings aren't really needed in n-cpu thus it is configured | 
|  | # only on nodes which runs controller services | 
|  | iniset $NOVA_CONF cache enabled $NOVA_ENABLE_CACHE | 
|  | iniset $NOVA_CONF cache backend $CACHE_BACKEND | 
|  | iniset $NOVA_CONF cache memcache_servers $MEMCACHE_SERVERS | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-api; then | 
|  | if is_service_enabled n-api-meta; then | 
|  | # If running n-api-meta as a separate service | 
|  | NOVA_ENABLED_APIS=$(echo $NOVA_ENABLED_APIS | sed "s/,metadata//") | 
|  | fi | 
|  | iniset $NOVA_CONF DEFAULT enabled_apis "$NOVA_ENABLED_APIS" | 
|  | if [[ "$NOVA_ENFORCE_SCOPE" == "True" || "$ENFORCE_SCOPE" == "True" ]]; then | 
|  | iniset $NOVA_CONF oslo_policy enforce_new_defaults True | 
|  | iniset $NOVA_CONF oslo_policy enforce_scope True | 
|  | else | 
|  | iniset $NOVA_CONF oslo_policy enforce_new_defaults False | 
|  | iniset $NOVA_CONF oslo_policy enforce_scope False | 
|  | fi | 
|  |  | 
|  | configure_keystone_authtoken_middleware $NOVA_CONF nova | 
|  | fi | 
|  |  | 
|  | if is_service_enabled cinder; then | 
|  | configure_cinder_access | 
|  | fi | 
|  |  | 
|  | if [ -n "$NOVA_STATE_PATH" ]; then | 
|  | iniset $NOVA_CONF DEFAULT state_path "$NOVA_STATE_PATH" | 
|  | iniset $NOVA_CONF oslo_concurrency lock_path "$NOVA_STATE_PATH" | 
|  | fi | 
|  | if [ -n "$NOVA_INSTANCES_PATH" ]; then | 
|  | iniset $NOVA_CONF DEFAULT instances_path "$NOVA_INSTANCES_PATH" | 
|  | fi | 
|  | if [ "$SYSLOG" != "False" ]; then | 
|  | iniset $NOVA_CONF DEFAULT use_syslog "True" | 
|  | fi | 
|  | if [ "$FORCE_CONFIG_DRIVE" != "False" ]; then | 
|  | iniset $NOVA_CONF DEFAULT force_config_drive "$FORCE_CONFIG_DRIVE" | 
|  | fi | 
|  |  | 
|  | # nova defaults to genisoimage but only mkisofs is available for 15.0+ | 
|  | # rhel provides mkisofs symlink to genisoimage or xorriso appropiately | 
|  | if is_fedora; then | 
|  | iniset $NOVA_CONF DEFAULT mkisofs_cmd /usr/bin/mkisofs | 
|  | fi | 
|  |  | 
|  | # Format logging | 
|  | setup_logging $NOVA_CONF | 
|  |  | 
|  | iniset $NOVA_CONF upgrade_levels compute "auto" | 
|  |  | 
|  | if is_service_enabled n-api; then | 
|  | write_uwsgi_config "$NOVA_UWSGI_CONF" "$NOVA_UWSGI" "/compute" "" "nova-api" | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-api-meta; then | 
|  | write_uwsgi_config "$NOVA_METADATA_UWSGI_CONF" "$NOVA_METADATA_UWSGI" "" "$SERVICE_LISTEN_ADDRESS:${METADATA_SERVICE_PORT}" "nova-metadata" | 
|  | fi | 
|  |  | 
|  | if is_service_enabled ceilometer; then | 
|  | iniset $NOVA_CONF DEFAULT instance_usage_audit "True" | 
|  | iniset $NOVA_CONF DEFAULT instance_usage_audit_period "hour" | 
|  | iniset $NOVA_CONF DEFAULT notify_on_state_change "vm_and_task_state" | 
|  | fi | 
|  |  | 
|  | # Set the oslo messaging driver to the typical default. This does not | 
|  | # enable notifications, but it will allow them to function when enabled. | 
|  | iniset $NOVA_CONF oslo_messaging_notifications driver "messagingv2" | 
|  | iniset $NOVA_CONF oslo_messaging_notifications transport_url $(get_notification_url) | 
|  | iniset $NOVA_CONF notifications notification_format "$NOVA_NOTIFICATION_FORMAT" | 
|  | iniset_rpc_backend nova $NOVA_CONF | 
|  |  | 
|  | iniset $NOVA_CONF DEFAULT osapi_compute_workers "$API_WORKERS" | 
|  | iniset $NOVA_CONF DEFAULT metadata_workers "$API_WORKERS" | 
|  | # don't let the conductor get out of control now that we're using a pure python db driver | 
|  | iniset $NOVA_CONF conductor workers "$API_WORKERS" | 
|  |  | 
|  | if is_service_enabled tls-proxy; then | 
|  | iniset $NOVA_CONF DEFAULT glance_protocol https | 
|  | iniset $NOVA_CONF oslo_middleware enable_proxy_headers_parsing True | 
|  | fi | 
|  |  | 
|  | iniset $NOVA_CONF DEFAULT graceful_shutdown_timeout "$SERVICE_GRACEFUL_SHUTDOWN_TIMEOUT" | 
|  |  | 
|  | if [ "$NOVA_USE_SERVICE_TOKEN" == "True" ]; then | 
|  | init_nova_service_user_conf | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-cond; then | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | local conf | 
|  | local vhost | 
|  | conf=$(conductor_conf $i) | 
|  | vhost="nova_cell${i}" | 
|  | # clean old conductor conf | 
|  | rm -f $conf | 
|  | iniset $conf database connection `database_connection_url nova_cell${i}` | 
|  | iniset $conf conductor workers "$API_WORKERS" | 
|  | iniset $conf DEFAULT debug "$ENABLE_DEBUG_LOG_LEVEL" | 
|  | # if we have a singleconductor, we don't have per host message queues. | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | iniset_rpc_backend nova $conf DEFAULT | 
|  | else | 
|  | rpc_backend_add_vhost $vhost | 
|  | iniset_rpc_backend nova $conf DEFAULT $vhost | 
|  | # When running in superconductor mode, the cell conductor | 
|  | # must be configured to talk to the placement service for | 
|  | # reschedules to work. | 
|  | if is_service_enabled placement placement-client; then | 
|  | configure_placement_nova_compute $conf | 
|  | fi | 
|  | fi | 
|  | # Format logging | 
|  | setup_logging $conf | 
|  | done | 
|  | fi | 
|  |  | 
|  | # Console proxy configuration has to go after conductor configuration | 
|  | # because the per cell config file nova_cellN.conf is cleared out as part | 
|  | # of conductor configuration. | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | configure_console_proxies | 
|  | else | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | local conf | 
|  | local offset | 
|  | conf=$(conductor_conf $i) | 
|  | offset=$((i - 1)) | 
|  | configure_console_proxies $conf $offset | 
|  | done | 
|  | fi | 
|  | } | 
|  |  | 
|  | # Configure access to placement from a nova service, usually | 
|  | # compute, but sometimes conductor. | 
|  | function configure_placement_nova_compute { | 
|  | # Use the provided config file path or default to $NOVA_CONF. | 
|  | local conf=${1:-$NOVA_CONF} | 
|  | iniset $conf placement auth_type "password" | 
|  | iniset $conf placement auth_url "$KEYSTONE_SERVICE_URI" | 
|  | iniset $conf placement username nova | 
|  | iniset $conf placement password "$SERVICE_PASSWORD" | 
|  | iniset $conf placement user_domain_name "$SERVICE_DOMAIN_NAME" | 
|  | iniset $conf placement project_name "$SERVICE_TENANT_NAME" | 
|  | iniset $conf placement project_domain_name "$SERVICE_DOMAIN_NAME" | 
|  | iniset $conf placement region_name "$REGION_NAME" | 
|  | } | 
|  |  | 
|  | # Configure access to cinder. | 
|  | function configure_cinder_access { | 
|  | iniset $NOVA_CONF cinder os_region_name "$REGION_NAME" | 
|  | iniset $NOVA_CONF cinder auth_type "password" | 
|  | iniset $NOVA_CONF cinder auth_url "$KEYSTONE_SERVICE_URI" | 
|  | # NOTE(mriedem): This looks a bit weird but we use the nova user here | 
|  | # since it has the admin role and the cinder user does not. This is | 
|  | # similar to using the nova user in init_nova_service_user_conf. We need | 
|  | # to use a user with the admin role for background tasks in nova to | 
|  | # be able to GET block-storage API resources owned by another project | 
|  | # since cinder has low-level "is_admin" checks in its DB API. | 
|  | iniset $NOVA_CONF cinder username nova | 
|  | iniset $NOVA_CONF cinder password "$SERVICE_PASSWORD" | 
|  | iniset $NOVA_CONF cinder user_domain_name "$SERVICE_DOMAIN_NAME" | 
|  | iniset $NOVA_CONF cinder project_name "$SERVICE_TENANT_NAME" | 
|  | iniset $NOVA_CONF cinder project_domain_name "$SERVICE_DOMAIN_NAME" | 
|  | if is_service_enabled tls-proxy; then | 
|  | CINDER_SERVICE_HOST=${CINDER_SERVICE_HOST:-$SERVICE_HOST} | 
|  | CINDER_SERVICE_PORT=${CINDER_SERVICE_PORT:-8776} | 
|  | iniset $NOVA_CONF cinder cafile $SSL_BUNDLE_FILE | 
|  | fi | 
|  | } | 
|  |  | 
|  | function configure_console_compute { | 
|  | # If we are running multiple cells (and thus multiple console proxies) on a | 
|  | # single host, we offset the ports to avoid collisions.  We need to | 
|  | # correspondingly configure the console proxy port for nova-compute and we | 
|  | # can use the NOVA_CPU_CELL variable to know which cell we are for | 
|  | # calculating the offset. | 
|  | # Stagger the offset based on the total number of possible console proxies | 
|  | # (novnc, spice, serial) so that their ports will not collide if | 
|  | # all are enabled. | 
|  | local offset | 
|  | offset=$(((NOVA_CPU_CELL - 1) * 3)) | 
|  |  | 
|  | # Use the host IP instead of the service host because for multi-node, the | 
|  | # service host will be the controller only. | 
|  | local default_proxyclient_addr | 
|  | default_proxyclient_addr=$(iniget $NOVA_CPU_CONF DEFAULT my_ip) | 
|  |  | 
|  | # All nova-compute workers need to know the vnc configuration options | 
|  | # These settings don't hurt anything if n-novnc is disabled | 
|  | if is_service_enabled n-cpu; then | 
|  | if [ "$NOVNC_FROM_PACKAGE" == "True" ]; then | 
|  | # Use the old URL when installing novnc packages. | 
|  | NOVNCPROXY_URL=${NOVNCPROXY_URL:-"http://$SERVICE_HOST:$((6080 + offset))/vnc_auto.html"} | 
|  | elif vercmp ${NOVNC_BRANCH} "<" "1.0.0"; then | 
|  | # Use the old URL when installing older novnc source. | 
|  | NOVNCPROXY_URL=${NOVNCPROXY_URL:-"http://$SERVICE_HOST:$((6080 + offset))/vnc_auto.html"} | 
|  | else | 
|  | # Use the new URL when building >=v1.0.0 from source. | 
|  | NOVNCPROXY_URL=${NOVNCPROXY_URL:-"http://$SERVICE_HOST:$((6080 + offset))/vnc_lite.html"} | 
|  | fi | 
|  | iniset $NOVA_CPU_CONF vnc novncproxy_base_url "$NOVNCPROXY_URL" | 
|  | SPICEHTML5PROXY_URL=${SPICEHTML5PROXY_URL:-"http://$SERVICE_HOST:$((6081 + offset))/spice_auto.html"} | 
|  | iniset $NOVA_CPU_CONF spice html5proxy_base_url "$SPICEHTML5PROXY_URL" | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-novnc || [ "$NOVA_VNC_ENABLED" != False ]; then | 
|  | # Address on which instance vncservers will listen on compute hosts. | 
|  | # For multi-host, this should be the management ip of the compute host. | 
|  | VNCSERVER_LISTEN=${VNCSERVER_LISTEN:-$NOVA_SERVICE_LISTEN_ADDRESS} | 
|  | VNCSERVER_PROXYCLIENT_ADDRESS=${VNCSERVER_PROXYCLIENT_ADDRESS:-$default_proxyclient_addr} | 
|  | iniset $NOVA_CPU_CONF vnc server_listen "$VNCSERVER_LISTEN" | 
|  | iniset $NOVA_CPU_CONF vnc server_proxyclient_address "$VNCSERVER_PROXYCLIENT_ADDRESS" | 
|  | else | 
|  | iniset $NOVA_CPU_CONF vnc enabled false | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-spice; then | 
|  | # Address on which instance spiceservers will listen on compute hosts. | 
|  | # For multi-host, this should be the management ip of the compute host. | 
|  | SPICESERVER_PROXYCLIENT_ADDRESS=${SPICESERVER_PROXYCLIENT_ADDRESS:-$default_proxyclient_addr} | 
|  | SPICESERVER_LISTEN=${SPICESERVER_LISTEN:-$NOVA_SERVICE_LISTEN_ADDRESS} | 
|  | iniset $NOVA_CPU_CONF spice enabled true | 
|  | iniset $NOVA_CPU_CONF spice server_listen "$SPICESERVER_LISTEN" | 
|  | iniset $NOVA_CPU_CONF spice server_proxyclient_address "$SPICESERVER_PROXYCLIENT_ADDRESS" | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-sproxy; then | 
|  | iniset $NOVA_CPU_CONF serial_console enabled True | 
|  | iniset $NOVA_CPU_CONF serial_console base_url "ws://$SERVICE_HOST:$((6082 + offset))/" | 
|  | fi | 
|  | } | 
|  |  | 
|  | function configure_console_proxies { | 
|  | # Use the provided config file path or default to $NOVA_CONF. | 
|  | local conf=${1:-$NOVA_CONF} | 
|  | local offset=${2:-0} | 
|  | # Stagger the offset based on the total number of possible console proxies | 
|  | # (novnc, spice, serial) so that their ports will not collide if | 
|  | # all are enabled. | 
|  | offset=$((offset * 3)) | 
|  |  | 
|  | if is_service_enabled n-novnc || [ "$NOVA_VNC_ENABLED" != False ]; then | 
|  | iniset $conf vnc novncproxy_host "$NOVA_SERVICE_LISTEN_ADDRESS" | 
|  | iniset $conf vnc novncproxy_port $((6080 + offset)) | 
|  |  | 
|  | if is_nova_console_proxy_compute_tls_enabled ; then | 
|  | iniset $conf vnc auth_schemes "vencrypt" | 
|  | iniset $conf vnc vencrypt_client_key "/etc/pki/nova-novnc/client-key.pem" | 
|  | iniset $conf vnc vencrypt_client_cert "/etc/pki/nova-novnc/client-cert.pem" | 
|  | iniset $conf vnc vencrypt_ca_certs "/etc/pki/nova-novnc/ca-cert.pem" | 
|  |  | 
|  | sudo mkdir -p /etc/pki/nova-novnc | 
|  | deploy_int_CA /etc/pki/nova-novnc/ca-cert.pem | 
|  | deploy_int_cert /etc/pki/nova-novnc/client-cert.pem /etc/pki/nova-novnc/client-key.pem | 
|  | # OpenSSL 1.1.0 generates the key file with permissions: 600, by | 
|  | # default, and the deploy_int* methods use 'sudo cp' to copy the | 
|  | # files, making them owned by root:root. | 
|  | # Change ownership of everything under /etc/pki/nova-novnc to | 
|  | # $STACK_USER:$(id -g ${STACK_USER}) so that $STACK_USER can read | 
|  | # the key file. | 
|  | sudo chown -R $STACK_USER:$(id -g ${STACK_USER}) /etc/pki/nova-novnc | 
|  | # This is needed to enable TLS in the proxy itself, example log: | 
|  | # WebSocket server settings: | 
|  | #   - Listen on 0.0.0.0:6080 | 
|  | #   - Flash security policy server | 
|  | #   - Web server (no directory listings). Web root: /usr/share/novnc | 
|  | #   - SSL/TLS support | 
|  | #   - proxying from 0.0.0.0:6080 to None:None | 
|  | iniset $conf DEFAULT key "/etc/pki/nova-novnc/client-key.pem" | 
|  | iniset $conf DEFAULT cert "/etc/pki/nova-novnc/client-cert.pem" | 
|  | fi | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-spice; then | 
|  | iniset $conf spice html5proxy_host "$NOVA_SERVICE_LISTEN_ADDRESS" | 
|  | iniset $conf spice html5proxy_port $((6081 + offset)) | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-sproxy; then | 
|  | iniset $conf serial_console serialproxy_host "$NOVA_SERVICE_LISTEN_ADDRESS" | 
|  | iniset $conf serial_console serialproxy_port $((6082 + offset)) | 
|  | fi | 
|  | } | 
|  |  | 
|  | function configure_nova_unified_limits { | 
|  | # Registered limit resources in keystone are system-specific resources. | 
|  | # Make sure we use a system-scoped token to interact with this API. | 
|  |  | 
|  | # Default limits here mirror the legacy config-based default values. | 
|  | # Note: disk quota is new in nova as of unified limits. | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 10 --region $REGION_NAME servers | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 20 --region $REGION_NAME class:VCPU | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit $((50 * 1024)) --region $REGION_NAME class:MEMORY_MB | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 20 --region $REGION_NAME class:DISK_GB | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 128 --region $REGION_NAME server_metadata_items | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 5 --region $REGION_NAME server_injected_files | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 10240 --region $REGION_NAME server_injected_file_content_bytes | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 255 --region $REGION_NAME server_injected_file_path_bytes | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 100 --region $REGION_NAME server_key_pairs | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 10 --region $REGION_NAME server_groups | 
|  | openstack --os-cloud devstack-system-admin registered limit create \ | 
|  | --service nova --default-limit 10 --region $REGION_NAME server_group_members | 
|  |  | 
|  | # Tell nova to use these limits | 
|  | iniset $NOVA_CONF quota driver "nova.quota.UnifiedLimitsDriver" | 
|  |  | 
|  | # Configure oslo_limit so it can talk to keystone | 
|  | iniset $NOVA_CONF oslo_limit user_domain_name $SERVICE_DOMAIN_NAME | 
|  | iniset $NOVA_CONF oslo_limit password $SERVICE_PASSWORD | 
|  | iniset $NOVA_CONF oslo_limit username nova | 
|  | iniset $NOVA_CONF oslo_limit auth_type password | 
|  | iniset $NOVA_CONF oslo_limit auth_url $KEYSTONE_SERVICE_URI | 
|  | iniset $NOVA_CONF oslo_limit system_scope all | 
|  | iniset $NOVA_CONF oslo_limit endpoint_id \ | 
|  | $(openstack endpoint list --service nova -f value -c ID) | 
|  |  | 
|  | # Allow the nova service user to read quotas | 
|  | openstack --os-cloud devstack-system-admin role add --user nova \ | 
|  | --user-domain $SERVICE_DOMAIN_NAME --system all reader | 
|  | } | 
|  |  | 
|  | function init_nova_service_user_conf { | 
|  | iniset $NOVA_CONF service_user send_service_user_token True | 
|  | iniset $NOVA_CONF service_user auth_type password | 
|  | iniset $NOVA_CONF service_user auth_url "$KEYSTONE_SERVICE_URI" | 
|  | iniset $NOVA_CONF service_user username nova | 
|  | iniset $NOVA_CONF service_user password "$SERVICE_PASSWORD" | 
|  | iniset $NOVA_CONF service_user user_domain_name "$SERVICE_DOMAIN_NAME" | 
|  | iniset $NOVA_CONF service_user project_name "$SERVICE_PROJECT_NAME" | 
|  | iniset $NOVA_CONF service_user project_domain_name "$SERVICE_DOMAIN_NAME" | 
|  | iniset $NOVA_CONF service_user auth_strategy keystone | 
|  | } | 
|  |  | 
|  | function conductor_conf { | 
|  | local cell="$1" | 
|  | echo "${NOVA_CONF_DIR}/nova_cell${cell}.conf" | 
|  | } | 
|  |  | 
|  | # create_nova_keys_dir() - Part of the init_nova() process | 
|  | function create_nova_keys_dir { | 
|  | # Create keys dir | 
|  | sudo install -d -o $STACK_USER ${NOVA_STATE_PATH} ${NOVA_STATE_PATH}/keys | 
|  | } | 
|  |  | 
|  | function init_nova_db { | 
|  | local dbname="$1" | 
|  | local conffile="$2" | 
|  | recreate_database $dbname | 
|  | $NOVA_BIN_DIR/nova-manage --config-file $conffile db sync --local_cell | 
|  | } | 
|  |  | 
|  | # init_nova() - Initialize databases, etc. | 
|  | function init_nova { | 
|  | # All nova components talk to a central database. | 
|  | # Only do this step once on the API node for an entire cluster. | 
|  | if is_service_enabled $DATABASE_BACKENDS && is_service_enabled n-api; then | 
|  | # (Re)create nova databases | 
|  | if [[ "$CELLSV2_SETUP" == "singleconductor" ]]; then | 
|  | # If we are doing singleconductor mode, we have some strange | 
|  | # interdependencies. in that the main config refers to cell1 | 
|  | # instead of cell0. In that case, just make sure the cell0 database | 
|  | # is created before we need it below, but don't db_sync it until | 
|  | # after the cellN databases are there. | 
|  | recreate_database nova_cell0 | 
|  | else | 
|  | async_run nova-cell-0 init_nova_db nova_cell0 $NOVA_CONF | 
|  | fi | 
|  |  | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | async_run nova-cell-$i init_nova_db nova_cell${i} $(conductor_conf $i) | 
|  | done | 
|  |  | 
|  | recreate_database $NOVA_API_DB | 
|  | $NOVA_BIN_DIR/nova-manage --config-file $NOVA_CONF api_db sync | 
|  |  | 
|  | # map_cell0 will create the cell mapping record in the nova_api DB so | 
|  | # this needs to come after the api_db sync happens. | 
|  | $NOVA_BIN_DIR/nova-manage cell_v2 map_cell0 --database_connection `database_connection_url nova_cell0` | 
|  |  | 
|  | # Wait for DBs to finish from above | 
|  | for i in $(seq 0 $NOVA_NUM_CELLS); do | 
|  | async_wait nova-cell-$i | 
|  | done | 
|  |  | 
|  | if [[ "$CELLSV2_SETUP" == "singleconductor" ]]; then | 
|  | # We didn't db sync cell0 above, so run it now | 
|  | $NOVA_BIN_DIR/nova-manage --config-file $NOVA_CONF db sync | 
|  | fi | 
|  |  | 
|  | # Run online migrations on the new databases | 
|  | # Needed for flavor conversion | 
|  | $NOVA_BIN_DIR/nova-manage --config-file $NOVA_CONF db online_data_migrations | 
|  |  | 
|  | # create the cell1 cell for the main nova db where the hosts live | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | $NOVA_BIN_DIR/nova-manage --config-file $NOVA_CONF --config-file $(conductor_conf $i) cell_v2 create_cell --name "cell$i" | 
|  | done | 
|  | fi | 
|  |  | 
|  | create_nova_keys_dir | 
|  |  | 
|  | if [[ "$NOVA_BACKEND" == "LVM" ]]; then | 
|  | init_default_lvm_volume_group | 
|  | fi | 
|  | } | 
|  |  | 
|  | # install_novaclient() - Collect source and prepare | 
|  | function install_novaclient { | 
|  | if use_library_from_git "python-novaclient"; then | 
|  | git_clone_by_name "python-novaclient" | 
|  | setup_dev_lib "python-novaclient" | 
|  | sudo install -D -m 0644 -o $STACK_USER {${GITDIR["python-novaclient"]}/tools/,/etc/bash_completion.d/}nova.bash_completion | 
|  | fi | 
|  | } | 
|  |  | 
|  | # install_nova() - Collect source and prepare | 
|  | function install_nova { | 
|  |  | 
|  | # Install os-vif | 
|  | if use_library_from_git "os-vif"; then | 
|  | git_clone_by_name "os-vif" | 
|  | setup_dev_lib "os-vif" | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-cpu && [[ -r $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER ]]; then | 
|  | install_nova_hypervisor | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-novnc; then | 
|  | # a websockets/html5 or flash powered VNC console for vm instances | 
|  | NOVNC_FROM_PACKAGE=$(trueorfalse False NOVNC_FROM_PACKAGE) | 
|  | if [ "$NOVNC_FROM_PACKAGE" = "True" ]; then | 
|  | # Installing novnc on Debian bullseye breaks the global pip | 
|  | # install. This happens because novnc pulls in distro cryptography | 
|  | # which will be prefered by distro pip, but if anything has | 
|  | # installed pyOpenSSL from pypi (keystone) that is not compatible | 
|  | # with distro cryptography. Fix this by installing | 
|  | # python3-openssl (pyOpenSSL) from the distro which pip will prefer | 
|  | # on Debian. Ubuntu has inverse problems so we only do this for | 
|  | # Debian. | 
|  | local novnc_packages | 
|  | novnc_packages="novnc" | 
|  | GetOSVersion | 
|  | if [[ "$os_VENDOR" = "Debian" ]] ; then | 
|  | novnc_packages="$novnc_packages python3-openssl" | 
|  | fi | 
|  |  | 
|  | NOVNC_WEB_DIR=/usr/share/novnc | 
|  | install_package $novnc_packages | 
|  | else | 
|  | NOVNC_WEB_DIR=$DEST/novnc | 
|  | git_clone $NOVNC_REPO $NOVNC_WEB_DIR $NOVNC_BRANCH | 
|  | fi | 
|  | fi | 
|  |  | 
|  | if is_service_enabled n-spice; then | 
|  | # a websockets/html5 or flash powered SPICE console for vm instances | 
|  | SPICE_FROM_PACKAGE=$(trueorfalse True SPICE_FROM_PACKAGE) | 
|  | if [ "$SPICE_FROM_PACKAGE" = "True" ]; then | 
|  | SPICE_WEB_DIR=/usr/share/spice-html5 | 
|  | install_package spice-html5 | 
|  | else | 
|  | SPICE_WEB_DIR=$DEST/spice-html5 | 
|  | git_clone $SPICE_REPO $SPICE_WEB_DIR $SPICE_BRANCH | 
|  | fi | 
|  | fi | 
|  |  | 
|  | git_clone $NOVA_REPO $NOVA_DIR $NOVA_BRANCH | 
|  | setup_develop $NOVA_DIR | 
|  | sudo install -D -m 0644 -o $STACK_USER {$NOVA_DIR/tools/,/etc/bash_completion.d/}nova-manage.bash_completion | 
|  | } | 
|  |  | 
|  | # start_nova_api() - Start the API process ahead of other things | 
|  | function start_nova_api { | 
|  | # Get right service port for testing | 
|  | local service_port=$NOVA_SERVICE_PORT | 
|  | local service_protocol=$NOVA_SERVICE_PROTOCOL | 
|  | local nova_url | 
|  | if is_service_enabled tls-proxy; then | 
|  | service_port=$NOVA_SERVICE_PORT_INT | 
|  | service_protocol="http" | 
|  | fi | 
|  |  | 
|  | # Hack to set the path for rootwrap | 
|  | local old_path=$PATH | 
|  | export PATH=$NOVA_BIN_DIR:$PATH | 
|  |  | 
|  | run_process "n-api" "$(which uwsgi) --procname-prefix nova-api --ini $NOVA_UWSGI_CONF" | 
|  | nova_url=$service_protocol://$SERVICE_HOST/compute/v2.1/ | 
|  |  | 
|  | echo "Waiting for nova-api to start..." | 
|  | if ! wait_for_service $SERVICE_TIMEOUT $nova_url; then | 
|  | die $LINENO "nova-api did not start" | 
|  | fi | 
|  |  | 
|  | export PATH=$old_path | 
|  | } | 
|  |  | 
|  |  | 
|  | # start_nova_compute() - Start the compute process | 
|  | function start_nova_compute { | 
|  | # Hack to set the path for rootwrap | 
|  | local old_path=$PATH | 
|  | export PATH=$NOVA_BIN_DIR:$PATH | 
|  |  | 
|  | local compute_cell_conf=$NOVA_CONF | 
|  |  | 
|  | # Bug #1802143: $NOVA_CPU_CONF is constructed by first copying $NOVA_CONF... | 
|  | cp $compute_cell_conf $NOVA_CPU_CONF | 
|  | # ...and then adding/overriding anything explicitly set in $NOVA_CPU_CONF | 
|  | merge_config_file $TOP_DIR/local.conf post-config '$NOVA_CPU_CONF' | 
|  |  | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | # NOTE(danms): Grenade doesn't setup multi-cell rabbit, so | 
|  | # skip these bits and use the normal config. | 
|  | echo "Skipping multi-cell conductor fleet setup" | 
|  | else | 
|  | # "${CELLSV2_SETUP}" is "superconductor" | 
|  | # FIXME(danms): Should this be configurable? | 
|  | iniset $NOVA_CPU_CONF workarounds disable_group_policy_check_upcall True | 
|  | # Since the nova-compute service cannot reach nova-scheduler over | 
|  | # RPC, we also disable track_instance_changes. | 
|  | iniset $NOVA_CPU_CONF filter_scheduler track_instance_changes False | 
|  | iniset_rpc_backend nova $NOVA_CPU_CONF DEFAULT "nova_cell${NOVA_CPU_CELL}" | 
|  | fi | 
|  |  | 
|  | # Make sure we nuke any database config | 
|  | inidelete $NOVA_CPU_CONF database connection | 
|  | inidelete $NOVA_CPU_CONF api_database connection | 
|  |  | 
|  | # Console proxies were configured earlier in create_nova_conf. Now that the | 
|  | # nova-cpu.conf has been created, configure the console settings required | 
|  | # by the compute process. | 
|  | configure_console_compute | 
|  |  | 
|  | # Set rebuild timeout longer for BFV instances because we likely have | 
|  | # slower disk than expected. Default is 20s/GB | 
|  | iniset $NOVA_CPU_CONF DEFAULT reimage_timeout_per_gb 180 | 
|  |  | 
|  | # Configure the OVSDB connection for os-vif | 
|  | if [ -n "$OVSDB_SERVER_LOCAL_HOST" ]; then | 
|  | iniset $NOVA_CPU_CONF os_vif_ovs ovsdb_connection "tcp:$OVSDB_SERVER_LOCAL_HOST:6640" | 
|  | fi | 
|  |  | 
|  | # Workaround bug #1939108 | 
|  | if [[ "$VIRT_DRIVER" == "libvirt" && "$LIBVIRT_TYPE" == "qemu" ]]; then | 
|  | iniset $NOVA_CPU_CONF workarounds libvirt_disable_apic True | 
|  | fi | 
|  |  | 
|  | if [[ "$NOVA_CPU_UUID" ]]; then | 
|  | echo -n $NOVA_CPU_UUID > $NOVA_CONF_DIR/compute_id | 
|  | fi | 
|  |  | 
|  | if [[ "$VIRT_DRIVER" = 'libvirt' ]]; then | 
|  | if [ ${NOVA_LIBVIRT_TB_CACHE_SIZE} -gt 0 ]; then | 
|  | iniset $NOVA_CPU_CONF libvirt tb_cache_size ${NOVA_LIBVIRT_TB_CACHE_SIZE} | 
|  | fi | 
|  | # The group **$LIBVIRT_GROUP** is added to the current user in this script. | 
|  | # ``sg`` is used in run_process to execute nova-compute as a member of the | 
|  | # **$LIBVIRT_GROUP** group. | 
|  | run_process n-cpu "$NOVA_BIN_DIR/nova-compute --config-file $NOVA_CPU_CONF" $LIBVIRT_GROUP | 
|  | elif [[ "$VIRT_DRIVER" = 'lxd' ]]; then | 
|  | run_process n-cpu "$NOVA_BIN_DIR/nova-compute --config-file $NOVA_CPU_CONF" $LXD_GROUP | 
|  | elif [[ "$VIRT_DRIVER" = 'docker' || "$VIRT_DRIVER" = 'zun' ]]; then | 
|  | run_process n-cpu "$NOVA_BIN_DIR/nova-compute --config-file $NOVA_CPU_CONF" $DOCKER_GROUP | 
|  | elif [[ "$VIRT_DRIVER" = 'fake' ]]; then | 
|  | local i | 
|  | for i in `seq 1 $NUMBER_FAKE_NOVA_COMPUTE`; do | 
|  | # Avoid process redirection of fake host configurations by | 
|  | # creating or modifying real configurations. Each fake | 
|  | # gets its own configuration and own log file. | 
|  | local fake_conf="${NOVA_FAKE_CONF}-${i}" | 
|  | iniset $fake_conf DEFAULT host "${HOSTNAME}${i}" | 
|  | run_process "n-cpu-${i}" "$NOVA_BIN_DIR/nova-compute --config-file $NOVA_CPU_CONF --config-file $fake_conf" | 
|  | done | 
|  | else | 
|  | if is_service_enabled n-cpu && [[ -r $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER ]]; then | 
|  | start_nova_hypervisor | 
|  | fi | 
|  | run_process n-cpu "$NOVA_BIN_DIR/nova-compute --config-file $NOVA_CPU_CONF" | 
|  | fi | 
|  |  | 
|  | export PATH=$old_path | 
|  | } | 
|  |  | 
|  | # start_nova() - Start running processes | 
|  | function start_nova_rest { | 
|  | # Hack to set the path for rootwrap | 
|  | local old_path=$PATH | 
|  | export PATH=$NOVA_BIN_DIR:$PATH | 
|  |  | 
|  | local compute_cell_conf=$NOVA_CONF | 
|  |  | 
|  | run_process n-sch "$NOVA_BIN_DIR/nova-scheduler --config-file $compute_cell_conf" | 
|  | run_process n-api-meta "$(which uwsgi) --procname-prefix nova-api-meta --ini $NOVA_METADATA_UWSGI_CONF" | 
|  |  | 
|  | export PATH=$old_path | 
|  | } | 
|  |  | 
|  | function enable_nova_console_proxies { | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | for srv in n-novnc n-spice n-sproxy; do | 
|  | if is_service_enabled $srv; then | 
|  | enable_service ${srv}-cell${i} | 
|  | fi | 
|  | done | 
|  | done | 
|  | } | 
|  |  | 
|  | function start_nova_console_proxies { | 
|  | # Hack to set the path for rootwrap | 
|  | local old_path=$PATH | 
|  | # This is needed to find the nova conf | 
|  | export PATH=$NOVA_BIN_DIR:$PATH | 
|  |  | 
|  | local api_cell_conf=$NOVA_CONF | 
|  | # console proxies run globally for singleconductor, else they run per cell | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | run_process n-novnc "$NOVA_BIN_DIR/nova-novncproxy --config-file $api_cell_conf --web $NOVNC_WEB_DIR" | 
|  | run_process n-spice "$NOVA_BIN_DIR/nova-spicehtml5proxy --config-file $api_cell_conf --web $SPICE_WEB_DIR" | 
|  | run_process n-sproxy "$NOVA_BIN_DIR/nova-serialproxy --config-file $api_cell_conf" | 
|  | else | 
|  | enable_nova_console_proxies | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | local conf | 
|  | conf=$(conductor_conf $i) | 
|  | run_process n-novnc-cell${i} "$NOVA_BIN_DIR/nova-novncproxy --config-file $conf --web $NOVNC_WEB_DIR" | 
|  | run_process n-spice-cell${i} "$NOVA_BIN_DIR/nova-spicehtml5proxy --config-file $conf --web $SPICE_WEB_DIR" | 
|  | run_process n-sproxy-cell${i} "$NOVA_BIN_DIR/nova-serialproxy --config-file $conf" | 
|  | done | 
|  | fi | 
|  |  | 
|  | export PATH=$old_path | 
|  | } | 
|  |  | 
|  | function enable_nova_fleet { | 
|  | if is_service_enabled n-cond; then | 
|  | enable_service n-super-cond | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | enable_service n-cond-cell${i} | 
|  | done | 
|  | fi | 
|  | } | 
|  |  | 
|  | function start_nova_conductor { | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | echo "Starting nova-conductor in a cellsv1-compatible way" | 
|  | run_process n-cond "$NOVA_BIN_DIR/nova-conductor --config-file $NOVA_COND_CONF" | 
|  | return | 
|  | fi | 
|  |  | 
|  | enable_nova_fleet | 
|  | if is_service_enabled n-super-cond; then | 
|  | run_process n-super-cond "$NOVA_BIN_DIR/nova-conductor --config-file $NOVA_COND_CONF" | 
|  | fi | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | if is_service_enabled n-cond-cell${i}; then | 
|  | local conf | 
|  | conf=$(conductor_conf $i) | 
|  | run_process n-cond-cell${i} "$NOVA_BIN_DIR/nova-conductor --config-file $conf" | 
|  | fi | 
|  | done | 
|  | } | 
|  |  | 
|  | function is_nova_ready { | 
|  | # NOTE(sdague): with cells v2 all the compute services must be up | 
|  | # and checked into the database before discover_hosts is run. This | 
|  | # happens in all in one installs by accident, because > 30 seconds | 
|  | # happen between here and the script ending. However, in multinode | 
|  | # tests this can very often not be the case. So ensure that the | 
|  | # compute is up before we move on. | 
|  | wait_for_compute $NOVA_READY_TIMEOUT | 
|  | } | 
|  |  | 
|  | function start_nova { | 
|  | start_nova_rest | 
|  | start_nova_console_proxies | 
|  | start_nova_conductor | 
|  | start_nova_compute | 
|  | if is_service_enabled n-api; then | 
|  | # dump the cell mapping to ensure life is good | 
|  | echo "Dumping cells_v2 mapping" | 
|  | $NOVA_BIN_DIR/nova-manage cell_v2 list_cells --verbose | 
|  | fi | 
|  | } | 
|  |  | 
|  | function stop_nova_compute { | 
|  | if [ "$VIRT_DRIVER" == "fake" ]; then | 
|  | local i | 
|  | for i in `seq 1 $NUMBER_FAKE_NOVA_COMPUTE`; do | 
|  | stop_process n-cpu-${i} | 
|  | done | 
|  | else | 
|  | stop_process n-cpu | 
|  | fi | 
|  | if is_service_enabled n-cpu && [[ -r $NOVA_PLUGINS/hypervisor-$VIRT_DRIVER ]]; then | 
|  | stop_nova_hypervisor | 
|  | fi | 
|  | } | 
|  |  | 
|  | function stop_nova_rest { | 
|  | # Kill the non-compute nova processes | 
|  | for serv in n-api n-api-meta n-sch; do | 
|  | stop_process $serv | 
|  | done | 
|  | } | 
|  |  | 
|  | function stop_nova_console_proxies { | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | for srv in n-novnc n-spice n-sproxy; do | 
|  | stop_process $srv | 
|  | done | 
|  | else | 
|  | enable_nova_console_proxies | 
|  | for i in $(seq 1 $NOVA_NUM_CELLS); do | 
|  | for srv in n-novnc n-spice n-sproxy; do | 
|  | stop_process ${srv}-cell${i} | 
|  | done | 
|  | done | 
|  | fi | 
|  | } | 
|  |  | 
|  | function stop_nova_conductor { | 
|  | if [[ "${CELLSV2_SETUP}" == "singleconductor" ]]; then | 
|  | stop_process n-cond | 
|  | return | 
|  | fi | 
|  |  | 
|  | enable_nova_fleet | 
|  | for srv in n-super-cond $(seq -f n-cond-cell%0.f 1 $NOVA_NUM_CELLS); do | 
|  | if is_service_enabled $srv; then | 
|  | stop_process $srv | 
|  | fi | 
|  | done | 
|  | } | 
|  |  | 
|  | # stop_nova() - Stop running processes | 
|  | function stop_nova { | 
|  | stop_nova_rest | 
|  | stop_nova_console_proxies | 
|  | stop_nova_conductor | 
|  | stop_nova_compute | 
|  | } | 
|  |  | 
|  | # create_instance_types(): Create default flavors | 
|  | function create_flavors { | 
|  | if is_service_enabled n-api; then | 
|  | if ! openstack --os-region-name="$REGION_NAME" flavor list | grep -q ds512M; then | 
|  | # Note that danms hates these flavors and apologizes for sdague | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id c1 --ram 256 --disk 1 --vcpus 1 --property hw_rng:allowed=True cirros256 | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id d1 --ram 512 --disk 5 --vcpus 1 --property hw_rng:allowed=True ds512M | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id d2 --ram 1024 --disk 10 --vcpus 1 --property hw_rng:allowed=True ds1G | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id d3 --ram 2048 --disk 10 --vcpus 2 --property hw_rng:allowed=True ds2G | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id d4 --ram 4096 --disk 20 --vcpus 4 --property hw_rng:allowed=True ds4G | 
|  | fi | 
|  |  | 
|  | if ! openstack --os-region-name="$REGION_NAME" flavor list | grep -q m1.tiny; then | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id 1 --ram 512 --disk 1 --vcpus 1 --property hw_rng:allowed=True m1.tiny | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id 2 --ram 2048 --disk 20 --vcpus 1 --property hw_rng:allowed=True m1.small | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id 3 --ram 4096 --disk 40 --vcpus 2 --property hw_rng:allowed=True m1.medium | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id 4 --ram 8192 --disk 80 --vcpus 4 --property hw_rng:allowed=True m1.large | 
|  | openstack --os-region-name="$REGION_NAME" flavor create --id 5 --ram 16384 --disk 160 --vcpus 8 --property hw_rng:allowed=True m1.xlarge | 
|  | fi | 
|  | fi | 
|  | } | 
|  |  | 
|  | # Restore xtrace | 
|  | $_XTRACE_LIB_NOVA | 
|  |  | 
|  | # Tell emacs to use shell-script-mode | 
|  | ## Local variables: | 
|  | ## mode: shell-script | 
|  | ## End: |