| # lib/swift |
| # Functions to control the configuration and operation of the **Swift** service |
| |
| # Dependencies: |
| # ``functions`` file |
| # ``apache`` file |
| # ``DEST``, ``SCREEN_NAME``, `SWIFT_HASH` must be defined |
| # ``STACK_USER`` must be defined |
| # ``SWIFT_DATA_DIR`` or ``DATA_DIR`` must be defined |
| # ``lib/keystone`` file |
| # ``stack.sh`` calls the entry points in this order: |
| # |
| # install_swift |
| # _config_swift_apache_wsgi |
| # configure_swift |
| # init_swift |
| # start_swift |
| # stop_swift |
| # cleanup_swift |
| # _cleanup_swift_apache_wsgi |
| |
| # Save trace setting |
| XTRACE=$(set +o | grep xtrace) |
| set +o xtrace |
| |
| |
| # Defaults |
| # -------- |
| |
| # Set up default directories |
| SWIFT_DIR=$DEST/swift |
| SWIFTCLIENT_DIR=$DEST/python-swiftclient |
| SWIFT_AUTH_CACHE_DIR=${SWIFT_AUTH_CACHE_DIR:-/var/cache/swift} |
| SWIFT_APACHE_WSGI_DIR=${SWIFT_APACHE_WSGI_DIR:-/var/www/swift} |
| SWIFT3_DIR=$DEST/swift3 |
| |
| # TODO: add logging to different location. |
| |
| # Set ``SWIFT_DATA_DIR`` to the location of swift drives and objects. |
| # Default is the common DevStack data directory. |
| SWIFT_DATA_DIR=${SWIFT_DATA_DIR:-${DATA_DIR}/swift} |
| |
| # Set ``SWIFT_CONF_DIR`` to the location of the configuration files. |
| # Default is ``/etc/swift``. |
| # TODO(dtroyer): remove SWIFT_CONFIG_DIR after cutting stable/grizzly |
| SWIFT_CONF_DIR=${SWIFT_CONF_DIR:-${SWIFT_CONFIG_DIR:-/etc/swift}} |
| |
| if is_service_enabled s-proxy && is_service_enabled swift3; then |
| # If we are using swift3, we can default the s3 port to swift instead |
| # of nova-objectstore |
| S3_SERVICE_PORT=${S3_SERVICE_PORT:-8080} |
| fi |
| |
| # DevStack will create a loop-back disk formatted as XFS to store the |
| # swift data. Set ``SWIFT_LOOPBACK_DISK_SIZE`` to the disk size in |
| # kilobytes. |
| # Default is 1 gigabyte. |
| SWIFT_LOOPBACK_DISK_SIZE=${SWIFT_LOOPBACK_DISK_SIZE:-1000000} |
| |
| # Set ``SWIFT_EXTRAS_MIDDLEWARE`` to extras middlewares. |
| # Default is ``staticweb, tempurl, bulk, formpost`` |
| SWIFT_EXTRAS_MIDDLEWARE=${SWIFT_EXTRAS_MIDDLEWARE:-tempurl formpost staticweb bulk} |
| |
| # The ring uses a configurable number of bits from a path’s MD5 hash as |
| # a partition index that designates a device. The number of bits kept |
| # from the hash is known as the partition power, and 2 to the partition |
| # power indicates the partition count. Partitioning the full MD5 hash |
| # ring allows other parts of the cluster to work in batches of items at |
| # once which ends up either more efficient or at least less complex than |
| # working with each item separately or the entire cluster all at once. |
| # By default we define 9 for the partition count (which mean 512). |
| SWIFT_PARTITION_POWER_SIZE=${SWIFT_PARTITION_POWER_SIZE:-9} |
| |
| # Set ``SWIFT_REPLICAS`` to configure how many replicas are to be |
| # configured for your Swift cluster. By default we are configuring |
| # only one replica since this is way less CPU and memory intensive. If |
| # you are planning to test swift replication you may want to set this |
| # up to 3. |
| SWIFT_REPLICAS=${SWIFT_REPLICAS:-1} |
| SWIFT_REPLICAS_SEQ=$(seq ${SWIFT_REPLICAS}) |
| |
| # Set ``OBJECT_PORT_BASE``, ``CONTAINER_PORT_BASE``, ``ACCOUNT_PORT_BASE`` |
| # Port bases used in port number calclution for the service "nodes" |
| # The specified port number will be used, the additinal ports calculated by |
| # base_port + node_num * 10 |
| OBJECT_PORT_BASE=${OBJECT_PORT_BASE:-6013} |
| CONTAINER_PORT_BASE=${CONTAINER_PORT_BASE:-6011} |
| ACCOUNT_PORT_BASE=${ACCOUNT_PORT_BASE:-6012} |
| |
| |
| # Functions |
| # --------- |
| |
| # cleanup_swift() - Remove residual data files |
| function cleanup_swift() { |
| rm -f ${SWIFT_CONF_DIR}{*.builder,*.ring.gz,backups/*.builder,backups/*.ring.gz} |
| if egrep -q ${SWIFT_DATA_DIR}/drives/sdb1 /proc/mounts; then |
| sudo umount ${SWIFT_DATA_DIR}/drives/sdb1 |
| fi |
| if [[ -e ${SWIFT_DATA_DIR}/drives/images/swift.img ]]; then |
| rm ${SWIFT_DATA_DIR}/drives/images/swift.img |
| fi |
| rm -rf ${SWIFT_DATA_DIR}/run/ |
| if is_apache_enabled_service swift; then |
| _cleanup_swift_apache_wsgi |
| fi |
| } |
| |
| # _cleanup_swift_apache_wsgi() - Remove wsgi files, disable and remove apache vhost file |
| function _cleanup_swift_apache_wsgi() { |
| sudo rm -f $SWIFT_APACHE_WSGI_DIR/*.wsgi |
| ! is_fedora && sudo a2dissite proxy-server |
| for node_number in ${SWIFT_REPLICAS_SEQ}; do |
| for type in object container account; do |
| site_name=${type}-server-${node_number} |
| ! is_fedora && sudo a2dissite ${site_name} |
| sudo rm -f /etc/$APACHE_NAME/$APACHE_CONF_DIR/${site_name} |
| done |
| done |
| } |
| |
| # _config_swift_apache_wsgi() - Set WSGI config files of Swift |
| function _config_swift_apache_wsgi() { |
| sudo mkdir -p ${SWIFT_APACHE_WSGI_DIR} |
| local apache_vhost_dir=/etc/${APACHE_NAME}/$APACHE_CONF_DIR |
| local proxy_port=${SWIFT_DEFAULT_BIND_PORT:-8080} |
| |
| # copy proxy vhost and wsgi file |
| sudo cp ${SWIFT_DIR}/examples/apache2/proxy-server.template ${apache_vhost_dir}/proxy-server |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%PORT%/$proxy_port/g; |
| s/%SERVICENAME%/proxy-server/g; |
| s/%APACHE_NAME%/${APACHE_NAME}/g; |
| " -i ${apache_vhost_dir}/proxy-server |
| |
| sudo cp ${SWIFT_DIR}/examples/wsgi/proxy-server.wsgi.template ${SWIFT_APACHE_WSGI_DIR}/proxy-server.wsgi |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%SERVICECONF%/proxy-server.conf/g; |
| " -i ${SWIFT_APACHE_WSGI_DIR}/proxy-server.wsgi |
| ! is_fedora && sudo a2ensite proxy-server |
| |
| # copy apache vhost file and set name and port |
| for node_number in ${SWIFT_REPLICAS_SEQ}; do |
| object_port=$[OBJECT_PORT_BASE + 10 * ($node_number - 1)] |
| container_port=$[CONTAINER_PORT_BASE + 10 * ($node_number - 1)] |
| account_port=$[ACCOUNT_PORT_BASE + 10 * ($node_number - 1)] |
| |
| sudo cp ${SWIFT_DIR}/examples/apache2/object-server.template ${apache_vhost_dir}/object-server-${node_number} |
| sudo sed -e " |
| s/%PORT%/$object_port/g; |
| s/%SERVICENAME%/object-server-${node_number}/g; |
| s/%APACHE_NAME%/${APACHE_NAME}/g; |
| " -i ${apache_vhost_dir}/object-server-${node_number} |
| ! is_fedora && sudo a2ensite object-server-${node_number} |
| |
| sudo cp ${SWIFT_DIR}/examples/wsgi/object-server.wsgi.template ${SWIFT_APACHE_WSGI_DIR}/object-server-${node_number}.wsgi |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%SERVICECONF%/object-server\/${node_number}.conf/g; |
| " -i ${SWIFT_APACHE_WSGI_DIR}/object-server-${node_number}.wsgi |
| |
| sudo cp ${SWIFT_DIR}/examples/apache2/container-server.template ${apache_vhost_dir}/container-server-${node_number} |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%PORT%/$container_port/g; |
| s/%SERVICENAME%/container-server-${node_number}/g; |
| s/%APACHE_NAME%/${APACHE_NAME}/g; |
| " -i ${apache_vhost_dir}/container-server-${node_number} |
| ! is_fedora && sudo a2ensite container-server-${node_number} |
| |
| sudo cp ${SWIFT_DIR}/examples/wsgi/container-server.wsgi.template ${SWIFT_APACHE_WSGI_DIR}/container-server-${node_number}.wsgi |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%SERVICECONF%/container-server\/${node_number}.conf/g; |
| " -i ${SWIFT_APACHE_WSGI_DIR}/container-server-${node_number}.wsgi |
| |
| sudo cp ${SWIFT_DIR}/examples/apache2/account-server.template ${apache_vhost_dir}/account-server-${node_number} |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%PORT%/$account_port/g; |
| s/%SERVICENAME%/account-server-${node_number}/g; |
| s/%APACHE_NAME%/${APACHE_NAME}/g; |
| " -i ${apache_vhost_dir}/account-server-${node_number} |
| ! is_fedora && sudo a2ensite account-server-${node_number} |
| |
| sudo cp ${SWIFT_DIR}/examples/wsgi/account-server.wsgi.template ${SWIFT_APACHE_WSGI_DIR}/account-server-${node_number}.wsgi |
| sudo sed -e " |
| /^#/d;/^$/d; |
| s/%SERVICECONF%/account-server\/${node_number}.conf/g; |
| " -i ${SWIFT_APACHE_WSGI_DIR}/account-server-${node_number}.wsgi |
| |
| done |
| |
| # run apache server as stack user |
| change_apache_user_group ${STACK_USER} |
| |
| # WSGI isn't enabled by default, enable it |
| ! is_fedora && sudo a2enmod wsgi |
| } |
| |
| # configure_swift() - Set config files, create data dirs and loop image |
| function configure_swift() { |
| local swift_pipeline=" " |
| local node_number |
| local swift_node_config |
| local swift_log_dir |
| |
| # Make sure to kill all swift processes first |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run all stop || true |
| |
| sudo mkdir -p ${SWIFT_CONF_DIR}/{object,container,account}-server |
| sudo chown -R $USER: ${SWIFT_CONF_DIR} |
| |
| if [[ "$SWIFT_CONF_DIR" != "/etc/swift" ]]; then |
| # Some swift tools are hard-coded to use ``/etc/swift`` and are apparently not going to be fixed. |
| # Create a symlink if the config dir is moved |
| sudo ln -sf ${SWIFT_CONF_DIR} /etc/swift |
| fi |
| |
| # Swift use rsync to synchronize between all the different |
| # partitions (which make more sense when you have a multi-node |
| # setup) we configure it with our version of rsync. |
| sed -e " |
| s/%GROUP%/${USER_GROUP}/; |
| s/%USER%/$USER/; |
| s,%SWIFT_DATA_DIR%,$SWIFT_DATA_DIR,; |
| " $FILES/swift/rsyncd.conf | sudo tee /etc/rsyncd.conf |
| # rsyncd.conf just prepared for 4 nodes |
| if is_ubuntu; then |
| sudo sed -i '/^RSYNC_ENABLE=false/ { s/false/true/ }' /etc/default/rsync |
| else |
| sudo sed -i '/disable *= *yes/ { s/yes/no/ }' /etc/xinetd.d/rsync |
| fi |
| |
| SWIFT_CONFIG_PROXY_SERVER=${SWIFT_CONF_DIR}/proxy-server.conf |
| cp ${SWIFT_DIR}/etc/proxy-server.conf-sample ${SWIFT_CONFIG_PROXY_SERVER} |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT user |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT user ${USER} |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT swift_dir |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT swift_dir ${SWIFT_CONF_DIR} |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT workers |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT workers 1 |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT log_level |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT log_level DEBUG |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT bind_port |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} DEFAULT bind_port ${SWIFT_DEFAULT_BIND_PORT:-8080} |
| |
| # By default Swift will be installed with keystone and tempauth middleware |
| # and add the swift3 middleware if its configured for it. The token for |
| # tempauth would be prefixed with the reseller_prefix setting TEMPAUTH_ the |
| # token for keystoneauth would have the standard reseller_prefix AUTH_ |
| if is_service_enabled swift3;then |
| swift_pipeline=" swift3 s3token " |
| fi |
| swift_pipeline+=" authtoken keystoneauth tempauth " |
| sed -i "/^pipeline/ { s/tempauth/${swift_pipeline} ${SWIFT_EXTRAS_MIDDLEWARE}/ ;}" ${SWIFT_CONFIG_PROXY_SERVER} |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} filter:tempauth account_autocreate |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} app:proxy-server account_autocreate true |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} filter:tempauth reseller_prefix |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:tempauth reseller_prefix "TEMPAUTH" |
| |
| # Configure Keystone |
| sed -i '/^# \[filter:authtoken\]/,/^# \[filter:keystoneauth\]$/ s/^#[ \t]*//' ${SWIFT_CONFIG_PROXY_SERVER} |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_host $KEYSTONE_AUTH_HOST |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_port $KEYSTONE_AUTH_PORT |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_protocol $KEYSTONE_AUTH_PROTOCOL |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken auth_uri $KEYSTONE_SERVICE_PROTOCOL://$KEYSTONE_SERVICE_HOST:$KEYSTONE_SERVICE_PORT/ |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken admin_tenant_name $SERVICE_TENANT_NAME |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken admin_user swift |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken admin_password $SERVICE_PASSWORD |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:authtoken signing_dir $SWIFT_AUTH_CACHE_DIR |
| |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} filter:keystoneauth use |
| iniuncomment ${SWIFT_CONFIG_PROXY_SERVER} filter:keystoneauth operator_roles |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:keystoneauth operator_roles "Member, admin" |
| |
| if is_service_enabled swift3; then |
| cat <<EOF >>${SWIFT_CONFIG_PROXY_SERVER} |
| # NOTE(chmou): s3token middleware is not updated yet to use only |
| # username and password. |
| [filter:s3token] |
| paste.filter_factory = keystone.middleware.s3_token:filter_factory |
| auth_port = ${KEYSTONE_AUTH_PORT} |
| auth_host = ${KEYSTONE_AUTH_HOST} |
| auth_protocol = ${KEYSTONE_AUTH_PROTOCOL} |
| auth_token = ${SERVICE_TOKEN} |
| admin_token = ${SERVICE_TOKEN} |
| |
| [filter:swift3] |
| use = egg:swift3#swift3 |
| EOF |
| fi |
| |
| cp ${SWIFT_DIR}/etc/swift.conf-sample ${SWIFT_CONF_DIR}/swift.conf |
| iniset ${SWIFT_CONF_DIR}/swift.conf swift-hash swift_hash_path_suffix ${SWIFT_HASH} |
| |
| # This function generates an object/account/proxy configuration |
| # emulating 4 nodes on different ports |
| function generate_swift_config() { |
| local swift_node_config=$1 |
| local node_id=$2 |
| local bind_port=$3 |
| local server_type=$4 |
| |
| log_facility=$[ node_id - 1 ] |
| node_path=${SWIFT_DATA_DIR}/${node_number} |
| |
| iniuncomment ${swift_node_config} DEFAULT user |
| iniset ${swift_node_config} DEFAULT user ${USER} |
| |
| iniuncomment ${swift_node_config} DEFAULT bind_port |
| iniset ${swift_node_config} DEFAULT bind_port ${bind_port} |
| |
| iniuncomment ${swift_node_config} DEFAULT swift_dir |
| iniset ${swift_node_config} DEFAULT swift_dir ${SWIFT_CONF_DIR} |
| |
| iniuncomment ${swift_node_config} DEFAULT devices |
| iniset ${swift_node_config} DEFAULT devices ${node_path} |
| |
| iniuncomment ${swift_node_config} DEFAULT log_facility |
| iniset ${swift_node_config} DEFAULT log_facility LOG_LOCAL${log_facility} |
| |
| iniuncomment ${swift_node_config} DEFAULT disable_fallocate |
| iniset ${swift_node_config} DEFAULT disable_fallocate true |
| |
| iniuncomment ${swift_node_config} DEFAULT mount_check |
| iniset ${swift_node_config} DEFAULT mount_check false |
| |
| iniuncomment ${swift_node_config} ${server_type}-replicator vm_test_mode |
| iniset ${swift_node_config} ${server_type}-replicator vm_test_mode yes |
| } |
| |
| for node_number in ${SWIFT_REPLICAS_SEQ}; do |
| swift_node_config=${SWIFT_CONF_DIR}/object-server/${node_number}.conf |
| cp ${SWIFT_DIR}/etc/object-server.conf-sample ${swift_node_config} |
| generate_swift_config ${swift_node_config} ${node_number} $[OBJECT_PORT_BASE + 10 * (node_number - 1)] object |
| iniset ${swift_node_config} filter:recon recon_cache_path ${SWIFT_DATA_DIR}/cache |
| # Using a sed and not iniset/iniuncomment because we want to a global |
| # modification and make sure it works for new sections. |
| sed -i -e "s,#[ ]*recon_cache_path .*,recon_cache_path = ${SWIFT_DATA_DIR}/cache," ${swift_node_config} |
| |
| swift_node_config=${SWIFT_CONF_DIR}/container-server/${node_number}.conf |
| cp ${SWIFT_DIR}/etc/container-server.conf-sample ${swift_node_config} |
| generate_swift_config ${swift_node_config} ${node_number} $[CONTAINER_PORT_BASE + 10 * (node_number - 1)] container |
| iniuncomment ${swift_node_config} app:container-server allow_versions |
| iniset ${swift_node_config} app:container-server allow_versions "true" |
| sed -i -e "s,#[ ]*recon_cache_path .*,recon_cache_path = ${SWIFT_DATA_DIR}/cache," ${swift_node_config} |
| |
| swift_node_config=${SWIFT_CONF_DIR}/account-server/${node_number}.conf |
| cp ${SWIFT_DIR}/etc/account-server.conf-sample ${swift_node_config} |
| generate_swift_config ${swift_node_config} ${node_number} $[ACCOUNT_PORT_BASE + 10 * (node_number - 1)] account |
| sed -i -e "s,#[ ]*recon_cache_path .*,recon_cache_path = ${SWIFT_DATA_DIR}/cache," ${swift_node_config} |
| done |
| |
| # Set new accounts in tempauth to match keystone tenant/user (to make testing easier) |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:tempauth user_swifttenanttest1_swiftusertest1 "testing .admin" |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:tempauth user_swifttenanttest2_swiftusertest2 "testing2 .admin" |
| iniset ${SWIFT_CONFIG_PROXY_SERVER} filter:tempauth user_swifttenanttest1_swiftusertest3 "testing3 .admin" |
| |
| testfile=${SWIFT_CONF_DIR}/test.conf |
| cp ${SWIFT_DIR}/test/sample.conf ${testfile} |
| |
| # Set accounts for functional tests |
| iniset ${testfile} func_test account swifttenanttest1 |
| iniset ${testfile} func_test username swiftusertest1 |
| iniset ${testfile} func_test username3 swiftusertest3 |
| iniset ${testfile} func_test account2 swifttenanttest2 |
| iniset ${testfile} func_test username2 swiftusertest2 |
| |
| if is_service_enabled key;then |
| iniuncomment ${testfile} func_test auth_version |
| iniset ${testfile} func_test auth_host ${KEYSTONE_SERVICE_HOST} |
| iniset ${testfile} func_test auth_port ${KEYSTONE_AUTH_PORT} |
| iniset ${testfile} func_test auth_prefix /v2.0/ |
| fi |
| |
| swift_log_dir=${SWIFT_DATA_DIR}/logs |
| rm -rf ${swift_log_dir} |
| mkdir -p ${swift_log_dir}/hourly |
| sudo chown -R $USER:adm ${swift_log_dir} |
| sed "s,%SWIFT_LOGDIR%,${swift_log_dir}," $FILES/swift/rsyslog.conf | sudo \ |
| tee /etc/rsyslog.d/10-swift.conf |
| if is_apache_enabled_service swift; then |
| _config_swift_apache_wsgi |
| fi |
| } |
| |
| # create_swift_disk - Create Swift backing disk |
| function create_swift_disk() { |
| local node_number |
| |
| # First do a bit of setup by creating the directories and |
| # changing the permissions so we can run it as our user. |
| |
| USER_GROUP=$(id -g) |
| sudo mkdir -p ${SWIFT_DATA_DIR}/{drives,cache,run,logs} |
| sudo chown -R $USER:${USER_GROUP} ${SWIFT_DATA_DIR} |
| |
| # Create a loopback disk and format it to XFS. |
| if [[ -e ${SWIFT_DATA_DIR}/drives/images/swift.img ]]; then |
| if egrep -q ${SWIFT_DATA_DIR}/drives/sdb1 /proc/mounts; then |
| sudo umount ${SWIFT_DATA_DIR}/drives/sdb1 |
| sudo rm -f ${SWIFT_DATA_DIR}/drives/images/swift.img |
| fi |
| fi |
| |
| mkdir -p ${SWIFT_DATA_DIR}/drives/images |
| sudo touch ${SWIFT_DATA_DIR}/drives/images/swift.img |
| sudo chown $USER: ${SWIFT_DATA_DIR}/drives/images/swift.img |
| |
| dd if=/dev/zero of=${SWIFT_DATA_DIR}/drives/images/swift.img \ |
| bs=1024 count=0 seek=${SWIFT_LOOPBACK_DISK_SIZE} |
| |
| # Make a fresh XFS filesystem |
| mkfs.xfs -f -i size=1024 ${SWIFT_DATA_DIR}/drives/images/swift.img |
| |
| # Mount the disk with mount options to make it as efficient as possible |
| mkdir -p ${SWIFT_DATA_DIR}/drives/sdb1 |
| if ! egrep -q ${SWIFT_DATA_DIR}/drives/sdb1 /proc/mounts; then |
| sudo mount -t xfs -o loop,noatime,nodiratime,nobarrier,logbufs=8 \ |
| ${SWIFT_DATA_DIR}/drives/images/swift.img ${SWIFT_DATA_DIR}/drives/sdb1 |
| fi |
| |
| # Create a link to the above mount and |
| # create all of the directories needed to emulate a few different servers |
| for node_number in ${SWIFT_REPLICAS_SEQ}; do |
| sudo ln -sf ${SWIFT_DATA_DIR}/drives/sdb1/$node_number ${SWIFT_DATA_DIR}/$node_number; |
| drive=${SWIFT_DATA_DIR}/drives/sdb1/${node_number} |
| node=${SWIFT_DATA_DIR}/${node_number}/node |
| node_device=${node}/sdb1 |
| [[ -d $node ]] && continue |
| [[ -d $drive ]] && continue |
| sudo install -o ${USER} -g $USER_GROUP -d $drive |
| sudo install -o ${USER} -g $USER_GROUP -d $node_device |
| sudo chown -R $USER: ${node} |
| done |
| } |
| # create_swift_accounts() - Set up standard swift accounts and extra |
| # one for tests we do this by attaching all words in the account name |
| # since we want to make it compatible with tempauth which use |
| # underscores for separators. |
| |
| # Tenant User Roles |
| # ------------------------------------------------------------------ |
| # service swift service |
| # swifttenanttest1 swiftusertest1 admin |
| # swifttenanttest1 swiftusertest3 anotherrole |
| # swifttenanttest2 swiftusertest2 admin |
| |
| function create_swift_accounts() { |
| KEYSTONE_CATALOG_BACKEND=${KEYSTONE_CATALOG_BACKEND:-sql} |
| |
| SERVICE_TENANT=$(keystone tenant-list | awk "/ $SERVICE_TENANT_NAME / { print \$2 }") |
| ADMIN_ROLE=$(keystone role-list | awk "/ admin / { print \$2 }") |
| |
| SWIFT_USER=$(keystone user-create --name=swift --pass="$SERVICE_PASSWORD" \ |
| --tenant_id $SERVICE_TENANT --email=swift@example.com | grep " id " | get_field 2) |
| keystone user-role-add --tenant_id $SERVICE_TENANT --user_id $SWIFT_USER --role_id $ADMIN_ROLE |
| |
| if [[ "$KEYSTONE_CATALOG_BACKEND" = 'sql' ]]; then |
| SWIFT_SERVICE=$(keystone service-create --name=swift --type="object-store" \ |
| --description="Swift Service" | grep " id " | get_field 2) |
| keystone endpoint-create \ |
| --region RegionOne \ |
| --service_id $SWIFT_SERVICE \ |
| --publicurl "http://$SERVICE_HOST:8080/v1/AUTH_\$(tenant_id)s" \ |
| --adminurl "http://$SERVICE_HOST:8080" \ |
| --internalurl "http://$SERVICE_HOST:8080/v1/AUTH_\$(tenant_id)s" |
| fi |
| |
| SWIFT_TENANT_TEST1=$(keystone tenant-create --name=swifttenanttest1 | grep " id " | get_field 2) |
| SWIFT_USER_TEST1=$(keystone user-create --name=swiftusertest1 --pass=testing --email=test@example.com | grep " id " | get_field 2) |
| keystone user-role-add --user_id $SWIFT_USER_TEST1 --role_id $ADMIN_ROLE --tenant_id $SWIFT_TENANT_TEST1 |
| |
| SWIFT_USER_TEST3=$(keystone user-create --name=swiftusertest3 --pass=testing3 --email=test3@example.com | grep " id " | get_field 2) |
| keystone user-role-add --user_id $SWIFT_USER_TEST3 --role_id $ANOTHER_ROLE --tenant_id $SWIFT_TENANT_TEST1 |
| |
| SWIFT_TENANT_TEST2=$(keystone tenant-create --name=swifttenanttest2 | grep " id " | get_field 2) |
| SWIFT_USER_TEST2=$(keystone user-create --name=swiftusertest2 --pass=testing2 --email=test2@example.com | grep " id " | get_field 2) |
| keystone user-role-add --user_id $SWIFT_USER_TEST2 --role_id $ADMIN_ROLE --tenant_id $SWIFT_TENANT_TEST2 |
| } |
| |
| # init_swift() - Initialize rings |
| function init_swift() { |
| local node_number |
| # Make sure to kill all swift processes first |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run all stop || true |
| |
| # Forcibly re-create the backing filesystem |
| create_swift_disk |
| |
| # This is where we create three different rings for swift with |
| # different object servers binding on different ports. |
| pushd ${SWIFT_CONF_DIR} >/dev/null && { |
| |
| rm -f *.builder *.ring.gz backups/*.builder backups/*.ring.gz |
| |
| swift-ring-builder object.builder create ${SWIFT_PARTITION_POWER_SIZE} ${SWIFT_REPLICAS} 1 |
| swift-ring-builder container.builder create ${SWIFT_PARTITION_POWER_SIZE} ${SWIFT_REPLICAS} 1 |
| swift-ring-builder account.builder create ${SWIFT_PARTITION_POWER_SIZE} ${SWIFT_REPLICAS} 1 |
| |
| for node_number in ${SWIFT_REPLICAS_SEQ}; do |
| swift-ring-builder object.builder add z${node_number}-127.0.0.1:$[OBJECT_PORT_BASE + 10 * (node_number - 1)]/sdb1 1 |
| swift-ring-builder container.builder add z${node_number}-127.0.0.1:$[CONTAINER_PORT_BASE + 10 * (node_number - 1)]/sdb1 1 |
| swift-ring-builder account.builder add z${node_number}-127.0.0.1:$[ACCOUNT_PORT_BASE + 10 * (node_number - 1)]/sdb1 1 |
| done |
| swift-ring-builder object.builder rebalance |
| swift-ring-builder container.builder rebalance |
| swift-ring-builder account.builder rebalance |
| } && popd >/dev/null |
| |
| # Create cache dir |
| sudo mkdir -p $SWIFT_AUTH_CACHE_DIR |
| sudo chown $STACK_USER $SWIFT_AUTH_CACHE_DIR |
| rm -f $SWIFT_AUTH_CACHE_DIR/* |
| } |
| |
| function install_swift() { |
| git_clone $SWIFT_REPO $SWIFT_DIR $SWIFT_BRANCH |
| setup_develop $SWIFT_DIR |
| if is_apache_enabled_service swift; then |
| install_apache_wsgi |
| fi |
| } |
| |
| function install_swiftclient() { |
| git_clone $SWIFTCLIENT_REPO $SWIFTCLIENT_DIR $SWIFTCLIENT_BRANCH |
| setup_develop $SWIFTCLIENT_DIR |
| } |
| |
| # start_swift() - Start running processes, including screen |
| function start_swift() { |
| # (re)start rsyslog |
| restart_service rsyslog |
| # (re)start memcached to make sure we have a clean memcache. |
| restart_service memcached |
| |
| # Start rsync |
| if is_ubuntu; then |
| sudo /etc/init.d/rsync restart || : |
| else |
| sudo systemctl start xinetd.service |
| fi |
| |
| if is_apache_enabled_service swift; then |
| # Make sure the apache lock dir is owned by $STACK_USER |
| # for running apache server to avoid failure of restarting |
| # apache server due to permission problem. |
| sudo chown -R $STACK_USER /var/run/lock/$APACHE_NAME |
| restart_apache_server |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run rest start |
| screen_it s-proxy "cd $SWIFT_DIR && sudo tail -f /var/log/$APACHE_NAME/proxy-server" |
| if [[ ${SWIFT_REPLICAS} == 1 ]]; then |
| for type in object container account; do |
| screen_it s-${type} "cd $SWIFT_DIR && sudo tail -f /var/log/$APACHE_NAME/${type}-server-1" |
| done |
| fi |
| return 0 |
| fi |
| |
| # By default with only one replica we are launching the proxy, |
| # container, account and object server in screen in foreground and |
| # other services in background. If we have SWIFT_REPLICAS set to something |
| # greater than one we first spawn all the swift services then kill the proxy |
| # service so we can run it in foreground in screen. ``swift-init ... |
| # {stop|restart}`` exits with '1' if no servers are running, ignore it just |
| # in case |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run all restart || true |
| if [[ ${SWIFT_REPLICAS} == 1 ]]; then |
| todo="object container account" |
| fi |
| for type in proxy ${todo}; do |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run ${type} stop || true |
| done |
| screen_it s-proxy "cd $SWIFT_DIR && $SWIFT_DIR/bin/swift-proxy-server ${SWIFT_CONF_DIR}/proxy-server.conf -v" |
| if [[ ${SWIFT_REPLICAS} == 1 ]]; then |
| for type in object container account; do |
| screen_it s-${type} "cd $SWIFT_DIR && $SWIFT_DIR/bin/swift-${type}-server ${SWIFT_CONF_DIR}/${type}-server/1.conf -v" |
| done |
| fi |
| } |
| |
| # stop_swift() - Stop running processes (non-screen) |
| function stop_swift() { |
| |
| if is_apache_enabled_service swift; then |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run rest stop && return 0 |
| fi |
| |
| # screen normally killed by unstack.sh |
| if type -p swift-init >/dev/null; then |
| swift-init --run-dir=${SWIFT_DATA_DIR}/run all stop || true |
| fi |
| # Dump the proxy server |
| sudo pkill -f swift-proxy-server |
| } |
| |
| # Restore xtrace |
| $XTRACE |
| |
| # Local variables: |
| # mode: shell-script |
| # End: |