merge origin/master
diff --git a/.gitignore b/.gitignore
index 22a7898..e482090 100644
--- a/.gitignore
+++ b/.gitignore
@@ -1,3 +1,5 @@
proto
*~
+*.log
+src
localrc
diff --git a/exercises/floating_ips.sh b/exercises/floating_ips.sh
old mode 100755
new mode 100644
index 5c38430..75046d1
--- a/exercises/floating_ips.sh
+++ b/exercises/floating_ips.sh
@@ -84,6 +84,15 @@
# Waiting for boot
# ----------------
+# Max time to wait while vm goes from build to active state
+ACTIVE_TIMEOUT=${ACTIVE_TIMEOUT:-10}
+
+# Max time till the vm is bootable
+BOOT_TIMEOUT=${BOOT_TIMEOUT:-15}
+
+# Max time to wait for proper association and dis-association.
+ASSOCIATE_TIMEOUT=${ASSOCIATE_TIMEOUT:-10}
+
# check that the status is active within ACTIVE_TIMEOUT seconds
if ! timeout $ACTIVE_TIMEOUT sh -c "while ! nova show $NAME | grep status | grep -q ACTIVE; do sleep 1; done"; then
echo "server didn't become active!"
@@ -158,7 +167,7 @@
nova secgroup-delete-rule $SECGROUP icmp -1 -1 0.0.0.0/0
# FIXME (anthony): make xs support security groups
-if [ "$VIRT_DRIVER" != "xenserver"]; then
+if [ "$VIRT_DRIVER" != "xenserver" ]; then
# test we can aren't able to ping our floating ip within ASSOCIATE_TIMEOUT seconds
if ! timeout $ASSOCIATE_TIMEOUT sh -c "while ping -c1 -w1 $FLOATING_IP; do sleep 1; done"; then
print "Security group failure - ping should not be allowed!"
diff --git a/files/apts/nova b/files/apts/nova
index 8ae74a2..9eefed7 100644
--- a/files/apts/nova
+++ b/files/apts/nova
@@ -1,7 +1,9 @@
dnsmasq-base
+dnsmasq-utils # for dhcp_release
kpartx
parted
-mysql-server
+arping # used for send_arp_for_ha option in nova-network
+mysql-server # NOPRIME
python-mysqldb
kvm
gawk
@@ -10,10 +12,10 @@
sqlite3
sudo
kvm
-libvirt-bin
+libvirt-bin # NOPRIME
vlan
curl
-rabbitmq-server
+rabbitmq-server # NOPRIME
socat # used by ajaxterm
python-mox
python-paste
diff --git a/files/apts/preseed b/files/apts/preseed
deleted file mode 100644
index 8712d5d..0000000
--- a/files/apts/preseed
+++ /dev/null
@@ -1,18 +0,0 @@
-# a collection of packages that speed up installation as they are dependencies
-# of packages we can't install during bootstraping (rabbitmq-server,
-# mysql-server, libvirt-bin)
-#
-# NOTE: only add packages to this file that aren't needed directly
-mysql-common
-mysql-client-5.1
-erlang-base
-erlang-ssl
-erlang-nox
-erlang-inets
-erlang-mnesia
-libhtml-template-perl
-gettext-base
-libavahi-client3
-libxml2-utils
-libpciaccess0
-libparted0debian1
diff --git a/files/apts/swift b/files/apts/swift
new file mode 100644
index 0000000..f298377
--- /dev/null
+++ b/files/apts/swift
@@ -0,0 +1,17 @@
+curl
+gcc
+memcached # NOPRIME
+python-configobj
+python-coverage
+python-dev
+python-eventlet
+python-greenlet
+python-netifaces
+python-nose
+python-pastedeploy
+python-setuptools
+python-simplejson
+python-webob
+python-xattr
+sqlite3
+xfsprogs
diff --git a/files/glance-api.conf b/files/glance-api.conf
index 3499ff7..bb758af 100644
--- a/files/glance-api.conf
+++ b/files/glance-api.conf
@@ -24,7 +24,7 @@
# Log to this file. Make sure you do not set the same log
# file for both the API and registry servers!
-log_file = %DEST%/glance/api.log
+#log_file = %DEST%/glance/api.log
# Send logs to syslog (/dev/log) instead of to file specified by `log_file`
use_syslog = %SYSLOG%
diff --git a/files/glance-registry.conf b/files/glance-registry.conf
index 351b09f..1e04186 100644
--- a/files/glance-registry.conf
+++ b/files/glance-registry.conf
@@ -13,7 +13,7 @@
# Log to this file. Make sure you do not set the same log
# file for both the API and registry servers!
-log_file = %DEST%/glance/registry.log
+#log_file = %DEST%/glance/registry.log
# Where to store images
filesystem_store_datadir = %DEST%/glance/images
diff --git a/files/keystone_data.sh b/files/keystone_data.sh
index a5e75a6..d926c52 100755
--- a/files/keystone_data.sh
+++ b/files/keystone_data.sh
@@ -30,12 +30,13 @@
$BIN_DIR/keystone-manage $* service add nova compute "Nova Compute Service"
$BIN_DIR/keystone-manage $* service add glance image "Glance Image Service"
$BIN_DIR/keystone-manage $* service add keystone identity "Keystone Identity Service"
+$BIN_DIR/keystone-manage $* service add swift object-store "Swift Service"
#endpointTemplates
$BIN_DIR/keystone-manage $* endpointTemplates add RegionOne nova http://%HOST_IP%:8774/v1.1/%tenant_id% http://%HOST_IP%:8774/v1.1/%tenant_id% http://%HOST_IP%:8774/v1.1/%tenant_id% 1 1
$BIN_DIR/keystone-manage $* endpointTemplates add RegionOne glance http://%HOST_IP%:9292/v1.1/%tenant_id% http://%HOST_IP%:9292/v1.1/%tenant_id% http://%HOST_IP%:9292/v1.1/%tenant_id% 1 1
$BIN_DIR/keystone-manage $* endpointTemplates add RegionOne keystone http://%HOST_IP%:5000/v2.0 http://%HOST_IP%:35357/v2.0 http://%HOST_IP%:5000/v2.0 1 1
-# $BIN_DIR/keystone-manage $* endpointTemplates add RegionOne swift http://%HOST_IP%:8080/v1/AUTH_%tenant_id% http://%HOST_IP%:8080/ http://%HOST_IP%:8080/v1/AUTH_%tenant_id% 1 1
+$BIN_DIR/keystone-manage $* endpointTemplates add RegionOne swift http://%HOST_IP%:8080/v1/AUTH_%tenant_id% http://%HOST_IP%:8080/ http://%HOST_IP%:8080/v1/AUTH_%tenant_id% 1 1
# Tokens
$BIN_DIR/keystone-manage $* token add %SERVICE_TOKEN% admin admin 2015-02-05T00:00
diff --git a/files/nova-api-paste.ini b/files/nova-api-paste.ini
new file mode 100644
index 0000000..2c642f8
--- /dev/null
+++ b/files/nova-api-paste.ini
@@ -0,0 +1,127 @@
+#######
+# EC2 #
+#######
+
+[composite:ec2]
+use = egg:Paste#urlmap
+/: ec2versions
+/services/Cloud: ec2cloud
+/services/Admin: ec2admin
+/latest: ec2metadata
+/2007-01-19: ec2metadata
+/2007-03-01: ec2metadata
+/2007-08-29: ec2metadata
+/2007-10-10: ec2metadata
+/2007-12-15: ec2metadata
+/2008-02-01: ec2metadata
+/2008-09-01: ec2metadata
+/2009-04-04: ec2metadata
+/1.0: ec2metadata
+
+[pipeline:ec2cloud]
+pipeline = logrequest totoken authtoken keystonecontext cloudrequest authorizer ec2executor
+
+[pipeline:ec2admin]
+pipeline = logrequest totoken authtoken keystonecontext adminrequest authorizer ec2executor
+
+[pipeline:ec2metadata]
+pipeline = logrequest ec2md
+
+[pipeline:ec2versions]
+pipeline = logrequest ec2ver
+
+[filter:logrequest]
+paste.filter_factory = nova.api.ec2:RequestLogging.factory
+
+[filter:ec2lockout]
+paste.filter_factory = nova.api.ec2:Lockout.factory
+
+[filter:totoken]
+paste.filter_factory = keystone.middleware.ec2_token:EC2Token.factory
+
+[filter:ec2noauth]
+paste.filter_factory = nova.api.ec2:NoAuth.factory
+
+[filter:authenticate]
+paste.filter_factory = nova.api.ec2:Authenticate.factory
+
+[filter:cloudrequest]
+controller = nova.api.ec2.cloud.CloudController
+paste.filter_factory = nova.api.ec2:Requestify.factory
+
+[filter:adminrequest]
+controller = nova.api.ec2.admin.AdminController
+paste.filter_factory = nova.api.ec2:Requestify.factory
+
+[filter:authorizer]
+paste.filter_factory = nova.api.ec2:Authorizer.factory
+
+[app:ec2executor]
+paste.app_factory = nova.api.ec2:Executor.factory
+
+[app:ec2ver]
+paste.app_factory = nova.api.ec2:Versions.factory
+
+[app:ec2md]
+paste.app_factory = nova.api.ec2.metadatarequesthandler:MetadataRequestHandler.factory
+
+#############
+# Openstack #
+#############
+
+[composite:osapi]
+use = egg:Paste#urlmap
+/: osversions
+/v1.0: openstackapi10
+/v1.1: openstackapi11
+
+[pipeline:openstackapi10]
+pipeline = faultwrap authtoken keystonecontext ratelimit osapiapp10
+
+[pipeline:openstackapi11]
+pipeline = faultwrap authtoken keystonecontext ratelimit extensions osapiapp11
+
+[filter:faultwrap]
+paste.filter_factory = nova.api.openstack:FaultWrapper.factory
+
+[filter:auth]
+paste.filter_factory = nova.api.openstack.auth:AuthMiddleware.factory
+
+[filter:noauth]
+paste.filter_factory = nova.api.openstack.auth:NoAuthMiddleware.factory
+
+[filter:ratelimit]
+paste.filter_factory = nova.api.openstack.limits:RateLimitingMiddleware.factory
+
+[filter:extensions]
+paste.filter_factory = nova.api.openstack.extensions:ExtensionMiddleware.factory
+
+[app:osapiapp10]
+paste.app_factory = nova.api.openstack:APIRouterV10.factory
+
+[app:osapiapp11]
+paste.app_factory = nova.api.openstack:APIRouterV11.factory
+
+[pipeline:osversions]
+pipeline = faultwrap osversionapp
+
+[app:osversionapp]
+paste.app_factory = nova.api.openstack.versions:Versions.factory
+
+##########
+# Shared #
+##########
+
+[filter:keystonecontext]
+paste.filter_factory = keystone.middleware.nova_keystone_context:NovaKeystoneContext.factory
+
+[filter:authtoken]
+paste.filter_factory = keystone.middleware.auth_token:filter_factory
+service_protocol = http
+service_host = 127.0.0.1
+service_port = 5000
+auth_host = 127.0.0.1
+auth_port = 35357
+auth_protocol = http
+auth_uri = http://127.0.0.1:5000/
+admin_token = %SERVICE_TOKEN%
diff --git a/files/sudo/nova b/files/sudo/nova
index 2ce1aac..62685b3 100644
--- a/files/sudo/nova
+++ b/files/sudo/nova
@@ -41,7 +41,7 @@
/usr/bin/socat, \
/sbin/parted, \
/usr/sbin/dnsmasq, \
- /usr/bin/arping
+ /usr/sbin/arping
%USER% ALL = (root) NOPASSWD: SETENV: NOVACMDS
diff --git a/files/swift/account-server.conf b/files/swift/account-server.conf
new file mode 100644
index 0000000..db0f097
--- /dev/null
+++ b/files/swift/account-server.conf
@@ -0,0 +1,20 @@
+[DEFAULT]
+devices = %NODE_PATH%/node
+mount_check = false
+bind_port = %BIND_PORT%
+user = %USER%
+log_facility = LOG_LOCAL%LOG_FACILITY%
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = account-server
+
+[app:account-server]
+use = egg:swift#account
+
+[account-replicator]
+vm_test_mode = yes
+
+[account-auditor]
+
+[account-reaper]
diff --git a/files/swift/container-server.conf b/files/swift/container-server.conf
new file mode 100644
index 0000000..bdc3e3a
--- /dev/null
+++ b/files/swift/container-server.conf
@@ -0,0 +1,22 @@
+[DEFAULT]
+devices = %NODE_PATH%/node
+mount_check = false
+bind_port = %BIND_PORT%
+user = %USER%
+log_facility = LOG_LOCAL%LOG_FACILITY%
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = container-server
+
+[app:container-server]
+use = egg:swift#container
+
+[container-replicator]
+vm_test_mode = yes
+
+[container-updater]
+
+[container-auditor]
+
+[container-sync]
diff --git a/files/swift/object-server.conf b/files/swift/object-server.conf
new file mode 100644
index 0000000..06fbffe
--- /dev/null
+++ b/files/swift/object-server.conf
@@ -0,0 +1,20 @@
+[DEFAULT]
+devices = %NODE_PATH%/node
+mount_check = false
+bind_port = %BIND_PORT%
+user = %USER%
+log_facility = LOG_LOCAL%LOG_FACILITY%
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = object-server
+
+[app:object-server]
+use = egg:swift#object
+
+[object-replicator]
+vm_test_mode = yes
+
+[object-updater]
+
+[object-auditor]
diff --git a/files/swift/proxy-server.conf b/files/swift/proxy-server.conf
new file mode 100644
index 0000000..fe7e39b
--- /dev/null
+++ b/files/swift/proxy-server.conf
@@ -0,0 +1,32 @@
+[DEFAULT]
+bind_port = 8080
+user = %USER%
+log_facility = LOG_LOCAL1
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = healthcheck cache %AUTH_SERVER% proxy-server
+
+[app:proxy-server]
+use = egg:swift#proxy
+allow_account_management = true
+account_autocreate = true
+
+[filter:keystone]
+use = egg:swiftkeystone2#keystone2
+keystone_admin_token = %SERVICE_TOKEN%
+keystone_url = http://localhost:35357/v2.0
+
+[filter:tempauth]
+use = egg:swift#tempauth
+user_admin_admin = admin .admin .reseller_admin
+user_test_tester = testing .admin
+user_test2_tester2 = testing2 .admin
+user_test_tester3 = testing3
+bind_ip = 0.0.0.0
+
+[filter:healthcheck]
+use = egg:swift#healthcheck
+
+[filter:cache]
+use = egg:swift#memcache
diff --git a/files/swift/rsyncd.conf b/files/swift/rsyncd.conf
new file mode 100644
index 0000000..66215c7
--- /dev/null
+++ b/files/swift/rsyncd.conf
@@ -0,0 +1,79 @@
+uid = %USER%
+gid = %GROUP%
+log file = /var/log/rsyncd.log
+pid file = /var/run/rsyncd.pid
+address = 127.0.0.1
+
+[account6012]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/1/node/
+read only = false
+lock file = /var/lock/account6012.lock
+
+[account6022]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/2/node/
+read only = false
+lock file = /var/lock/account6022.lock
+
+[account6032]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/3/node/
+read only = false
+lock file = /var/lock/account6032.lock
+
+[account6042]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/4/node/
+read only = false
+lock file = /var/lock/account6042.lock
+
+
+[container6011]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/1/node/
+read only = false
+lock file = /var/lock/container6011.lock
+
+[container6021]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/2/node/
+read only = false
+lock file = /var/lock/container6021.lock
+
+[container6031]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/3/node/
+read only = false
+lock file = /var/lock/container6031.lock
+
+[container6041]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/4/node/
+read only = false
+lock file = /var/lock/container6041.lock
+
+
+[object6010]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/1/node/
+read only = false
+lock file = /var/lock/object6010.lock
+
+[object6020]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/2/node/
+read only = false
+lock file = /var/lock/object6020.lock
+
+[object6030]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/3/node/
+read only = false
+lock file = /var/lock/object6030.lock
+
+[object6040]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/4/node/
+read only = false
+lock file = /var/lock/object6040.lock
diff --git a/files/swift/swift-remakerings b/files/swift/swift-remakerings
new file mode 100755
index 0000000..c65353c
--- /dev/null
+++ b/files/swift/swift-remakerings
@@ -0,0 +1,26 @@
+#!/bin/bash
+
+cd %SWIFT_CONFIG_LOCATION%
+
+rm -f *.builder *.ring.gz backups/*.builder backups/*.ring.gz
+
+swift-ring-builder object.builder create %SWIFT_PARTITION_POWER_SIZE% 3 1
+swift-ring-builder object.builder add z1-127.0.0.1:6010/sdb1 1
+swift-ring-builder object.builder add z2-127.0.0.1:6020/sdb2 1
+swift-ring-builder object.builder add z3-127.0.0.1:6030/sdb3 1
+swift-ring-builder object.builder add z4-127.0.0.1:6040/sdb4 1
+swift-ring-builder object.builder rebalance
+
+swift-ring-builder container.builder create %SWIFT_PARTITION_POWER_SIZE% 3 1
+swift-ring-builder container.builder add z1-127.0.0.1:6011/sdb1 1
+swift-ring-builder container.builder add z2-127.0.0.1:6021/sdb2 1
+swift-ring-builder container.builder add z3-127.0.0.1:6031/sdb3 1
+swift-ring-builder container.builder add z4-127.0.0.1:6041/sdb4 1
+swift-ring-builder container.builder rebalance
+
+swift-ring-builder account.builder create %SWIFT_PARTITION_POWER_SIZE% 3 1
+swift-ring-builder account.builder add z1-127.0.0.1:6012/sdb1 1
+swift-ring-builder account.builder add z2-127.0.0.1:6022/sdb2 1
+swift-ring-builder account.builder add z3-127.0.0.1:6032/sdb3 1
+swift-ring-builder account.builder add z4-127.0.0.1:6042/sdb4 1
+swift-ring-builder account.builder rebalance
diff --git a/files/swift/swift-startmain b/files/swift/swift-startmain
new file mode 100755
index 0000000..69efebd
--- /dev/null
+++ b/files/swift/swift-startmain
@@ -0,0 +1,3 @@
+#!/bin/bash
+
+swift-init all restart
diff --git a/files/swift/swift.conf b/files/swift/swift.conf
new file mode 100644
index 0000000..98df466
--- /dev/null
+++ b/files/swift/swift.conf
@@ -0,0 +1,3 @@
+[swift-hash]
+# random unique string that can never change (DO NOT LOSE)
+swift_hash_path_suffix = %SWIFT_HASH%
diff --git a/stack.sh b/stack.sh
index 9466512..78851b9 100755
--- a/stack.sh
+++ b/stack.sh
@@ -70,7 +70,7 @@
# called ``localrc``
#
# If ``localrc`` exists, then ``stackrc`` will load those settings. This is
-# useful for changing a branch or repostiory to test other versions. Also you
+# useful for changing a branch or repository to test other versions. Also you
# can store your other settings like **MYSQL_PASSWORD** or **ADMIN_PASSWORD** instead
# of letting devstack generate random ones for you.
source ./stackrc
@@ -103,8 +103,7 @@
# since this script runs as a normal user, we need to give that user
# ability to run sudo
- apt_get update
- apt_get install sudo
+ dpkg -l sudo || apt_get update && apt_get install sudo
if ! getent passwd stack >/dev/null; then
echo "Creating a user called stack"
@@ -150,6 +149,12 @@
NOVACLIENT_DIR=$DEST/python-novaclient
OPENSTACKX_DIR=$DEST/openstackx
NOVNC_DIR=$DEST/noVNC
+SWIFT_DIR=$DEST/swift
+SWIFT_KEYSTONE_DIR=$DEST/swift-keystone2
+QUANTUM_DIR=$DEST/quantum
+
+# Default Quantum Plugin
+Q_PLUGIN=${Q_PLUGIN:-openvswitch}
# Specify which services to launch. These generally correspond to screen tabs
ENABLED_SERVICES=${ENABLED_SERVICES:-g-api,g-reg,key,n-api,n-cpu,n-net,n-sch,n-vnc,horizon,mysql,rabbit}
@@ -169,6 +174,9 @@
HOST_IP=`LC_ALL=C /sbin/ifconfig | grep -m 1 'inet addr:'| cut -d: -f2 | awk '{print $1}'`
fi
+# Service startup timeout
+SERVICE_TIMEOUT=${SERVICE_TIMEOUT:-60}
+
# Generic helper to configure passwords
function read_password {
set +o xtrace
@@ -224,7 +232,7 @@
# Multi-host is a mode where each compute node runs its own network node. This
# allows network operations and routing for a VM to occur on the server that is
# running the VM - removing a SPOF and bandwidth bottleneck.
-MULTI_HOST=${MULTI_HOST:-0}
+MULTI_HOST=${MULTI_HOST:-False}
# If you are using FlatDHCP on multiple hosts, set the ``FLAT_INTERFACE``
# variable but make sure that the interface doesn't already have an
@@ -239,11 +247,22 @@
# If you are running on a single node and don't need to access the VMs from
# devices other than that node, you can set the flat interface to the same
# value as ``FLAT_NETWORK_BRIDGE``. This will stop the network hiccup from
-# occuring.
+# occurring.
FLAT_INTERFACE=${FLAT_INTERFACE:-eth0}
## FIXME(ja): should/can we check that FLAT_INTERFACE is sane?
+# Using Quantum networking:
+#
+# Make sure that q-svc is enabled in ENABLED_SERVICES. If it is the network
+# manager will be set to the QuantumManager.
+#
+# If you're planning to use the Quantum openvswitch plugin, set Q_PLUGIN to
+# "openvswitch" and make sure the q-agt service is enabled in
+# ENABLED_SERVICES.
+#
+# With Quantum networking the NET_MAN variable is ignored.
+
# MySQL & RabbitMQ
# ----------------
@@ -270,6 +289,42 @@
# Glance connection info. Note the port must be specified.
GLANCE_HOSTPORT=${GLANCE_HOSTPORT:-$HOST_IP:9292}
+# SWIFT
+# -----
+# TODO: implement glance support
+# TODO: add logging to different location.
+
+# By default the location of swift drives and objects is located inside
+# the swift source directory. SWIFT_DATA_LOCATION variable allow you to redefine
+# this.
+SWIFT_DATA_LOCATION=${SWIFT_DATA_LOCATION:-${SWIFT_DIR}/data}
+
+# We are going to have the configuration files inside the source
+# directory, change SWIFT_CONFIG_LOCATION if you want to adjust that.
+SWIFT_CONFIG_LOCATION=${SWIFT_CONFIG_LOCATION:-${SWIFT_DIR}/config}
+
+# devstack will create a loop-back disk formatted as XFS to store the
+# swift data. By default the disk size is 1 gigabyte. The variable
+# SWIFT_LOOPBACK_DISK_SIZE specified in bytes allow you to change
+# that.
+SWIFT_LOOPBACK_DISK_SIZE=${SWIFT_LOOPBACK_DISK_SIZE:-1000000}
+
+# The ring uses a configurable number of bits from a path’s MD5 hash as
+# a partition index that designates a device. The number of bits kept
+# from the hash is known as the partition power, and 2 to the partition
+# power indicates the partition count. Partitioning the full MD5 hash
+# ring allows other parts of the cluster to work in batches of items at
+# once which ends up either more efficient or at least less complex than
+# working with each item separately or the entire cluster all at once.
+# By default we define 9 for the partition count (which mean 512).
+SWIFT_PARTITION_POWER_SIZE=${SWIFT_PARTITION_POWER_SIZE:-9}
+
+# We only ask for Swift Hash if we have enabled swift service.
+if [[ "$ENABLED_SERVICES" =~ "swift" ]]; then
+ # SWIFT_HASH is a random unique string for a swift cluster that
+ # can never change.
+ read_password SWIFT_HASH "ENTER A RANDOM SWIFT HASH."
+fi
# Keystone
# --------
@@ -283,7 +338,7 @@
LOGFILE=${LOGFILE:-"$PWD/stack.sh.$$.log"}
(
# So that errors don't compound we exit on any errors so you see only the
-# first error that occured.
+# first error that occurred.
trap failed ERR
failed() {
local r=$?
@@ -310,7 +365,7 @@
# install apt requirements
apt_get update
-apt_get install `cat $FILES/apts/* | cut -d\# -f1 | grep -Ev "mysql-server|rabbitmq-server"`
+apt_get install `cat $FILES/apts/* | cut -d\# -f1 | grep -Ev "mysql-server|rabbitmq-server|memcached"`
# install python requirements
sudo PIP_DOWNLOAD_CACHE=/var/cache/pip pip install `cat $FILES/pips/*`
@@ -349,6 +404,10 @@
# compute service
git_clone $NOVA_REPO $NOVA_DIR $NOVA_BRANCH
+# storage service
+git_clone $SWIFT_REPO $SWIFT_DIR $SWIFT_BRANCH
+# swift + keystone middleware
+git_clone $SWIFT_KEYSTONE_REPO $SWIFT_KEYSTONE_DIR $SWIFT_KEYSTONE_BRANCH
# image catalog service
git_clone $GLANCE_REPO $GLANCE_DIR $GLANCE_BRANCH
# unified auth system (manages accounts/tokens)
@@ -362,6 +421,8 @@
# openstackx is a collection of extensions to openstack.compute & nova
# that is *deprecated*. The code is being moved into python-novaclient & nova.
git_clone $OPENSTACKX_REPO $OPENSTACKX_DIR $OPENSTACKX_BRANCH
+# quantum
+git_clone $QUANTUM_REPO $QUANTUM_DIR $QUANTUM_BRANCH
# Initialization
# ==============
@@ -370,12 +431,15 @@
# setup our checkouts so they are installed into python path
# allowing ``import nova`` or ``import glance.client``
cd $KEYSTONE_DIR; sudo python setup.py develop
+cd $SWIFT_DIR; sudo python setup.py develop
+cd $SWIFT_KEYSTONE_DIR; sudo python setup.py develop
cd $GLANCE_DIR; sudo python setup.py develop
cd $NOVACLIENT_DIR; sudo python setup.py develop
cd $NOVA_DIR; sudo python setup.py develop
cd $OPENSTACKX_DIR; sudo python setup.py develop
cd $HORIZON_DIR/django-openstack; sudo python setup.py develop
cd $HORIZON_DIR/openstack-dashboard; sudo python setup.py develop
+cd $QUANTUM_DIR; sudo python setup.py develop
# Add a useful screenrc. This isn't required to run openstack but is we do
# it since we are going to run the services in screen for simple
@@ -500,13 +564,12 @@
# ----
if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then
- # We are going to use the sample http middleware configuration from the
- # keystone project to launch nova. This paste config adds the configuration
- # required for nova to validate keystone tokens - except we need to switch
- # the config to use our service token instead (instead of the invalid token
- # 999888777666).
- cp $KEYSTONE_DIR/examples/paste/nova-api-paste.ini $NOVA_DIR/bin
- sed -e "s,999888777666,$SERVICE_TOKEN,g" -i $NOVA_DIR/bin/nova-api-paste.ini
+ # We are going to use a sample http middleware configuration based on the
+ # one from the keystone project to launch nova. This paste config adds
+ # the configuration required for nova to validate keystone tokens. We add
+ # our own service token to the configuration.
+ cp $FILES/nova-api-paste.ini $NOVA_DIR/bin
+ sed -e "s,%SERVICE_TOKEN%,$SERVICE_TOKEN,g" -i $NOVA_DIR/bin/nova-api-paste.ini
fi
if [[ "$ENABLED_SERVICES" =~ "n-cpu" ]]; then
@@ -580,6 +643,129 @@
mkdir -p $NOVA_DIR/networks
fi
+# Storage Service
+if [[ "$ENABLED_SERVICES" =~ "swift" ]]; then
+ # We first do a bit of setup by creating the directories and
+ # changing the permissions so we can run it as our user.
+
+ USER_GROUP=$(id -g)
+ sudo mkdir -p ${SWIFT_DATA_LOCATION}/drives
+ sudo chown -R $USER:${USER_GROUP} ${SWIFT_DATA_LOCATION}/drives
+
+ # We then create a loopback disk and format it to XFS.
+ if [[ ! -e ${SWIFT_DATA_LOCATION}/drives/images/swift.img ]];then
+ mkdir -p ${SWIFT_DATA_LOCATION}/drives/images
+ sudo touch ${SWIFT_DATA_LOCATION}/drives/images/swift.img
+ sudo chown $USER: ${SWIFT_DATA_LOCATION}/drives/images/swift.img
+
+ dd if=/dev/zero of=${SWIFT_DATA_LOCATION}/drives/images/swift.img \
+ bs=1024 count=0 seek=${SWIFT_LOOPBACK_DISK_SIZE}
+ mkfs.xfs -f -i size=1024 ${SWIFT_DATA_LOCATION}/drives/images/swift.img
+ fi
+
+ # After the drive being created we mount the disk with a few mount
+ # options to make it most efficient as possible for swift.
+ mkdir -p ${SWIFT_DATA_LOCATION}/drives/sdb1
+ if ! egrep -q ${SWIFT_DATA_LOCATION}/drives/sdb1 /proc/mounts;then
+ sudo mount -t xfs -o loop,noatime,nodiratime,nobarrier,logbufs=8 \
+ ${SWIFT_DATA_LOCATION}/drives/images/swift.img ${SWIFT_DATA_LOCATION}/drives/sdb1
+ fi
+
+ # We then create link to that mounted location so swift would know
+ # where to go.
+ for x in {1..4}; do sudo ln -sf ${SWIFT_DATA_LOCATION}/drives/sdb1/$x ${SWIFT_DATA_LOCATION}/$x; done
+
+ # We now have to emulate a few different servers into one we
+ # create all the directories needed for swift
+ tmpd=""
+ for d in ${SWIFT_DATA_LOCATION}/drives/sdb1/{1..4} \
+ ${SWIFT_CONFIG_LOCATION}/{object,container,account}-server \
+ ${SWIFT_DATA_LOCATION}/{1..4}/node/sdb1 /var/run/swift ;do
+ [[ -d $d ]] && continue
+ sudo install -o ${USER} -g $USER_GROUP -d $d
+ done
+
+ # We do want to make sure this is all owned by our user.
+ sudo chown -R $USER: ${SWIFT_DATA_LOCATION}/{1..4}/node
+ sudo chown -R $USER: ${SWIFT_CONFIG_LOCATION}
+
+ # swift-init has a bug using /etc/swift until bug #885595 is fixed
+ # we have to create a link
+ sudo ln -s ${SWIFT_CONFIG_LOCATION} /etc/swift
+
+ # Swift use rsync to syncronize between all the different
+ # partitions (which make more sense when you have a multi-node
+ # setup) we configure it with our version of rsync.
+ sed -e "s/%GROUP%/${USER_GROUP}/;s/%USER%/$USER/;s,%SWIFT_DATA_LOCATION%,$SWIFT_DATA_LOCATION," $FILES/swift/rsyncd.conf | sudo tee /etc/rsyncd.conf
+ sudo sed -i '/^RSYNC_ENABLE=false/ { s/false/true/ }' /etc/default/rsync
+
+ # By default Swift will be installed with the tempauth middleware
+ # which has some default username and password if you have
+ # configured keystone it will checkout the directory.
+ if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
+ swift_auth_server=keystone
+ # We need a special version of bin/swift which understand the
+ # OpenStack api 2.0, we download it until this is getting
+ # integrated in swift.
+ sudo curl -s -o/usr/local/bin/swift \
+ 'https://review.openstack.org/gitweb?p=openstack/swift.git;a=blob_plain;f=bin/swift;hb=48bfda6e2fdf3886c98bd15649887d54b9a2574e'
+ else
+ swift_auth_server=tempauth
+ fi
+
+ # We do the install of the proxy-server and swift configuration
+ # replacing a few directives to match our configuration.
+ sed "s,%SWIFT_CONFIG_LOCATION%,${SWIFT_CONFIG_LOCATION},;s/%USER%/$USER/;s/%SERVICE_TOKEN%/${SERVICE_TOKEN}/;s/%AUTH_SERVER%/${swift_auth_server}/" \
+ $FILES/swift/proxy-server.conf|sudo tee ${SWIFT_CONFIG_LOCATION}/proxy-server.conf
+
+ sed -e "s/%SWIFT_HASH%/$SWIFT_HASH/" $FILES/swift/swift.conf > ${SWIFT_CONFIG_LOCATION}/swift.conf
+
+ # We need to generate a object/account/proxy configuration
+ # emulating 4 nodes on different ports we have a little function
+ # that help us doing that.
+ function generate_swift_configuration() {
+ local server_type=$1
+ local bind_port=$2
+ local log_facility=$3
+ local node_number
+
+ for node_number in {1..4};do
+ node_path=${SWIFT_DATA_LOCATION}/${node_number}
+ sed -e "s,%SWIFT_CONFIG_LOCATION%,${SWIFT_CONFIG_LOCATION},;s,%USER%,$USER,;s,%NODE_PATH%,${node_path},;s,%BIND_PORT%,${bind_port},;s,%LOG_FACILITY%,${log_facility}," \
+ $FILES/swift/${server_type}-server.conf > ${SWIFT_CONFIG_LOCATION}/${server_type}-server/${node_number}.conf
+ bind_port=$(( ${bind_port} + 10 ))
+ log_facility=$(( ${log_facility} + 1 ))
+ done
+ }
+ generate_swift_configuration object 6010 2
+ generate_swift_configuration container 6011 2
+ generate_swift_configuration account 6012 2
+
+ # We create two helper scripts :
+ #
+ # - swift-remakerings
+ # Allow to recreate rings from scratch.
+ # - swift-startmain
+ # Restart your full cluster.
+ #
+ sed -e "s,%SWIFT_CONFIG_LOCATION%,${SWIFT_CONFIG_LOCATION},;s/%SWIFT_PARTITION_POWER_SIZE%/$SWIFT_PARTITION_POWER_SIZE/" $FILES/swift/swift-remakerings | \
+ sudo tee /usr/local/bin/swift-remakerings
+ sudo install -m755 $FILES/swift/swift-startmain /usr/local/bin/
+ sudo chmod +x /usr/local/bin/swift-*
+
+ # We then can start rsync.
+ sudo /etc/init.d/rsync restart || :
+
+ # Create our ring for the object/container/account.
+ /usr/local/bin/swift-remakerings
+
+ # And now we launch swift-startmain to get our cluster running
+ # ready to be tested.
+ /usr/local/bin/swift-startmain || :
+
+ unset s swift_hash swift_auth_server tmpd
+fi
+
# Volume Service
# --------------
@@ -593,7 +779,7 @@
# By default, the backing file is 2G in size, and is stored in /opt/stack.
#
if ! sudo vgdisplay | grep -q nova-volumes; then
- VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-/opt/stack/nova-volumes-backing-file}
+ VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-$DEST/nova-volumes-backing-file}
VOLUME_BACKING_FILE_SIZE=${VOLUME_BACKING_FILE_SIZE:-2052M}
truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE
DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE`
@@ -616,8 +802,16 @@
add_nova_flag "--allow_admin_api"
add_nova_flag "--scheduler_driver=$SCHEDULER"
add_nova_flag "--dhcpbridge_flagfile=$NOVA_DIR/bin/nova.conf"
-add_nova_flag "--network_manager=nova.network.manager.$NET_MAN"
add_nova_flag "--fixed_range=$FIXED_RANGE"
+if [[ "$ENABLED_SERVICES" =~ "q-svc" ]]; then
+ add_nova_flag "--network_manager=nova.network.quantum.manager.QuantumManager"
+ if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
+ add_nova_flag "--libvirt_vif_type=ethernet"
+ add_nova_flag "--libvirt_vif_driver=nova.virt.libvirt.vif.LibvirtOpenVswitchDriver"
+ fi
+else
+ add_nova_flag "--network_manager=nova.network.manager.$NET_MAN"
+fi
add_nova_flag "--my_ip=$HOST_IP"
add_nova_flag "--public_interface=$PUBLIC_INTERFACE"
add_nova_flag "--vlan_interface=$VLAN_INTERFACE"
@@ -632,15 +826,16 @@
add_nova_flag "--rabbit_host=$RABBIT_HOST"
add_nova_flag "--rabbit_password=$RABBIT_PASSWORD"
add_nova_flag "--glance_api_servers=$GLANCE_HOSTPORT"
+add_nova_flag "--force_dhcp_release"
if [ -n "$INSTANCES_PATH" ]; then
add_nova_flag "--instances_path=$INSTANCES_PATH"
fi
-if [ -n "$MULTI_HOST" ]; then
- add_nova_flag "--multi_host=$MULTI_HOST"
- add_nova_flag "--send_arp_for_ha=1"
+if [ "$MULTI_HOST" != "False" ]; then
+ add_nova_flag "--multi_host"
+ add_nova_flag "--send_arp_for_ha"
fi
if [ "$SYSLOG" != "False" ]; then
- add_nova_flag "--use_syslog=1"
+ add_nova_flag "--use_syslog"
fi
# XenServer
@@ -676,12 +871,6 @@
# (re)create nova database
$NOVA_DIR/bin/nova-manage db sync
-
- # create a small network
- $NOVA_DIR/bin/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE
-
- # create some floating ips
- $NOVA_DIR/bin/nova-manage floating create $FLOATING_RANGE
fi
@@ -721,16 +910,20 @@
function screen_it {
NL=`echo -ne '\015'`
if [[ "$ENABLED_SERVICES" =~ "$1" ]]; then
- screen -S nova -X screen -t $1
- screen -S nova -p $1 -X stuff "$2$NL"
+ screen -S stack -X screen -t $1
+ # sleep to allow bash to be ready to be send the command - we are
+ # creating a new window in screen and then sends characters, so if
+ # bash isn't running by the time we send the command, nothing happens
+ sleep 1
+ screen -S stack -p $1 -X stuff "$2$NL"
fi
}
# create a new named screen to run processes in
-screen -d -m -S nova -t nova
+screen -d -m -S stack -t stack
sleep 1
-# launch the glance registery service
+# launch the glance registry service
if [[ "$ENABLED_SERVICES" =~ "g-reg" ]]; then
screen_it g-reg "cd $GLANCE_DIR; bin/glance-registry --config-file=etc/glance-registry.conf"
fi
@@ -739,7 +932,7 @@
if [[ "$ENABLED_SERVICES" =~ "g-api" ]]; then
screen_it g-api "cd $GLANCE_DIR; bin/glance-api --config-file=etc/glance-api.conf"
echo "Waiting for g-api ($GLANCE_HOSTPORT) to start..."
- if ! timeout 60 sh -c "while ! wget -q -O- http://$GLANCE_HOSTPORT; do sleep 1; done"; then
+ if ! timeout $SERVICE_TIMEOUT sh -c "while ! wget -q -O- http://$GLANCE_HOSTPORT; do sleep 1; done"; then
echo "g-api did not start"
exit 1
fi
@@ -749,7 +942,7 @@
if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
screen_it key "cd $KEYSTONE_DIR && $KEYSTONE_DIR/bin/keystone --config-file $KEYSTONE_CONF -d"
echo "Waiting for keystone to start..."
- if ! timeout 60 sh -c "while ! wget -q -O- http://127.0.0.1:5000; do sleep 1; done"; then
+ if ! timeout $SERVICE_TIMEOUT sh -c "while ! wget -q -O- http://127.0.0.1:5000; do sleep 1; done"; then
echo "keystone did not start"
exit 1
fi
@@ -759,11 +952,62 @@
if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then
screen_it n-api "cd $NOVA_DIR && $NOVA_DIR/bin/nova-api"
echo "Waiting for nova-api to start..."
- if ! timeout 60 sh -c "while ! wget -q -O- http://127.0.0.1:8774; do sleep 1; done"; then
+ if ! timeout $SERVICE_TIMEOUT sh -c "while ! wget -q -O- http://127.0.0.1:8774; do sleep 1; done"; then
echo "nova-api did not start"
exit 1
fi
fi
+
+# Quantum
+if [[ "$ENABLED_SERVICES" =~ "q-svc" ]]; then
+ # Install deps
+ # FIXME add to files/apts/quantum, but don't install if not needed!
+ apt_get install openvswitch-switch openvswitch-datapath-dkms
+
+ # Create database for the plugin/agent
+ if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
+ if [[ "$ENABLED_SERVICES" =~ "mysql" ]]; then
+ mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE IF NOT EXISTS ovs_quantum;'
+ else
+ echo "mysql must be enabled in order to use the $Q_PLUGIN Quantum plugin."
+ exit 1
+ fi
+ fi
+
+ QUANTUM_PLUGIN_INI_FILE=$QUANTUM_DIR/quantum/plugins.ini
+ # Make sure we're using the openvswitch plugin
+ sed -i -e "s/^provider =.*$/provider = quantum.plugins.openvswitch.ovs_quantum_plugin.OVSQuantumPlugin/g" $QUANTUM_PLUGIN_INI_FILE
+ screen_it q-svc "cd $QUANTUM_DIR && export PYTHONPATH=.:$PYTHONPATH; python $QUANTUM_DIR/bin/quantum $QUANTUM_DIR/etc/quantum.conf"
+fi
+
+# Quantum agent (for compute nodes)
+if [[ "$ENABLED_SERVICES" =~ "q-agt" ]]; then
+ if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
+ # Set up integration bridge
+ OVS_BRIDGE=${OVS_BRIDGE:-br-int}
+ sudo ovs-vsctl --no-wait -- --if-exists del-br $OVS_BRIDGE
+ sudo ovs-vsctl --no-wait add-br $OVS_BRIDGE
+ sudo ovs-vsctl --no-wait br-set-external-id $OVS_BRIDGE bridge-id br-int
+ fi
+
+ # Start up the quantum <-> openvswitch agent
+ screen_it q-agt "sleep 4; sudo python $QUANTUM_DIR/quantum/plugins/openvswitch/agent/ovs_quantum_agent.py $QUANTUM_DIR/quantum/plugins/openvswitch/ovs_quantum_plugin.ini -v"
+fi
+
+# If we're using Quantum (i.e. q-svc is enabled), network creation has to
+# happen after we've started the Quantum service.
+if [[ "$ENABLED_SERVICES" =~ "mysql" ]]; then
+ # create a small network
+ $NOVA_DIR/bin/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE
+
+ if [[ "$ENABLED_SERVICES" =~ "q-svc" ]]; then
+ echo "Not creating floating IPs (not supported by QuantumManager)"
+ else
+ # create some floating ips
+ $NOVA_DIR/bin/nova-manage floating create $FLOATING_RANGE
+ fi
+fi
+
# Launching nova-compute should be as simple as running ``nova-compute`` but
# have to do a little more than that in our script. Since we add the group
# ``libvirtd`` to our user in this script, when nova-compute is run it is
@@ -787,7 +1031,7 @@
# TTY also uses cloud-init, supporting login via keypair and sending scripts as
# userdata. See https://help.ubuntu.com/community/CloudInit for more on cloud-init
#
-# Override ``IMAGE_URLS`` with a comma-seperated list of uec images.
+# Override ``IMAGE_URLS`` with a comma-separated list of uec images.
#
# * **natty**: http://uec-images.ubuntu.com/natty/current/natty-server-cloudimg-amd64.tar.gz
# * **oneiric**: http://uec-images.ubuntu.com/oneiric/current/oneiric-server-cloudimg-amd64.tar.gz
@@ -843,6 +1087,10 @@
# Using the cloud
# ===============
+echo ""
+echo ""
+echo ""
+
# If you installed the horizon on this server, then you should be able
# to access the site using your browser.
if [[ "$ENABLED_SERVICES" =~ "horizon" ]]; then
diff --git a/stackrc b/stackrc
index 9b110a3..b541711 100644
--- a/stackrc
+++ b/stackrc
@@ -2,6 +2,14 @@
NOVA_REPO=https://github.com/cloudbuilders/nova.git
NOVA_BRANCH=diablo
+# storage service
+SWIFT_REPO=https://github.com/openstack/swift.git
+SWIFT_BRANCH=stable/diablo
+
+# swift and keystone integration
+SWIFT_KEYSTONE_REPO=https://github.com/cloudbuilders/swift-keystone2.git
+SWIFT_KEYSTONE_BRANCH=master
+
# image catalog service
GLANCE_REPO=https://github.com/cloudbuilders/glance.git
GLANCE_BRANCH=diablo
@@ -27,6 +35,10 @@
OPENSTACKX_REPO=https://github.com/cloudbuilders/openstackx.git
OPENSTACKX_BRANCH=diablo
+# quantum service
+QUANTUM_REPO=https://github.com/openstack/quantum
+QUANTUM_BRANCH=stable/diablo
+
# Specify a comma-separated list of uec images to download and install into glance.
IMAGE_URLS=http://smoser.brickies.net/ubuntu/ttylinux-uec/ttylinux-uec-amd64-11.2_2.6.35-15_1.tar.gz
diff --git a/tools/build_kvm.sh b/tools/build_kvm.sh
deleted file mode 100755
index 1b33926..0000000
--- a/tools/build_kvm.sh
+++ /dev/null
@@ -1,402 +0,0 @@
-#!/usr/bin/env bash
-
-# exit on error to stop unexpected errors
-set -o errexit
-
-# Make sure that we have the proper version of ubuntu
-UBUNTU_VERSION=`cat /etc/lsb-release | grep CODENAME | sed 's/.*=//g'`
-if [ ! "oneiric" = "$UBUNTU_VERSION" ]; then
- if [ ! "natty" = "$UBUNTU_VERSION" ]; then
- echo "This script only works with oneiric and natty"
- exit 1
- fi
-fi
-
-# Echo commands
-set -o xtrace
-
-# Keep track of the current directory
-TOOLS_DIR=$(cd $(dirname "$0") && pwd)
-TOP_DIR=$TOOLS_DIR/..
-
-# Where to store files and instances
-KVMSTACK_DIR=${KVMSTACK_DIR:-/opt/kvmstack}
-
-# Where to store images
-IMAGES_DIR=$KVMSTACK_DIR/images
-
-# Create images dir
-mkdir -p $IMAGES_DIR
-
-# Move to top devstack dir
-cd $TOP_DIR
-
-# Abort if localrc is not set
-if [ ! -e ./localrc ]; then
- echo "You must have a localrc with ALL necessary passwords defined before proceeding."
- echo "See stack.sh for required passwords."
- exit 1
-fi
-
-# Source params
-source ./stackrc
-
-# Configure the root password of the vm to be the same as ``ADMIN_PASSWORD``
-ROOT_PASSWORD=${ADMIN_PASSWORD:-password}
-
-
-# Base image (natty by default)
-DIST_NAME=${DIST_NAME:-natty}
-IMAGE_FNAME=$DIST_NAME.raw
-
-# Name of our instance, used by libvirt
-GUEST_NAME=${GUEST_NAME:-kvmstack}
-
-# Original version of built image
-BASE_IMAGE=$KVMSTACK_DIR/images/$DIST_NAME.raw
-
-# Copy of base image, which we pre-install with tasty treats
-VM_IMAGE=$IMAGES_DIR/$DIST_NAME.$GUEST_NAME.raw
-
-# Mop up after previous runs
-virsh destroy $GUEST_NAME || true
-
-# Where this vm is stored
-VM_DIR=$KVMSTACK_DIR/instances/$GUEST_NAME
-
-# Create vm dir
-mkdir -p $VM_DIR
-
-# Mount point into copied base image
-COPY_DIR=$VM_DIR/copy
-mkdir -p $COPY_DIR
-
-# Create the base image if it does not yet exist
-if [ ! -e $IMAGES_DIR/$IMAGE_FNAME ]; then
- cd $TOOLS_DIR
- ./make_image.sh -m -r 5000 $DIST_NAME raw
- mv $DIST_NAME.raw $BASE_IMAGE
- cd $TOP_DIR
-fi
-
-# Create a copy of the base image
-if [ ! -e $VM_IMAGE ]; then
- cp -p $BASE_IMAGE $VM_IMAGE
-fi
-
-# Unmount the copied base image
-function unmount_images() {
- # unmount the filesystem
- while df | grep -q $COPY_DIR; do
- umount $COPY_DIR || echo 'ok'
- sleep 1
- done
-}
-
-# Unmount from failed runs
-unmount_images
-
-# Ctrl-c catcher
-function kill_unmount() {
- unmount_images
- exit 1
-}
-
-# Install deps if needed
-dpkg -l kvm libvirt-bin kpartx || apt-get install -y --force-yes kvm libvirt-bin kpartx
-
-# Let Ctrl-c kill tail and exit
-trap kill_unmount SIGINT
-
-# Where Openstack code will live in image
-DEST=${DEST:-/opt/stack}
-
-# Mount the file system
-mount -o loop,offset=32256 $VM_IMAGE $COPY_DIR
-
-# git clone only if directory doesn't exist already. Since ``DEST`` might not
-# be owned by the installation user, we create the directory and change the
-# ownership to the proper user.
-function git_clone {
- if [ ! -d $2 ]; then
- sudo mkdir $2
- sudo chown `whoami` $2
- git clone $1 $2
- cd $2
- # This checkout syntax works for both branches and tags
- git checkout $3
- fi
-}
-
-# Make sure that base requirements are installed
-cp /etc/resolv.conf $COPY_DIR/etc/resolv.conf
-chroot $COPY_DIR apt-get update
-chroot $COPY_DIR apt-get install -y --force-yes `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-chroot $COPY_DIR apt-get install -y --download-only rabbitmq-server libvirt-bin mysql-server
-chroot $COPY_DIR pip install `cat files/pips/*`
-
-# Clean out code repos if directed to do so
-if [ "$CLEAN" = "1" ]; then
- rm -rf $COPY_DIR/$DEST
-fi
-
-# Cache openstack code
-mkdir -p $COPY_DIR/$DEST
-git_clone $NOVA_REPO $COPY_DIR/$DEST/nova $NOVA_BRANCH
-git_clone $GLANCE_REPO $COPY_DIR/$DEST/glance $GLANCE_BRANCH
-git_clone $KEYSTONE_REPO $COPY_DIR/$DESTkeystone $KEYSTONE_BRANCH
-git_clone $NOVNC_REPO $COPY_DIR/$DEST/noVNC $NOVNC_BRANCH
-git_clone $HORIZON_REPO $COPY_DIR/$DEST/horizon $HORIZON_BRANCH $HORIZON_TAG
-git_clone $NOVACLIENT_REPO $COPY_DIR/$DEST/python-novaclient $NOVACLIENT_BRANCH
-git_clone $OPENSTACKX_REPO $COPY_DIR/$DEST/openstackx $OPENSTACKX_BRANCH
-git_clone $KEYSTONE_REPO $COPY_DIR/$DEST/keystone $KEYSTONE_BRANCH
-git_clone $NOVNC_REPO $COPY_DIR/$DEST/noVNC $NOVNC_BRANCH
-
-# Back to devstack
-cd $TOP_DIR
-
-# Unmount the filesystems
-unmount_images
-
-# Network configuration variables
-BRIDGE=${BRIDGE:-br0}
-GUEST_IP=${GUEST_IP:-192.168.1.50}
-GUEST_CIDR=${GUEST_CIDR:-$GUEST_IP/24}
-GUEST_NETMASK=${GUEST_NETMASK:-255.255.255.0}
-GUEST_GATEWAY=${GUEST_GATEWAY:-192.168.1.1}
-GUEST_MAC=${GUEST_MAC:-"02:16:3e:07:69:`printf '%02X' $(echo $GUEST_IP | sed "s/.*\.//")`"}
-GUEST_RAM=${GUEST_RAM:-1524288}
-GUEST_CORES=${GUEST_CORES:-1}
-
-# libvirt.xml configuration
-LIBVIRT_XML=$VM_DIR/libvirt.xml
-cat > $LIBVIRT_XML <<EOF
-<domain type='kvm'>
- <name>$GUEST_NAME</name>
- <memory>$GUEST_RAM</memory>
- <os>
- <type>hvm</type>
- <bootmenu enable='yes'/>
- </os>
- <features>
- <acpi/>
- </features>
- <vcpu>$GUEST_CORES</vcpu>
- <devices>
- <disk type='file'>
- <driver type='qcow2'/>
- <source file='$VM_DIR/disk'/>
- <target dev='vda' bus='virtio'/>
- </disk>
-
- <interface type='bridge'>
- <source bridge='$BRIDGE'/>
- <mac address='$GUEST_MAC'/>
- </interface>
-
- <!-- The order is significant here. File must be defined first -->
- <serial type="file">
- <source path='$VM_DIR/console.log'/>
- <target port='1'/>
- </serial>
-
- <console type='pty' tty='/dev/pts/2'>
- <source path='/dev/pts/2'/>
- <target port='0'/>
- </console>
-
- <serial type='pty'>
- <source path='/dev/pts/2'/>
- <target port='0'/>
- </serial>
-
- <graphics type='vnc' port='-1' autoport='yes' keymap='en-us' listen='0.0.0.0'/>
- </devices>
-</domain>
-EOF
-
-# Mount point for instance fs
-ROOTFS=$VM_DIR/root
-mkdir -p $ROOTFS
-
-# Make sure we have nbd-ness
-modprobe nbd max_part=63
-
-# Which NBD device to use?
-NBD=${NBD:-/dev/nbd5}
-
-# Clean up from previous runs
-umount $ROOTFS || echo 'ok'
-qemu-nbd -d $NBD || echo 'ok'
-
-# Clean up old runs
-cd $VM_DIR
-rm -f $VM_DIR/disk
-
-# Create our instance fs
-qemu-img create -f qcow2 -b $VM_IMAGE disk
-
-# Connect our nbd and wait till it is mountable
-qemu-nbd -c $NBD disk
-NBD_DEV=`basename $NBD`
-if ! timeout 60 sh -c "while ! [ -e /sys/block/$NBD_DEV/pid ]; do sleep 1; done"; then
- echo "Couldn't connect $NBD"
- exit 1
-fi
-
-# Mount the instance
-mount $NBD $ROOTFS -o offset=32256 -t ext4
-
-# Configure instance network
-INTERFACES=$ROOTFS/etc/network/interfaces
-cat > $INTERFACES <<EOF
-auto lo
-iface lo inet loopback
-
-auto eth0
-iface eth0 inet static
- address $GUEST_IP
- netmask $GUEST_NETMASK
- gateway $GUEST_GATEWAY
-EOF
-
-# User configuration for the instance
-chroot $ROOTFS groupadd libvirtd || true
-chroot $ROOTFS useradd stack -s /bin/bash -d $DEST -G libvirtd
-cp -pr $TOOLS_DIR/.. $ROOTFS/$DEST/devstack
-echo "root:$ROOT_PASSWORD" | chroot $ROOTFS chpasswd
-echo "stack:pass" | chroot $ROOTFS chpasswd
-echo "stack ALL=(ALL) NOPASSWD: ALL" >> $ROOTFS/etc/sudoers
-
-# Gracefully cp only if source file/dir exists
-function cp_it {
- if [ -e $1 ] || [ -d $1 ]; then
- cp -pRL $1 $2
- fi
-}
-
-# Copy over your ssh keys and env if desired
-COPYENV=${COPYENV:-1}
-if [ "$COPYENV" = "1" ]; then
- cp_it ~/.ssh $ROOTFS/$DEST/.ssh
- cp_it ~/.ssh/id_rsa.pub $ROOTFS/$DEST/.ssh/authorized_keys
- cp_it ~/.gitconfig $ROOTFS/$DEST/.gitconfig
- cp_it ~/.vimrc $ROOTFS/$DEST/.vimrc
- cp_it ~/.bashrc $ROOTFS/$DEST/.bashrc
-fi
-
-# pre-cache uec images
-for image_url in ${IMAGE_URLS//,/ }; do
- IMAGE_FNAME=`basename "$image_url"`
- if [ ! -f $IMAGES_DIR/$IMAGE_FNAME ]; then
- wget -c $image_url -O $IMAGES_DIR/$IMAGE_FNAME
- fi
- cp $IMAGES_DIR/$IMAGE_FNAME $ROOTFS/$DEST/devstack/files
-done
-
-# Configure the runner
-RUN_SH=$ROOTFS/$DEST/run.sh
-cat > $RUN_SH <<EOF
-#!/usr/bin/env bash
-
-# Kill any existing screens
-killall screen
-
-# Install and run stack.sh
-sudo apt-get update
-sudo apt-get -y --force-yes install git-core vim-nox sudo
-if [ ! -d "$DEST/devstack" ]; then
- git clone git://github.com/cloudbuilders/devstack.git $DEST/devstack
-fi
-cd $DEST/devstack && $STACKSH_PARAMS FORCE=yes ./stack.sh > /$DEST/run.sh.log
-echo >> /$DEST/run.sh.log
-echo >> /$DEST/run.sh.log
-echo "All done! Time to start clicking." >> /$DEST/run.sh.log
-cat $DEST/run.sh.log
-EOF
-chmod 755 $RUN_SH
-
-# Make runner launch on boot
-RC_LOCAL=$ROOTFS/etc/init.d/local
-cat > $RC_LOCAL <<EOF
-#!/bin/sh -e
-# Reboot if this is our first run to enable console log on $DIST_NAME :(
-if [ ! -e /root/firstlaunch ]; then
- touch /root/firstlaunch
- reboot -f
- exit 0
-fi
-su -c "$DEST/run.sh" stack
-EOF
-chmod +x $RC_LOCAL
-chroot $ROOTFS sudo update-rc.d local defaults 80
-
-# Make our ip address hostnames look nice at the command prompt
-echo "export PS1='${debian_chroot:+($debian_chroot)}\\u@\\H:\\w\\$ '" >> $ROOTFS/$DEST/.bashrc
-echo "export PS1='${debian_chroot:+($debian_chroot)}\\u@\\H:\\w\\$ '" >> $ROOTFS/etc/profile
-
-# Give stack ownership over $DEST so it may do the work needed
-chroot $ROOTFS chown -R stack $DEST
-
-# Change boot params so that we get a console log
-sudo sed -e "s/quiet splash/splash console=ttyS0 console=ttyS1,19200n8/g" -i $ROOTFS/boot/grub/menu.lst
-sudo sed -e "s/^hiddenmenu//g" -i $ROOTFS/boot/grub/menu.lst
-
-# Set the hostname
-echo $GUEST_NAME > $ROOTFS/etc/hostname
-
-# We need the hostname to resolve for rabbit to launch
-if ! grep -q $GUEST_NAME $ROOTFS/etc/hosts; then
- echo "$GUEST_IP $GUEST_NAME" >> $ROOTFS/etc/hosts
-fi
-
-# Unmount
-umount $ROOTFS || echo 'ok'
-qemu-nbd -d $NBD
-
-# Create the instance
-cd $VM_DIR && virsh create libvirt.xml
-
-# Tail the console log till we are done
-WAIT_TILL_LAUNCH=${WAIT_TILL_LAUNCH:-1}
-if [ "$WAIT_TILL_LAUNCH" = "1" ]; then
- # Done creating the container, let's tail the log
- echo
- echo "============================================================="
- echo " -- YAY! --"
- echo "============================================================="
- echo
- echo "We're done launching the vm, about to start tailing the"
- echo "stack.sh log. It will take a second or two to start."
- echo
- echo "Just CTRL-C at any time to stop tailing."
-
- while [ ! -e "$VM_DIR/console.log" ]; do
- sleep 1
- done
-
- tail -F $VM_DIR/console.log &
-
- TAIL_PID=$!
-
- function kill_tail() {
- kill $TAIL_PID
- exit 1
- }
-
- # Let Ctrl-c kill tail and exit
- trap kill_tail SIGINT
-
- echo "Waiting stack.sh to finish..."
- while ! cat $VM_DIR/console.log | grep -q 'All done' ; do
- sleep 5
- done
-
- kill $TAIL_PID
-
- if grep -q "stack.sh failed" $VM_DIR/console.log; then
- exit 1
- fi
- echo ""
- echo "Finished - Zip-a-dee Doo-dah!"
-fi
diff --git a/tools/build_libvirt.sh b/tools/build_libvirt.sh
index 5ae8fda..d192879 100755
--- a/tools/build_libvirt.sh
+++ b/tools/build_libvirt.sh
@@ -12,6 +12,27 @@
fi
fi
+# Clean up any resources that may be in use
+cleanup() {
+ set +o errexit
+ unmount_images
+
+ if [ -n "$ROOTFS" ]; then
+ umount $ROOTFS/dev
+ umount $ROOTFS
+ fi
+
+ # Release NBD devices
+ if [ -n "$NBD" ]; then
+ qemu-nbd -d $NBD
+ fi
+
+ # Kill ourselves to signal any calling process
+ trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
# Echo commands
set -o xtrace
@@ -100,9 +121,6 @@
# Install deps if needed
dpkg -l kvm libvirt-bin kpartx || apt-get install -y --force-yes kvm libvirt-bin kpartx
-# Let Ctrl-c kill tail and exit
-trap kill_unmount SIGINT
-
# Where Openstack code will live in image
DEST=${DEST:-/opt/stack}
@@ -127,8 +145,8 @@
# Make sure that base requirements are installed
cp /etc/resolv.conf $COPY_DIR/etc/resolv.conf
chroot $COPY_DIR apt-get update
-chroot $COPY_DIR apt-get install -y --force-yes `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-chroot $COPY_DIR apt-get install -y --download-only rabbitmq-server libvirt-bin mysql-server
+chroot $COPY_DIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+chroot $COPY_DIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
chroot $COPY_DIR pip install `cat files/pips/*`
# Clean out code repos if directed to do so
@@ -156,6 +174,7 @@
# Network configuration variables
GUEST_NETWORK=${GUEST_NETWORK:-1}
+GUEST_RECREATE_NET=${GUEST_RECREATE_NET:-yes}
GUEST_IP=${GUEST_IP:-192.168.$GUEST_NETWORK.50}
GUEST_CIDR=${GUEST_CIDR:-$GUEST_IP/24}
@@ -176,8 +195,10 @@
</network>
EOF
-virsh net-destroy devstack-$GUEST_NETWORK || true
-virsh net-create $VM_DIR/net.xml
+if [[ "$GUEST_RECREATE_NET" == "yes" ]]; then
+ virsh net-destroy devstack-$GUEST_NETWORK || true
+ virsh net-create $VM_DIR/net.xml
+fi
# libvirt.xml configuration
LIBVIRT_XML=$VM_DIR/libvirt.xml
@@ -239,26 +260,35 @@
# Create our instance fs
qemu-img create -f qcow2 -b $VM_IMAGE disk
+# Finds the next available NBD device
+# Exits script if error connecting or none free
+# map_nbd image
+# returns full nbd device path
+function map_nbd {
+ for i in `seq 0 15`; do
+ if [ ! -e /sys/block/nbd$i/pid ]; then
+ NBD=/dev/nbd$i
+ # Connect to nbd and wait till it is ready
+ qemu-nbd -c $NBD $1
+ if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
+ echo "Couldn't connect $NBD"
+ exit 1
+ fi
+ break
+ fi
+ done
+ if [ -z "$NBD" ]; then
+ echo "No free NBD slots"
+ exit 1
+ fi
+ echo $NBD
+}
+
# Make sure we have nbd-ness
modprobe nbd max_part=63
# Set up nbd
-for i in `seq 0 15`; do
- if [ ! -e /sys/block/nbd$i/pid ]; then
- NBD=/dev/nbd$i
- # Connect to nbd and wait till it is ready
- qemu-nbd -c $NBD disk
- if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
- echo "Couldn't connect $NBD"
- exit 1
- fi
- break
- fi
-done
-if [ -z "$NBD" ]; then
- echo "No free NBD slots"
- exit 1
-fi
+NBD=`map_nbd disk`
NBD_DEV=`basename $NBD`
# Mount the instance
@@ -381,7 +411,9 @@
# Unmount
umount $ROOTFS || echo 'ok'
+ROOTFS=""
qemu-nbd -d $NBD
+NBD=""
# Create the instance
cd $VM_DIR && virsh create libvirt.xml
diff --git a/tools/build_lxc.sh b/tools/build_lxc.sh
index a2c5a22..9d8ce92 100755
--- a/tools/build_lxc.sh
+++ b/tools/build_lxc.sh
@@ -125,8 +125,8 @@
# Make sure that base requirements are installed
chroot $CACHEDIR apt-get update
-chroot $CACHEDIR apt-get install -y --force-yes `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-chroot $CACHEDIR apt-get install -y --download-only rabbitmq-server libvirt-bin mysql-server
+chroot $CACHEDIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+chroot $CACHEDIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
chroot $CACHEDIR pip install `cat files/pips/*`
# Clean out code repos if directed to do so
diff --git a/tools/build_nfs.sh b/tools/build_nfs.sh
index 5c591e4..39a2cf0 100755
--- a/tools/build_nfs.sh
+++ b/tools/build_nfs.sh
@@ -32,8 +32,9 @@
# prime natty with as many apt/pips as we can
if [ ! -d $CHROOTCACHE/natty-dev ]; then
rsync -azH $CHROOTCACHE/natty-base/ $CHROOTCACHE/natty-dev/
- chroot $CHROOTCACHE/natty-dev apt-get install -y `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
- chroot $CHROOTCACHE/natty-dev pip install `cat files/pips/*`
+ chroot $CHROOTCACHE apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+ chroot $CHROOTCACHE apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
+ chroot $CHROOTCACHE pip install `cat files/pips/*`
# Create a stack user that is a member of the libvirtd group so that stack
# is able to interact with libvirt.
diff --git a/tools/build_pxe_boot.sh b/tools/build_pxe_boot.sh
index da8bbcc..ab64098 100755
--- a/tools/build_pxe_boot.sh
+++ b/tools/build_pxe_boot.sh
@@ -11,6 +11,22 @@
OPWD=`pwd`
PROGDIR=`dirname $0`
+# Clean up any resources that may be in use
+cleanup() {
+ set +o errexit
+
+ # Mop up temporary files
+ if [ -n "$MNTDIR" -a -d "$MNTDIR" ]; then
+ umount $MNTDIR
+ rmdir $MNTDIR
+ fi
+
+ # Kill ourselves to signal any calling process
+ trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
mkdir -p $DEST_DIR/pxelinux.cfg
cd $DEST_DIR
for i in memdisk menu.c32 pxelinux.0; do
diff --git a/tools/build_ramdisk.sh b/tools/build_ramdisk.sh
index 43f8999..2c914dc 100755
--- a/tools/build_ramdisk.sh
+++ b/tools/build_ramdisk.sh
@@ -1,60 +1,120 @@
#!/bin/bash
# build_ramdisk.sh - Build RAM disk images
+# exit on error to stop unexpected errors
+set -o errexit
+
if [ ! "$#" -eq "1" ]; then
- echo "$0 builds a gziped natty openstack install"
+ echo "$0 builds a gziped Ubuntu OpenStack install"
echo "usage: $0 dest"
exit 1
fi
+# Clean up any resources that may be in use
+cleanup() {
+ set +o errexit
+
+ # Mop up temporary files
+ if [ -n "$MNTDIR" -a -d "$MNTDIR" ]; then
+ umount $MNTDIR
+ rmdir $MNTDIR
+ fi
+ if [ -n "$DEV_FILE_TMP" -a -e "$DEV_FILE_TMP "]; then
+ rm -f $DEV_FILE_TMP
+ fi
+ if [ -n "$IMG_FILE_TMP" -a -e "$IMG_FILE_TMP" ]; then
+ rm -f $IMG_FILE_TMP
+ fi
+
+ # Release NBD devices
+ if [ -n "$NBD" ]; then
+ qemu-nbd -d $NBD
+ fi
+
+ # Kill ourselves to signal any calling process
+ trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
+# Set up nbd
+modprobe nbd max_part=63
+
+# Echo commands
+set -o xtrace
+
IMG_FILE=$1
-PROGDIR=`dirname $0`
-CHROOTCACHE=${CHROOTCACHE:-/var/cache/devstack}
-
-# Source params
-source ./stackrc
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
# Store cwd
CWD=`pwd`
+cd $TOP_DIR
+
+# Source params
+source ./stackrc
+
+CACHEDIR=${CACHEDIR:-/var/cache/devstack}
+
DEST=${DEST:-/opt/stack}
+# Configure the root password of the vm to be the same as ``ADMIN_PASSWORD``
+ROOT_PASSWORD=${ADMIN_PASSWORD:-password}
+
+# Base image (natty by default)
+DIST_NAME=${DIST_NAME:-natty}
+
# Param string to pass to stack.sh. Like "EC2_DMZ_HOST=192.168.1.1 MYSQL_USER=nova"
STACKSH_PARAMS=${STACKSH_PARAMS:-}
# Option to use the version of devstack on which we are currently working
USE_CURRENT_DEVSTACK=${USE_CURRENT_DEVSTACK:-1}
-# Set up nbd
-modprobe nbd max_part=63
-NBD=${NBD:-/dev/nbd9}
-NBD_DEV=`basename $NBD`
-
-# clean install of natty
-if [ ! -r $CHROOTCACHE/natty-base.img ]; then
- $PROGDIR/get_uec_image.sh natty $CHROOTCACHE/natty-base.img
-# # copy kernel modules...
-# # NOTE(ja): is there a better way to do this?
-# cp -pr /lib/modules/`uname -r` $CHROOTCACHE/natty-base/lib/modules
-# # a simple password - pass
-# echo root:pass | chroot $CHROOTCACHE/natty-base chpasswd
+# clean install
+if [ ! -r $CACHEDIR/$DIST_NAME-base.img ]; then
+ $TOOLS_DIR/get_uec_image.sh $DIST_NAME $CACHEDIR/$DIST_NAME-base.img
fi
-# prime natty with as many apt/pips as we can
-if [ ! -r $CHROOTCACHE/natty-dev.img ]; then
- cp -p $CHROOTCACHE/natty-base.img $CHROOTCACHE/natty-dev.img
-
- qemu-nbd -c $NBD $CHROOTCACHE/natty-dev.img
- if ! timeout 60 sh -c "while ! [ -e /sys/block/$NBD_DEV/pid ]; do sleep 1; done"; then
- echo "Couldn't connect $NBD"
+# Finds the next available NBD device
+# Exits script if error connecting or none free
+# map_nbd image
+# returns full nbd device path
+function map_nbd {
+ for i in `seq 0 15`; do
+ if [ ! -e /sys/block/nbd$i/pid ]; then
+ NBD=/dev/nbd$i
+ # Connect to nbd and wait till it is ready
+ qemu-nbd -c $NBD $1
+ if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
+ echo "Couldn't connect $NBD"
+ exit 1
+ fi
+ break
+ fi
+ done
+ if [ -z "$NBD" ]; then
+ echo "No free NBD slots"
exit 1
fi
+ echo $NBD
+}
+
+# prime image with as many apt/pips as we can
+DEV_FILE=$CACHEDIR/$DIST_NAME-dev.img
+DEV_FILE_TMP=`mktemp $DEV_FILE.XXXXXX`
+if [ ! -r $DEV_FILE ]; then
+ cp -p $CACHEDIR/$DIST_NAME-base.img $DEV_FILE_TMP
+
+ NBD=`map_nbd $DEV_FILE_TMP`
MNTDIR=`mktemp -d --tmpdir mntXXXXXXXX`
mount -t ext4 ${NBD}p1 $MNTDIR
cp -p /etc/resolv.conf $MNTDIR/etc/resolv.conf
- chroot $MNTDIR apt-get install -y `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
+ chroot $MNTDIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+ chroot $MNTDIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
chroot $MNTDIR pip install `cat files/pips/*`
# Create a stack user that is a member of the libvirtd group so that stack
@@ -66,6 +126,7 @@
# a simple password - pass
echo stack:pass | chroot $MNTDIR chpasswd
+ echo root:$ROOT_PASSWORD | chroot $MNTDIR chpasswd
# and has sudo ability (in the future this should be limited to only what
# stack requires)
@@ -74,27 +135,31 @@
umount $MNTDIR
rmdir $MNTDIR
qemu-nbd -d $NBD
+ NBD=""
+ mv $DEV_FILE_TMP $DEV_FILE
fi
+rm -f $DEV_FILE_TMP
# clone git repositories onto the system
# ======================================
+IMG_FILE_TMP=`mktemp $IMG_FILE.XXXXXX`
+
if [ ! -r $IMG_FILE ]; then
- qemu-nbd -c $NBD $CHROOTCACHE/natty-dev.img
- if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
- echo "Couldn't connect $NBD"
- exit 1
- fi
+ NBD=`map_nbd $DEV_FILE`
# Pre-create the image file
# FIXME(dt): This should really get the partition size to
# pre-create the image file
- dd if=/dev/zero of=$IMG_FILE bs=1 count=1 seek=$((2*1024*1024*1024))
+ dd if=/dev/zero of=$IMG_FILE_TMP bs=1 count=1 seek=$((2*1024*1024*1024))
# Create filesystem image for RAM disk
- dd if=${NBD}p1 of=$IMG_FILE bs=1M
+ dd if=${NBD}p1 of=$IMG_FILE_TMP bs=1M
qemu-nbd -d $NBD
+ NBD=""
+ mv $IMG_FILE_TMP $IMG_FILE
fi
+rm -f $IMG_FILE_TMP
MNTDIR=`mktemp -d --tmpdir mntXXXXXXXX`
mount -t ext4 -o loop $IMG_FILE $MNTDIR
diff --git a/tools/build_uec.sh b/tools/build_uec.sh
new file mode 100755
index 0000000..d57cb29
--- /dev/null
+++ b/tools/build_uec.sh
@@ -0,0 +1,248 @@
+#!/usr/bin/env bash
+
+# Make sure that we have the proper version of ubuntu (only works on oneiric)
+if ! egrep -q "oneiric" /etc/lsb-release; then
+ echo "This script only works with ubuntu oneiric."
+ exit 1
+fi
+
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+
+cd $TOP_DIR
+
+# Source params
+source ./stackrc
+
+# Ubuntu distro to install
+DIST_NAME=${DIST_NAME:-oneiric}
+
+# Configure how large the VM should be
+GUEST_SIZE=${GUEST_SIZE:-10G}
+
+# exit on error to stop unexpected errors
+set -o errexit
+set -o xtrace
+
+# Abort if localrc is not set
+if [ ! -e $TOP_DIR/localrc ]; then
+ echo "You must have a localrc with ALL necessary passwords defined before proceeding."
+ echo "See stack.sh for required passwords."
+ exit 1
+fi
+
+# Install deps if needed
+DEPS="kvm libvirt-bin kpartx cloud-utils"
+dpkg -l $DEPS || apt-get install -y --force-yes $DEPS
+
+# Where to store files and instances
+WORK_DIR=${WORK_DIR:-/opt/kvmstack}
+
+# Where to store images
+image_dir=$WORK_DIR/images/$DIST_NAME
+mkdir -p $image_dir
+
+# Original version of built image
+uec_url=http://uec-images.ubuntu.com/$DIST_NAME/current/$DIST_NAME-server-cloudimg-amd64.tar.gz
+tarball=$image_dir/$(basename $uec_url)
+
+# download the base uec image if we haven't already
+if [ ! -f $tarball ]; then
+ curl $uec_url -o $tarball
+ (cd $image_dir && tar -Sxvzf $tarball)
+ resize-part-image $image_dir/*.img $GUEST_SIZE $image_dir/disk
+ cp $image_dir/*-vmlinuz-virtual $image_dir/kernel
+fi
+
+
+# Configure the root password of the vm to be the same as ``ADMIN_PASSWORD``
+ROOT_PASSWORD=${ADMIN_PASSWORD:-password}
+
+# Name of our instance, used by libvirt
+GUEST_NAME=${GUEST_NAME:-devstack}
+
+# Mop up after previous runs
+virsh destroy $GUEST_NAME || true
+
+# Where this vm is stored
+vm_dir=$WORK_DIR/instances/$GUEST_NAME
+
+# Create vm dir and remove old disk
+mkdir -p $vm_dir
+rm -f $vm_dir/disk
+
+# Create a copy of the base image
+qemu-img create -f qcow2 -b $image_dir/disk $vm_dir/disk
+
+# Back to devstack
+cd $TOP_DIR
+
+GUEST_NETWORK=${GUEST_NETWORK:-1}
+GUEST_RECREATE_NET=${GUEST_RECREATE_NET:-yes}
+GUEST_IP=${GUEST_IP:-192.168.$GUEST_NETWORK.50}
+GUEST_CIDR=${GUEST_CIDR:-$GUEST_IP/24}
+GUEST_NETMASK=${GUEST_NETMASK:-255.255.255.0}
+GUEST_GATEWAY=${GUEST_GATEWAY:-192.168.$GUEST_NETWORK.1}
+GUEST_MAC=${GUEST_MAC:-"02:16:3e:07:69:`printf '%02X' $GUEST_NETWORK`"}
+GUEST_RAM=${GUEST_RAM:-1524288}
+GUEST_CORES=${GUEST_CORES:-1}
+
+# libvirt.xml configuration
+NET_XML=$vm_dir/net.xml
+cat > $NET_XML <<EOF
+<network>
+ <name>devstack-$GUEST_NETWORK</name>
+ <bridge name="stackbr%d" />
+ <forward/>
+ <ip address="$GUEST_GATEWAY" netmask="$GUEST_NETMASK">
+ <dhcp>
+ <range start='192.168.$GUEST_NETWORK.2' end='192.168.$GUEST_NETWORK.127' />
+ </dhcp>
+ </ip>
+</network>
+EOF
+
+if [[ "$GUEST_RECREATE_NET" == "yes" ]]; then
+ virsh net-destroy devstack-$GUEST_NETWORK || true
+ # destroying the network isn't enough to delete the leases
+ rm -f /var/lib/libvirt/dnsmasq/devstack-$GUEST_NETWORK.leases
+ virsh net-create $vm_dir/net.xml
+fi
+
+# libvirt.xml configuration
+LIBVIRT_XML=$vm_dir/libvirt.xml
+cat > $LIBVIRT_XML <<EOF
+<domain type='kvm'>
+ <name>$GUEST_NAME</name>
+ <memory>$GUEST_RAM</memory>
+ <os>
+ <type>hvm</type>
+ <kernel>$image_dir/kernel</kernel>
+ <cmdline>root=/dev/vda ro console=ttyS0 init=/usr/lib/cloud-init/uncloud-init ds=nocloud-net;s=http://192.168.$GUEST_NETWORK.1:4567/ ubuntu-pass=ubuntu</cmdline>
+ </os>
+ <features>
+ <acpi/>
+ </features>
+ <clock offset='utc'/>
+ <vcpu>$GUEST_CORES</vcpu>
+ <devices>
+ <disk type='file'>
+ <driver type='qcow2'/>
+ <source file='$vm_dir/disk'/>
+ <target dev='vda' bus='virtio'/>
+ </disk>
+
+ <interface type='network'>
+ <source network='devstack-$GUEST_NETWORK'/>
+ </interface>
+
+ <!-- The order is significant here. File must be defined first -->
+ <serial type="file">
+ <source path='$vm_dir/console.log'/>
+ <target port='1'/>
+ </serial>
+
+ <console type='pty' tty='/dev/pts/2'>
+ <source path='/dev/pts/2'/>
+ <target port='0'/>
+ </console>
+
+ <serial type='pty'>
+ <source path='/dev/pts/2'/>
+ <target port='0'/>
+ </serial>
+
+ <graphics type='vnc' port='-1' autoport='yes' keymap='en-us' listen='0.0.0.0'/>
+ </devices>
+</domain>
+EOF
+
+
+rm -rf $vm_dir/uec
+cp -r $TOOLS_DIR/uec $vm_dir/uec
+
+# set metadata
+cat > $vm_dir/uec/meta-data<<EOF
+hostname: $GUEST_NAME
+instance-id: i-hop
+instance-type: m1.ignore
+local-hostname: $GUEST_NAME.local
+EOF
+
+# set metadata
+cat > $vm_dir/uec/user-data<<EOF
+#!/bin/bash
+# hostname needs to resolve for rabbit
+sed -i "s/127.0.0.1/127.0.0.1 \`hostname\`/" /etc/hosts
+apt-get update
+apt-get install git sudo -y
+git clone https://github.com/cloudbuilders/devstack.git
+cd devstack
+git remote set-url origin `cd $TOP_DIR; git remote show origin | grep Fetch | awk '{print $3}'`
+git fetch
+git checkout `git rev-parse HEAD`
+cat > localrc <<LOCAL_EOF
+ROOTSLEEP=0
+`cat $TOP_DIR/localrc`
+LOCAL_EOF
+./stack.sh
+EOF
+
+# (re)start a metadata service
+(
+ pid=`lsof -iTCP@192.168.$GUEST_NETWORK.1:4567 -n | awk '{print $2}' | tail -1`
+ [ -z "$pid" ] || kill -9 $pid
+)
+cd $vm_dir/uec
+python meta.py 192.168.$GUEST_NETWORK.1:4567 &
+
+# Create the instance
+virsh create $vm_dir/libvirt.xml
+
+# Tail the console log till we are done
+WAIT_TILL_LAUNCH=${WAIT_TILL_LAUNCH:-1}
+if [ "$WAIT_TILL_LAUNCH" = "1" ]; then
+ set +o xtrace
+ # Done creating the container, let's tail the log
+ echo
+ echo "============================================================="
+ echo " -- YAY! --"
+ echo "============================================================="
+ echo
+ echo "We're done launching the vm, about to start tailing the"
+ echo "stack.sh log. It will take a second or two to start."
+ echo
+ echo "Just CTRL-C at any time to stop tailing."
+
+ while [ ! -e "$vm_dir/console.log" ]; do
+ sleep 1
+ done
+
+ tail -F $vm_dir/console.log &
+
+ TAIL_PID=$!
+
+ function kill_tail() {
+ kill $TAIL_PID
+ exit 1
+ }
+
+ # Let Ctrl-c kill tail and exit
+ trap kill_tail SIGINT
+
+ echo "Waiting stack.sh to finish..."
+ while ! egrep -q '^stack.sh (completed|failed)' $vm_dir/console.log ; do
+ sleep 1
+ done
+
+ set -o xtrace
+
+ kill $TAIL_PID
+
+ if ! grep -q "^stack.sh completed in" $vm_dir/console.log; then
+ exit 1
+ fi
+ echo ""
+ echo "Finished - Zip-a-dee Doo-dah!"
+fi
diff --git a/tools/build_usb_boot.sh b/tools/build_usb_boot.sh
index fc5e969..e4dabc0 100755
--- a/tools/build_usb_boot.sh
+++ b/tools/build_usb_boot.sh
@@ -11,6 +11,26 @@
OPWD=`pwd`
PROGDIR=`dirname $0`
+# Clean up any resources that may be in use
+cleanup() {
+ set +o errexit
+
+ # Mop up temporary files
+ if [ -n "$DEST_DEV" ]; then
+ umount $DEST_DIR
+ rmdir $DEST_DIR
+ fi
+ if [ -n "$MNTDIR" -a -d "$MNTDIR" ]; then
+ umount $MNTDIR
+ rmdir $MNTDIR
+ fi
+
+ # Kill ourselves to signal any calling process
+ trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
if [ -b $DEST_DIR ]; then
# We have a block device, install syslinux and mount it
DEST_DEV=$DEST_DIR
diff --git a/tools/get_uec_image.sh b/tools/get_uec_image.sh
index 935feba..7b95aab 100755
--- a/tools/get_uec_image.sh
+++ b/tools/get_uec_image.sh
@@ -14,6 +14,9 @@
TOOLS_DIR=$(cd $(dirname "$0") && pwd)
TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+# exit on error to stop unexpected errors
+set -o errexit
+
usage() {
echo "Usage: $0 - Prepare Ubuntu images"
echo ""
@@ -26,6 +29,32 @@
exit 1
}
+# Clean up any resources that may be in use
+cleanup() {
+ set +o errexit
+
+ # Mop up temporary files
+ if [ -n "$IMG_FILE_TMP" -a -e "$IMG_FILE_TMP" ]; then
+ rm -f $IMG_FILE_TMP
+ fi
+
+ # Release NBD devices
+ if [ -n "$NBD" ]; then
+ qemu-nbd -d $NBD
+ fi
+
+ # Kill ourselves to signal any calling process
+ trap 2; kill -2 $$
+}
+
+# apt-get wrapper to just get arguments set correctly
+function apt_get() {
+ local sudo="sudo"
+ [ "$(id -u)" = "0" ] && sudo="env"
+ $sudo DEBIAN_FRONTEND=noninteractive apt-get \
+ --option "Dpkg::Options::=--force-confold" --assume-yes "$@"
+}
+
while getopts f:hmr: c; do
case $c in
f) FORMAT=$OPTARG
@@ -89,11 +118,21 @@
;;
esac
+trap cleanup SIGHUP SIGINT SIGTERM SIGQUIT
+
+# Check for dependencies
+
+if [ ! -x "`which qemu-img`" -o ! -x "`which qemu-nbd`" ]; then
+ # Missing KVM?
+ apt_get install qemu-kvm
+fi
+
# Prepare the base image
# Get the UEC image
UEC_NAME=$DIST_NAME-server-cloudimg-amd64
if [ ! -e $CACHEDIR/$UEC_NAME-disk1.img ]; then
+ mkdir -p $CACHEDIR
(cd $CACHEDIR && wget -N http://uec-images.ubuntu.com/$DIST_NAME/current/$UEC_NAME-disk1.img)
fi
@@ -111,25 +150,33 @@
qemu-img resize $IMG_FILE_TMP +$((ROOTSIZE - 2000))M
fi
+# Finds the next available NBD device
+# Exits script if error connecting or none free
+# map_nbd image
+# returns full nbd device path
+function map_nbd {
+ for i in `seq 0 15`; do
+ if [ ! -e /sys/block/nbd$i/pid ]; then
+ NBD=/dev/nbd$i
+ # Connect to nbd and wait till it is ready
+ qemu-nbd -c $NBD $1
+ if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
+ echo "Couldn't connect $NBD"
+ exit 1
+ fi
+ break
+ fi
+ done
+ if [ -z "$NBD" ]; then
+ echo "No free NBD slots"
+ exit 1
+ fi
+ echo $NBD
+}
+
# Set up nbd
modprobe nbd max_part=63
-for i in `seq 1 15`; do
- if [ ! -e /sys/block/nbd$i/pid ]; then
- NBD=/dev/nbd$i
- # Connect to nbd and wait till it is ready
- qemu-nbd -c $NBD $IMG_FILE_TMP
- if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
- echo "Couldn't connect $NBD"
- exit 1
- fi
- break
- fi
-done
-if [ -z "$NBD" ]; then
- echo "No free NBD slots"
- exit 1
-fi
-NBD_DEV=`basename $NBD`
+NBD=`map_nbd $IMG_FILE_TMP`
# Resize partition 1 to full size of the disk image
echo "d
@@ -162,5 +209,6 @@
umount $MNTDIR
rmdir $MNTDIR
qemu-nbd -d $NBD
+NBD=""
mv $IMG_FILE_TMP $IMG_FILE
diff --git a/tools/setup_stack_user.sh b/tools/setup_stack_user.sh
new file mode 100755
index 0000000..231a20f
--- /dev/null
+++ b/tools/setup_stack_user.sh
@@ -0,0 +1,74 @@
+#!/usr/bin/env bash
+
+# Echo commands
+set -o xtrace
+
+# Exit on error to stop unexpected errors
+set -o errexit
+
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+
+# Change dir to top of devstack
+cd $TOP_DIR
+
+# Echo usage
+usage() {
+ echo "Add stack user and keys"
+ echo ""
+ echo "Usage: $0 [full path to raw uec base image]"
+}
+
+# Make sure this is a raw image
+if ! qemu-img info $1 | grep -q "file format: raw"; then
+ usage
+ exit 1
+fi
+
+# Mount the image
+DEST=/opt/stack
+STAGING_DIR=/tmp/`echo $1 | sed "s/\//_/g"`.stage.user
+mkdir -p $STAGING_DIR
+umount $STAGING_DIR || true
+sleep 1
+mount -t ext4 -o loop $1 $STAGING_DIR
+mkdir -p $STAGING_DIR/$DEST
+
+# Create a stack user that is a member of the libvirtd group so that stack
+# is able to interact with libvirt.
+chroot $STAGING_DIR groupadd libvirtd || true
+chroot $STAGING_DIR useradd stack -s /bin/bash -d $DEST -G libvirtd || true
+
+# Add a simple password - pass
+echo stack:pass | chroot $STAGING_DIR chpasswd
+
+# Configure sudo
+grep -q "^#includedir.*/etc/sudoers.d" $STAGING_DIR/etc/sudoers ||
+ echo "#includedir /etc/sudoers.d" | sudo tee -a $STAGING_DIR/etc/sudoers
+cp $TOP_DIR/files/sudo/* $STAGING_DIR/etc/sudoers.d/
+sed -e "s,%USER%,$USER,g" -i $STAGING_DIR/etc/sudoers.d/*
+
+# and has sudo ability (in the future this should be limited to only what
+# stack requires)
+echo "stack ALL=(ALL) NOPASSWD: ALL" >> $STAGING_DIR/etc/sudoers
+
+# Gracefully cp only if source file/dir exists
+function cp_it {
+ if [ -e $1 ] || [ -d $1 ]; then
+ cp -pRL $1 $2
+ fi
+}
+
+# Copy over your ssh keys and env if desired
+cp_it ~/.ssh $STAGING_DIR/$DEST/.ssh
+cp_it ~/.ssh/id_rsa.pub $STAGING_DIR/$DEST/.ssh/authorized_keys
+cp_it ~/.gitconfig $STAGING_DIR/$DEST/.gitconfig
+cp_it ~/.vimrc $STAGING_DIR/$DEST/.vimrc
+cp_it ~/.bashrc $STAGING_DIR/$DEST/.bashrc
+
+# Give stack ownership over $DEST so it may do the work needed
+chroot $STAGING_DIR chown -R stack $DEST
+
+# Unmount
+umount $STAGING_DIR
diff --git a/tools/uec/meta.py b/tools/uec/meta.py
new file mode 100644
index 0000000..5b845d8
--- /dev/null
+++ b/tools/uec/meta.py
@@ -0,0 +1,29 @@
+import sys
+from BaseHTTPServer import HTTPServer, BaseHTTPRequestHandler
+from SimpleHTTPServer import SimpleHTTPRequestHandler
+
+def main(host, port, HandlerClass = SimpleHTTPRequestHandler,
+ ServerClass = HTTPServer, protocol="HTTP/1.0"):
+ """simple http server that listens on a give address:port"""
+
+ server_address = (host, port)
+
+ HandlerClass.protocol_version = protocol
+ httpd = ServerClass(server_address, HandlerClass)
+
+ sa = httpd.socket.getsockname()
+ print "Serving HTTP on", sa[0], "port", sa[1], "..."
+ httpd.serve_forever()
+
+if __name__ == '__main__':
+ if sys.argv[1:]:
+ address = sys.argv[1]
+ else:
+ address = '0.0.0.0'
+ if ':' in address:
+ host, port = address.split(':')
+ else:
+ host = address
+ port = 8080
+
+ main(host, int(port))
diff --git a/tools/warm_apts_and_pips.sh b/tools/warm_apts_and_pips.sh
new file mode 100755
index 0000000..ec7e916
--- /dev/null
+++ b/tools/warm_apts_and_pips.sh
@@ -0,0 +1,53 @@
+#!/usr/bin/env bash
+
+# Echo commands
+set -o xtrace
+
+# Exit on error to stop unexpected errors
+set -o errexit
+
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+
+# Change dir to top of devstack
+cd $TOP_DIR
+
+# Echo usage
+usage() {
+ echo "Cache OpenStack dependencies on a uec image to speed up performance."
+ echo ""
+ echo "Usage: $0 [full path to raw uec base image]"
+}
+
+# Make sure this is a raw image
+if ! qemu-img info $1 | grep -q "file format: raw"; then
+ usage
+ exit 1
+fi
+
+# Make sure we are in the correct dir
+if [ ! -d files/apts ]; then
+ echo "Please run this script from devstack/tools/"
+ exit 1
+fi
+
+# Mount the image
+STAGING_DIR=/tmp/`echo $1 | sed "s/\//_/g"`.stage
+mkdir -p $STAGING_DIR
+umount $STAGING_DIR || true
+sleep 1
+mount -t ext4 -o loop $1 $STAGING_DIR
+
+# Make sure that base requirements are installed
+cp /etc/resolv.conf $STAGING_DIR/etc/resolv.conf
+
+# Perform caching on the base image to speed up subsequent runs
+chroot $STAGING_DIR apt-get update
+chroot $STAGING_DIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+chroot $STAGING_DIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1` || true
+mkdir -p $STAGING_DIR/var/cache/pip
+PIP_DOWNLOAD_CACHE=/var/cache/pip chroot $STAGING_DIR pip install `cat files/pips/*` || true
+
+# Unmount
+umount $STAGING_DIR
diff --git a/tools/xen/build_domU.sh b/tools/xen/build_domU.sh
index 8e40225..6362849 100755
--- a/tools/xen/build_domU.sh
+++ b/tools/xen/build_domU.sh
@@ -7,6 +7,12 @@
exit 1
fi
+# This directory
+TOP_DIR=$(cd $(dirname "$0") && pwd)
+
+# Source params
+cd ../.. && source ./stackrc && cd $TOP_DIR
+
# Echo commands
set -o xtrace
@@ -41,9 +47,6 @@
# Size of image
VDI_MB=${VDI_MB:-2500}
-# This directory
-TOP_DIR=$(cd $(dirname "$0") && pwd)
-
# Make sure we have git
if ! which git; then
GITDIR=/tmp/git-1.7.7
@@ -223,15 +226,21 @@
SR_UUID=`xe sr-list --minimal name-label="Local storage"`
xe sr-param-set uuid=$SR_UUID other-config:i18n-key=local-storage
-# Uninstall previous runs
-xe vm-list --minimal name-label="$LABEL" | xargs ./scripts/uninstall-os-vpx.sh
-# Destroy any instances that were launched
-for uuid in `xe vm-list | grep -1 instance | grep uuid | sed "s/.*\: //g"`; do
- echo "Shutting down nova instance $uuid"
- xe vm-shutdown uuid=$uuid
- xe vm-destroy uuid=$uuid
-done
+# Shutdown previous runs
+DO_SHUTDOWN=${DO_SHUTDOWN:-1}
+if [ "$DO_SHUTDOWN" = "1" ]; then
+ # Shutdown all domU's that created previously
+ xe vm-list --minimal name-label="$LABEL" | xargs ./scripts/uninstall-os-vpx.sh
+
+ # Destroy any instances that were launched
+ for uuid in `xe vm-list | grep -1 instance | grep uuid | sed "s/.*\: //g"`; do
+ echo "Shutting down nova instance $uuid"
+ xe vm-unpause uuid=$uuid || true
+ xe vm-shutdown uuid=$uuid
+ xe vm-destroy uuid=$uuid
+ done
+fi
# Path to head xva. By default keep overwriting the same one to save space
USE_SEPARATE_XVAS=${USE_SEPARATE_XVAS:-0}
diff --git a/tools/xen/build_domU_multi.sh b/tools/xen/build_domU_multi.sh
new file mode 100755
index 0000000..130bec5
--- /dev/null
+++ b/tools/xen/build_domU_multi.sh
@@ -0,0 +1,35 @@
+#!/usr/bin/env bash
+
+# Echo commands
+set -o xtrace
+
+# Head node host, which runs glance, api, keystone
+HEAD_PUB_IP=${HEAD_PUB_IP:-192.168.1.57}
+HEAD_MGT_IP=${HEAD_MGT_IP:-172.16.100.57}
+
+COMPUTE_PUB_IP=${COMPUTE_PUB_IP:-192.168.1.58}
+COMPUTE_MGT_IP=${COMPUTE_MGT_IP:-172.16.100.58}
+
+# Networking params
+FLOATING_RANGE=${FLOATING_RANGE:-192.168.1.196/30}
+
+# Variables common amongst all hosts in the cluster
+COMMON_VARS="$STACKSH_PARAMS MYSQL_HOST=$HEAD_MGT_IP RABBIT_HOST=$HEAD_MGT_IP GLANCE_HOSTPORT=$HEAD_MGT_IP:9292 FLOATING_RANGE=$FLOATING_RANGE"
+
+# Helper to launch containers
+function build_domU {
+ GUEST_NAME=$1 PUB_IP=$2 MGT_IP=$3 DO_SHUTDOWN=$4 TERMINATE=$TERMINATE STACKSH_PARAMS="$COMMON_VARS $5" ./build_domU.sh
+}
+
+# Launch the head node - headnode uses a non-ip domain name,
+# because rabbit won't launch with an ip addr hostname :(
+build_domU HEADNODE $HEAD_PUB_IP $HEAD_MGT_IP 1 "ENABLED_SERVICES=g-api,g-reg,key,n-api,n-sch,n-vnc,horizon,mysql,rabbit"
+
+# Wait till the head node is up
+while ! curl -L http://$HEAD_PUB_IP | grep -q username; do
+ echo "Waiting for head node ($HEAD_PUB_IP) to start..."
+ sleep 5
+done
+
+# Build the HA compute host
+build_domU $COMPUTE_PUB_IP $COMPUTE_PUB_IP $COMPUTE_MGT_IP 0 "ENABLED_SERVICES=n-cpu,n-net,n-api"