merge origin/master
diff --git a/.gitignore b/.gitignore
index 22a7898..e482090 100644
--- a/.gitignore
+++ b/.gitignore
@@ -1,3 +1,5 @@
 proto
 *~
+*.log
+src
 localrc
diff --git a/exercises/floating_ips.sh b/exercises/floating_ips.sh
old mode 100755
new mode 100644
index 5c38430..75046d1
--- a/exercises/floating_ips.sh
+++ b/exercises/floating_ips.sh
@@ -84,6 +84,15 @@
 # Waiting for boot
 # ----------------
 
+# Max time to wait while vm goes from build to active state
+ACTIVE_TIMEOUT=${ACTIVE_TIMEOUT:-10}
+
+# Max time till the vm is bootable
+BOOT_TIMEOUT=${BOOT_TIMEOUT:-15}
+
+# Max time to wait for proper association and dis-association.
+ASSOCIATE_TIMEOUT=${ASSOCIATE_TIMEOUT:-10}
+
 # check that the status is active within ACTIVE_TIMEOUT seconds
 if ! timeout $ACTIVE_TIMEOUT sh -c "while ! nova show $NAME | grep status | grep -q ACTIVE; do sleep 1; done"; then
     echo "server didn't become active!"
@@ -158,7 +167,7 @@
 nova secgroup-delete-rule $SECGROUP icmp -1 -1 0.0.0.0/0
 
 # FIXME (anthony): make xs support security groups
-if [ "$VIRT_DRIVER" != "xenserver"]; then
+if [ "$VIRT_DRIVER" != "xenserver" ]; then
     # test we can aren't able to ping our floating ip within ASSOCIATE_TIMEOUT seconds
     if ! timeout $ASSOCIATE_TIMEOUT sh -c "while ping -c1 -w1 $FLOATING_IP; do sleep 1; done"; then
         print "Security group failure - ping should not be allowed!"
diff --git a/files/apts/nova b/files/apts/nova
index 8ae74a2..9eefed7 100644
--- a/files/apts/nova
+++ b/files/apts/nova
@@ -1,7 +1,9 @@
 dnsmasq-base
+dnsmasq-utils # for dhcp_release
 kpartx
 parted
-mysql-server
+arping # used for send_arp_for_ha option in nova-network
+mysql-server # NOPRIME
 python-mysqldb
 kvm
 gawk
@@ -10,10 +12,10 @@
 sqlite3
 sudo
 kvm
-libvirt-bin
+libvirt-bin # NOPRIME
 vlan
 curl
-rabbitmq-server
+rabbitmq-server # NOPRIME
 socat # used by ajaxterm
 python-mox
 python-paste
diff --git a/files/apts/preseed b/files/apts/preseed
deleted file mode 100644
index 8712d5d..0000000
--- a/files/apts/preseed
+++ /dev/null
@@ -1,18 +0,0 @@
-# a collection of packages that speed up installation as they are dependencies
-# of packages we can't install during bootstraping (rabbitmq-server, 
-# mysql-server, libvirt-bin)
-#
-# NOTE: only add packages to this file that aren't needed directly
-mysql-common
-mysql-client-5.1
-erlang-base
-erlang-ssl 
-erlang-nox
-erlang-inets
-erlang-mnesia
-libhtml-template-perl
-gettext-base
-libavahi-client3
-libxml2-utils
-libpciaccess0
-libparted0debian1
diff --git a/files/apts/swift b/files/apts/swift
new file mode 100644
index 0000000..f298377
--- /dev/null
+++ b/files/apts/swift
@@ -0,0 +1,17 @@
+curl
+gcc
+memcached # NOPRIME
+python-configobj
+python-coverage
+python-dev
+python-eventlet
+python-greenlet
+python-netifaces
+python-nose
+python-pastedeploy
+python-setuptools
+python-simplejson
+python-webob
+python-xattr
+sqlite3
+xfsprogs
diff --git a/files/glance-api.conf b/files/glance-api.conf
index 3499ff7..bb758af 100644
--- a/files/glance-api.conf
+++ b/files/glance-api.conf
@@ -24,7 +24,7 @@
 
 # Log to this file. Make sure you do not set the same log
 # file for both the API and registry servers!
-log_file = %DEST%/glance/api.log
+#log_file = %DEST%/glance/api.log
 
 # Send logs to syslog (/dev/log) instead of to file specified by `log_file`
 use_syslog = %SYSLOG%
diff --git a/files/glance-registry.conf b/files/glance-registry.conf
index 351b09f..1e04186 100644
--- a/files/glance-registry.conf
+++ b/files/glance-registry.conf
@@ -13,7 +13,7 @@
 
 # Log to this file. Make sure you do not set the same log
 # file for both the API and registry servers!
-log_file = %DEST%/glance/registry.log
+#log_file = %DEST%/glance/registry.log
 
 # Where to store images
 filesystem_store_datadir = %DEST%/glance/images
diff --git a/files/keystone_data.sh b/files/keystone_data.sh
index a5e75a6..d926c52 100755
--- a/files/keystone_data.sh
+++ b/files/keystone_data.sh
@@ -30,12 +30,13 @@
 $BIN_DIR/keystone-manage $* service add nova compute "Nova Compute Service"
 $BIN_DIR/keystone-manage $* service add glance image "Glance Image Service"
 $BIN_DIR/keystone-manage $* service add keystone identity "Keystone Identity Service"
+$BIN_DIR/keystone-manage $* service add swift object-store "Swift Service"
 
 #endpointTemplates
 $BIN_DIR/keystone-manage $* endpointTemplates add RegionOne nova http://%HOST_IP%:8774/v1.1/%tenant_id% http://%HOST_IP%:8774/v1.1/%tenant_id%  http://%HOST_IP%:8774/v1.1/%tenant_id% 1 1
 $BIN_DIR/keystone-manage $* endpointTemplates add RegionOne glance http://%HOST_IP%:9292/v1.1/%tenant_id% http://%HOST_IP%:9292/v1.1/%tenant_id% http://%HOST_IP%:9292/v1.1/%tenant_id% 1 1
 $BIN_DIR/keystone-manage $* endpointTemplates add RegionOne keystone http://%HOST_IP%:5000/v2.0 http://%HOST_IP%:35357/v2.0 http://%HOST_IP%:5000/v2.0 1 1
-# $BIN_DIR/keystone-manage $* endpointTemplates add RegionOne swift http://%HOST_IP%:8080/v1/AUTH_%tenant_id% http://%HOST_IP%:8080/ http://%HOST_IP%:8080/v1/AUTH_%tenant_id% 1 1
+$BIN_DIR/keystone-manage $* endpointTemplates add RegionOne swift http://%HOST_IP%:8080/v1/AUTH_%tenant_id% http://%HOST_IP%:8080/ http://%HOST_IP%:8080/v1/AUTH_%tenant_id% 1 1
 
 # Tokens
 $BIN_DIR/keystone-manage $* token add %SERVICE_TOKEN% admin admin 2015-02-05T00:00
diff --git a/files/nova-api-paste.ini b/files/nova-api-paste.ini
new file mode 100644
index 0000000..2c642f8
--- /dev/null
+++ b/files/nova-api-paste.ini
@@ -0,0 +1,127 @@
+#######
+# EC2 #
+#######
+
+[composite:ec2]
+use = egg:Paste#urlmap
+/: ec2versions
+/services/Cloud: ec2cloud
+/services/Admin: ec2admin
+/latest: ec2metadata
+/2007-01-19: ec2metadata
+/2007-03-01: ec2metadata
+/2007-08-29: ec2metadata
+/2007-10-10: ec2metadata
+/2007-12-15: ec2metadata
+/2008-02-01: ec2metadata
+/2008-09-01: ec2metadata
+/2009-04-04: ec2metadata
+/1.0: ec2metadata
+
+[pipeline:ec2cloud]
+pipeline = logrequest totoken authtoken keystonecontext cloudrequest authorizer ec2executor
+
+[pipeline:ec2admin]
+pipeline = logrequest totoken authtoken keystonecontext adminrequest authorizer ec2executor
+
+[pipeline:ec2metadata]
+pipeline = logrequest ec2md
+
+[pipeline:ec2versions]
+pipeline = logrequest ec2ver
+
+[filter:logrequest]
+paste.filter_factory = nova.api.ec2:RequestLogging.factory
+
+[filter:ec2lockout]
+paste.filter_factory = nova.api.ec2:Lockout.factory
+
+[filter:totoken]
+paste.filter_factory = keystone.middleware.ec2_token:EC2Token.factory
+
+[filter:ec2noauth]
+paste.filter_factory = nova.api.ec2:NoAuth.factory
+
+[filter:authenticate]
+paste.filter_factory = nova.api.ec2:Authenticate.factory
+
+[filter:cloudrequest]
+controller = nova.api.ec2.cloud.CloudController
+paste.filter_factory = nova.api.ec2:Requestify.factory
+
+[filter:adminrequest]
+controller = nova.api.ec2.admin.AdminController
+paste.filter_factory = nova.api.ec2:Requestify.factory
+
+[filter:authorizer]
+paste.filter_factory = nova.api.ec2:Authorizer.factory
+
+[app:ec2executor]
+paste.app_factory = nova.api.ec2:Executor.factory
+
+[app:ec2ver]
+paste.app_factory = nova.api.ec2:Versions.factory
+
+[app:ec2md]
+paste.app_factory = nova.api.ec2.metadatarequesthandler:MetadataRequestHandler.factory
+
+#############
+# Openstack #
+#############
+
+[composite:osapi]
+use = egg:Paste#urlmap
+/: osversions
+/v1.0: openstackapi10
+/v1.1: openstackapi11
+
+[pipeline:openstackapi10]
+pipeline = faultwrap authtoken keystonecontext ratelimit osapiapp10
+
+[pipeline:openstackapi11]
+pipeline = faultwrap authtoken keystonecontext ratelimit extensions osapiapp11
+
+[filter:faultwrap]
+paste.filter_factory = nova.api.openstack:FaultWrapper.factory
+
+[filter:auth]
+paste.filter_factory = nova.api.openstack.auth:AuthMiddleware.factory
+
+[filter:noauth]
+paste.filter_factory = nova.api.openstack.auth:NoAuthMiddleware.factory
+
+[filter:ratelimit]
+paste.filter_factory = nova.api.openstack.limits:RateLimitingMiddleware.factory
+
+[filter:extensions]
+paste.filter_factory = nova.api.openstack.extensions:ExtensionMiddleware.factory
+
+[app:osapiapp10]
+paste.app_factory = nova.api.openstack:APIRouterV10.factory
+
+[app:osapiapp11]
+paste.app_factory = nova.api.openstack:APIRouterV11.factory
+
+[pipeline:osversions]
+pipeline = faultwrap osversionapp
+
+[app:osversionapp]
+paste.app_factory = nova.api.openstack.versions:Versions.factory
+
+##########
+# Shared #
+##########
+
+[filter:keystonecontext]
+paste.filter_factory = keystone.middleware.nova_keystone_context:NovaKeystoneContext.factory
+
+[filter:authtoken]
+paste.filter_factory = keystone.middleware.auth_token:filter_factory
+service_protocol = http
+service_host = 127.0.0.1
+service_port = 5000
+auth_host = 127.0.0.1
+auth_port = 35357
+auth_protocol = http
+auth_uri = http://127.0.0.1:5000/
+admin_token = %SERVICE_TOKEN%
diff --git a/files/sudo/nova b/files/sudo/nova
index 2ce1aac..62685b3 100644
--- a/files/sudo/nova
+++ b/files/sudo/nova
@@ -41,7 +41,7 @@
                       /usr/bin/socat,                           \
                       /sbin/parted,                             \
                       /usr/sbin/dnsmasq,                        \
-                      /usr/bin/arping
+                      /usr/sbin/arping
 
 %USER% ALL = (root) NOPASSWD: SETENV: NOVACMDS
 
diff --git a/files/swift/account-server.conf b/files/swift/account-server.conf
new file mode 100644
index 0000000..db0f097
--- /dev/null
+++ b/files/swift/account-server.conf
@@ -0,0 +1,20 @@
+[DEFAULT]
+devices = %NODE_PATH%/node
+mount_check = false
+bind_port = %BIND_PORT%
+user = %USER%
+log_facility = LOG_LOCAL%LOG_FACILITY%
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = account-server
+
+[app:account-server]
+use = egg:swift#account
+
+[account-replicator]
+vm_test_mode = yes
+
+[account-auditor]
+
+[account-reaper]
diff --git a/files/swift/container-server.conf b/files/swift/container-server.conf
new file mode 100644
index 0000000..bdc3e3a
--- /dev/null
+++ b/files/swift/container-server.conf
@@ -0,0 +1,22 @@
+[DEFAULT]
+devices = %NODE_PATH%/node
+mount_check = false
+bind_port = %BIND_PORT%
+user = %USER%
+log_facility = LOG_LOCAL%LOG_FACILITY%
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = container-server
+
+[app:container-server]
+use = egg:swift#container
+
+[container-replicator]
+vm_test_mode = yes
+
+[container-updater]
+
+[container-auditor]
+
+[container-sync]
diff --git a/files/swift/object-server.conf b/files/swift/object-server.conf
new file mode 100644
index 0000000..06fbffe
--- /dev/null
+++ b/files/swift/object-server.conf
@@ -0,0 +1,20 @@
+[DEFAULT]
+devices = %NODE_PATH%/node
+mount_check = false
+bind_port = %BIND_PORT%
+user = %USER%
+log_facility = LOG_LOCAL%LOG_FACILITY%
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = object-server
+
+[app:object-server]
+use = egg:swift#object
+
+[object-replicator]
+vm_test_mode = yes
+
+[object-updater]
+
+[object-auditor]
diff --git a/files/swift/proxy-server.conf b/files/swift/proxy-server.conf
new file mode 100644
index 0000000..fe7e39b
--- /dev/null
+++ b/files/swift/proxy-server.conf
@@ -0,0 +1,32 @@
+[DEFAULT]
+bind_port = 8080
+user = %USER%
+log_facility = LOG_LOCAL1
+swift_dir = %SWIFT_CONFIG_LOCATION%
+
+[pipeline:main]
+pipeline = healthcheck cache %AUTH_SERVER% proxy-server
+
+[app:proxy-server]
+use = egg:swift#proxy
+allow_account_management = true
+account_autocreate = true
+
+[filter:keystone]
+use = egg:swiftkeystone2#keystone2
+keystone_admin_token = %SERVICE_TOKEN%
+keystone_url = http://localhost:35357/v2.0
+
+[filter:tempauth]
+use = egg:swift#tempauth
+user_admin_admin = admin .admin .reseller_admin
+user_test_tester = testing .admin
+user_test2_tester2 = testing2 .admin
+user_test_tester3 = testing3
+bind_ip = 0.0.0.0
+
+[filter:healthcheck]
+use = egg:swift#healthcheck
+
+[filter:cache]
+use = egg:swift#memcache
diff --git a/files/swift/rsyncd.conf b/files/swift/rsyncd.conf
new file mode 100644
index 0000000..66215c7
--- /dev/null
+++ b/files/swift/rsyncd.conf
@@ -0,0 +1,79 @@
+uid = %USER%
+gid = %GROUP%
+log file = /var/log/rsyncd.log
+pid file = /var/run/rsyncd.pid
+address = 127.0.0.1
+
+[account6012]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/1/node/
+read only = false
+lock file = /var/lock/account6012.lock
+
+[account6022]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/2/node/
+read only = false
+lock file = /var/lock/account6022.lock
+
+[account6032]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/3/node/
+read only = false
+lock file = /var/lock/account6032.lock
+
+[account6042]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/4/node/
+read only = false
+lock file = /var/lock/account6042.lock
+
+
+[container6011]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/1/node/
+read only = false
+lock file = /var/lock/container6011.lock
+
+[container6021]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/2/node/
+read only = false
+lock file = /var/lock/container6021.lock
+
+[container6031]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/3/node/
+read only = false
+lock file = /var/lock/container6031.lock
+
+[container6041]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/4/node/
+read only = false
+lock file = /var/lock/container6041.lock
+
+
+[object6010]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/1/node/
+read only = false
+lock file = /var/lock/object6010.lock
+
+[object6020]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/2/node/
+read only = false
+lock file = /var/lock/object6020.lock
+
+[object6030]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/3/node/
+read only = false
+lock file = /var/lock/object6030.lock
+
+[object6040]
+max connections = 25
+path = %SWIFT_DATA_LOCATION%/4/node/
+read only = false
+lock file = /var/lock/object6040.lock
diff --git a/files/swift/swift-remakerings b/files/swift/swift-remakerings
new file mode 100755
index 0000000..c65353c
--- /dev/null
+++ b/files/swift/swift-remakerings
@@ -0,0 +1,26 @@
+#!/bin/bash
+
+cd %SWIFT_CONFIG_LOCATION%
+
+rm -f *.builder *.ring.gz backups/*.builder backups/*.ring.gz
+
+swift-ring-builder object.builder create %SWIFT_PARTITION_POWER_SIZE% 3 1
+swift-ring-builder object.builder add z1-127.0.0.1:6010/sdb1 1
+swift-ring-builder object.builder add z2-127.0.0.1:6020/sdb2 1
+swift-ring-builder object.builder add z3-127.0.0.1:6030/sdb3 1
+swift-ring-builder object.builder add z4-127.0.0.1:6040/sdb4 1
+swift-ring-builder object.builder rebalance
+
+swift-ring-builder container.builder create %SWIFT_PARTITION_POWER_SIZE% 3 1
+swift-ring-builder container.builder add z1-127.0.0.1:6011/sdb1 1
+swift-ring-builder container.builder add z2-127.0.0.1:6021/sdb2 1
+swift-ring-builder container.builder add z3-127.0.0.1:6031/sdb3 1
+swift-ring-builder container.builder add z4-127.0.0.1:6041/sdb4 1
+swift-ring-builder container.builder rebalance
+
+swift-ring-builder account.builder create %SWIFT_PARTITION_POWER_SIZE% 3 1
+swift-ring-builder account.builder add z1-127.0.0.1:6012/sdb1 1
+swift-ring-builder account.builder add z2-127.0.0.1:6022/sdb2 1
+swift-ring-builder account.builder add z3-127.0.0.1:6032/sdb3 1
+swift-ring-builder account.builder add z4-127.0.0.1:6042/sdb4 1
+swift-ring-builder account.builder rebalance
diff --git a/files/swift/swift-startmain b/files/swift/swift-startmain
new file mode 100755
index 0000000..69efebd
--- /dev/null
+++ b/files/swift/swift-startmain
@@ -0,0 +1,3 @@
+#!/bin/bash
+
+swift-init all restart
diff --git a/files/swift/swift.conf b/files/swift/swift.conf
new file mode 100644
index 0000000..98df466
--- /dev/null
+++ b/files/swift/swift.conf
@@ -0,0 +1,3 @@
+[swift-hash]
+# random unique string that can never change (DO NOT LOSE)
+swift_hash_path_suffix = %SWIFT_HASH%
diff --git a/stack.sh b/stack.sh
index 9466512..78851b9 100755
--- a/stack.sh
+++ b/stack.sh
@@ -70,7 +70,7 @@
 # called ``localrc``
 #
 # If ``localrc`` exists, then ``stackrc`` will load those settings.  This is
-# useful for changing a branch or repostiory to test other versions.  Also you
+# useful for changing a branch or repository to test other versions.  Also you
 # can store your other settings like **MYSQL_PASSWORD** or **ADMIN_PASSWORD** instead
 # of letting devstack generate random ones for you.
 source ./stackrc
@@ -103,8 +103,7 @@
 
     # since this script runs as a normal user, we need to give that user
     # ability to run sudo
-    apt_get update
-    apt_get install sudo
+    dpkg -l sudo || apt_get update && apt_get install sudo
 
     if ! getent passwd stack >/dev/null; then
         echo "Creating a user called stack"
@@ -150,6 +149,12 @@
 NOVACLIENT_DIR=$DEST/python-novaclient
 OPENSTACKX_DIR=$DEST/openstackx
 NOVNC_DIR=$DEST/noVNC
+SWIFT_DIR=$DEST/swift
+SWIFT_KEYSTONE_DIR=$DEST/swift-keystone2
+QUANTUM_DIR=$DEST/quantum
+
+# Default Quantum Plugin
+Q_PLUGIN=${Q_PLUGIN:-openvswitch}
 
 # Specify which services to launch.  These generally correspond to screen tabs
 ENABLED_SERVICES=${ENABLED_SERVICES:-g-api,g-reg,key,n-api,n-cpu,n-net,n-sch,n-vnc,horizon,mysql,rabbit}
@@ -169,6 +174,9 @@
     HOST_IP=`LC_ALL=C /sbin/ifconfig  | grep -m 1 'inet addr:'| cut -d: -f2 | awk '{print $1}'`
 fi
 
+# Service startup timeout
+SERVICE_TIMEOUT=${SERVICE_TIMEOUT:-60}
+
 # Generic helper to configure passwords
 function read_password {
     set +o xtrace
@@ -224,7 +232,7 @@
 # Multi-host is a mode where each compute node runs its own network node.  This
 # allows network operations and routing for a VM to occur on the server that is
 # running the VM - removing a SPOF and bandwidth bottleneck.
-MULTI_HOST=${MULTI_HOST:-0}
+MULTI_HOST=${MULTI_HOST:-False}
 
 # If you are using FlatDHCP on multiple hosts, set the ``FLAT_INTERFACE``
 # variable but make sure that the interface doesn't already have an
@@ -239,11 +247,22 @@
 # If you are running on a single node and don't need to access the VMs from
 # devices other than that node, you can set the flat interface to the same
 # value as ``FLAT_NETWORK_BRIDGE``.  This will stop the network hiccup from
-# occuring.
+# occurring.
 FLAT_INTERFACE=${FLAT_INTERFACE:-eth0}
 
 ## FIXME(ja): should/can we check that FLAT_INTERFACE is sane?
 
+# Using Quantum networking:
+#
+# Make sure that q-svc is enabled in ENABLED_SERVICES.  If it is the network
+# manager will be set to the QuantumManager.
+#
+# If you're planning to use the Quantum openvswitch plugin, set Q_PLUGIN to
+# "openvswitch" and make sure the q-agt service is enabled in
+# ENABLED_SERVICES.
+#
+# With Quantum networking the NET_MAN variable is ignored.
+
 
 # MySQL & RabbitMQ
 # ----------------
@@ -270,6 +289,42 @@
 # Glance connection info.  Note the port must be specified.
 GLANCE_HOSTPORT=${GLANCE_HOSTPORT:-$HOST_IP:9292}
 
+# SWIFT
+# -----
+# TODO: implement glance support
+# TODO: add logging to different location.
+
+# By default the location of swift drives and objects is located inside
+# the swift source directory. SWIFT_DATA_LOCATION variable allow you to redefine
+# this.
+SWIFT_DATA_LOCATION=${SWIFT_DATA_LOCATION:-${SWIFT_DIR}/data}
+
+# We are going to have the configuration files inside the source
+# directory, change SWIFT_CONFIG_LOCATION if you want to adjust that.
+SWIFT_CONFIG_LOCATION=${SWIFT_CONFIG_LOCATION:-${SWIFT_DIR}/config}
+
+# devstack will create a loop-back disk formatted as XFS to store the
+# swift data. By default the disk size is 1 gigabyte. The variable
+# SWIFT_LOOPBACK_DISK_SIZE specified in bytes allow you to change
+# that.
+SWIFT_LOOPBACK_DISK_SIZE=${SWIFT_LOOPBACK_DISK_SIZE:-1000000}
+
+# The ring uses a configurable number of bits from a path’s MD5 hash as
+# a partition index that designates a device. The number of bits kept
+# from the hash is known as the partition power, and 2 to the partition
+# power indicates the partition count. Partitioning the full MD5 hash
+# ring allows other parts of the cluster to work in batches of items at
+# once which ends up either more efficient or at least less complex than
+# working with each item separately or the entire cluster all at once.
+# By default we define 9 for the partition count (which mean 512).
+SWIFT_PARTITION_POWER_SIZE=${SWIFT_PARTITION_POWER_SIZE:-9}
+
+# We only ask for Swift Hash if we have enabled swift service.
+if [[ "$ENABLED_SERVICES" =~ "swift" ]]; then
+    # SWIFT_HASH is a random unique string for a swift cluster that
+    # can never change.
+    read_password SWIFT_HASH "ENTER A RANDOM SWIFT HASH."
+fi
 
 # Keystone
 # --------
@@ -283,7 +338,7 @@
 LOGFILE=${LOGFILE:-"$PWD/stack.sh.$$.log"}
 (
 # So that errors don't compound we exit on any errors so you see only the
-# first error that occured.
+# first error that occurred.
 trap failed ERR
 failed() {
     local r=$?
@@ -310,7 +365,7 @@
 
 # install apt requirements
 apt_get update
-apt_get install `cat $FILES/apts/* | cut -d\# -f1 | grep -Ev "mysql-server|rabbitmq-server"`
+apt_get install `cat $FILES/apts/* | cut -d\# -f1 | grep -Ev "mysql-server|rabbitmq-server|memcached"`
 
 # install python requirements
 sudo PIP_DOWNLOAD_CACHE=/var/cache/pip pip install `cat $FILES/pips/*`
@@ -349,6 +404,10 @@
 
 # compute service
 git_clone $NOVA_REPO $NOVA_DIR $NOVA_BRANCH
+# storage service
+git_clone $SWIFT_REPO $SWIFT_DIR $SWIFT_BRANCH
+# swift + keystone middleware
+git_clone $SWIFT_KEYSTONE_REPO $SWIFT_KEYSTONE_DIR $SWIFT_KEYSTONE_BRANCH
 # image catalog service
 git_clone $GLANCE_REPO $GLANCE_DIR $GLANCE_BRANCH
 # unified auth system (manages accounts/tokens)
@@ -362,6 +421,8 @@
 # openstackx is a collection of extensions to openstack.compute & nova
 # that is *deprecated*.  The code is being moved into python-novaclient & nova.
 git_clone $OPENSTACKX_REPO $OPENSTACKX_DIR $OPENSTACKX_BRANCH
+# quantum
+git_clone $QUANTUM_REPO $QUANTUM_DIR $QUANTUM_BRANCH
 
 # Initialization
 # ==============
@@ -370,12 +431,15 @@
 # setup our checkouts so they are installed into python path
 # allowing ``import nova`` or ``import glance.client``
 cd $KEYSTONE_DIR; sudo python setup.py develop
+cd $SWIFT_DIR; sudo python setup.py develop
+cd $SWIFT_KEYSTONE_DIR; sudo python setup.py develop
 cd $GLANCE_DIR; sudo python setup.py develop
 cd $NOVACLIENT_DIR; sudo python setup.py develop
 cd $NOVA_DIR; sudo python setup.py develop
 cd $OPENSTACKX_DIR; sudo python setup.py develop
 cd $HORIZON_DIR/django-openstack; sudo python setup.py develop
 cd $HORIZON_DIR/openstack-dashboard; sudo python setup.py develop
+cd $QUANTUM_DIR; sudo python setup.py develop
 
 # Add a useful screenrc.  This isn't required to run openstack but is we do
 # it since we are going to run the services in screen for simple
@@ -500,13 +564,12 @@
 # ----
 
 if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then
-    # We are going to use the sample http middleware configuration from the
-    # keystone project to launch nova.  This paste config adds the configuration
-    # required for nova to validate keystone tokens - except we need to switch
-    # the config to use our service token instead (instead of the invalid token
-    # 999888777666).
-    cp $KEYSTONE_DIR/examples/paste/nova-api-paste.ini $NOVA_DIR/bin
-    sed -e "s,999888777666,$SERVICE_TOKEN,g" -i $NOVA_DIR/bin/nova-api-paste.ini
+    # We are going to use a sample http middleware configuration based on the
+    # one from the keystone project to launch nova.  This paste config adds
+    # the configuration required for nova to validate keystone tokens. We add
+    # our own service token to the configuration.
+    cp $FILES/nova-api-paste.ini $NOVA_DIR/bin
+    sed -e "s,%SERVICE_TOKEN%,$SERVICE_TOKEN,g" -i $NOVA_DIR/bin/nova-api-paste.ini
 fi
 
 if [[ "$ENABLED_SERVICES" =~ "n-cpu" ]]; then
@@ -580,6 +643,129 @@
     mkdir -p $NOVA_DIR/networks
 fi
 
+# Storage Service
+if [[ "$ENABLED_SERVICES" =~ "swift" ]]; then
+    # We first do a bit of setup by creating the directories and
+    # changing the permissions so we can run it as our user.
+
+    USER_GROUP=$(id -g)
+    sudo mkdir -p ${SWIFT_DATA_LOCATION}/drives
+    sudo chown -R $USER:${USER_GROUP} ${SWIFT_DATA_LOCATION}/drives
+
+    # We then create a loopback disk and format it to XFS.
+    if [[ ! -e ${SWIFT_DATA_LOCATION}/drives/images/swift.img ]];then
+        mkdir -p  ${SWIFT_DATA_LOCATION}/drives/images
+        sudo touch  ${SWIFT_DATA_LOCATION}/drives/images/swift.img
+        sudo chown $USER: ${SWIFT_DATA_LOCATION}/drives/images/swift.img
+
+        dd if=/dev/zero of=${SWIFT_DATA_LOCATION}/drives/images/swift.img \
+            bs=1024 count=0 seek=${SWIFT_LOOPBACK_DISK_SIZE}
+        mkfs.xfs -f -i size=1024  ${SWIFT_DATA_LOCATION}/drives/images/swift.img
+    fi
+
+    # After the drive being created we mount the disk with a few mount
+    # options to make it most efficient as possible for swift.
+    mkdir -p ${SWIFT_DATA_LOCATION}/drives/sdb1
+    if ! egrep -q ${SWIFT_DATA_LOCATION}/drives/sdb1 /proc/mounts;then
+        sudo mount -t xfs -o loop,noatime,nodiratime,nobarrier,logbufs=8  \
+            ${SWIFT_DATA_LOCATION}/drives/images/swift.img ${SWIFT_DATA_LOCATION}/drives/sdb1
+    fi
+
+    # We then create link to that mounted location so swift would know
+    # where to go.
+    for x in {1..4}; do sudo ln -sf ${SWIFT_DATA_LOCATION}/drives/sdb1/$x ${SWIFT_DATA_LOCATION}/$x; done
+
+    # We now have to emulate a few different servers into one we
+    # create all the directories needed for swift
+    tmpd=""
+    for d in ${SWIFT_DATA_LOCATION}/drives/sdb1/{1..4} \
+        ${SWIFT_CONFIG_LOCATION}/{object,container,account}-server \
+        ${SWIFT_DATA_LOCATION}/{1..4}/node/sdb1 /var/run/swift ;do
+        [[ -d $d ]] && continue
+        sudo install -o ${USER} -g $USER_GROUP -d $d
+    done
+
+   # We do want to make sure this is all owned by our user.
+   sudo chown -R $USER: ${SWIFT_DATA_LOCATION}/{1..4}/node
+   sudo chown -R $USER: ${SWIFT_CONFIG_LOCATION}
+
+   # swift-init has a bug using /etc/swift until bug #885595 is fixed
+   # we have to create a link
+   sudo ln -s ${SWIFT_CONFIG_LOCATION} /etc/swift
+
+   # Swift use rsync to syncronize between all the different
+   # partitions (which make more sense when you have a multi-node
+   # setup) we configure it with our version of rsync.
+   sed -e "s/%GROUP%/${USER_GROUP}/;s/%USER%/$USER/;s,%SWIFT_DATA_LOCATION%,$SWIFT_DATA_LOCATION," $FILES/swift/rsyncd.conf | sudo tee /etc/rsyncd.conf
+   sudo sed -i '/^RSYNC_ENABLE=false/ { s/false/true/ }' /etc/default/rsync
+
+   # By default Swift will be installed with the tempauth middleware
+   # which has some default username and password if you have
+   # configured keystone it will checkout the directory.
+   if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
+       swift_auth_server=keystone
+       # We need a special version of bin/swift which understand the
+       # OpenStack api 2.0, we download it until this is getting
+       # integrated in swift.
+       sudo curl -s -o/usr/local/bin/swift \
+           'https://review.openstack.org/gitweb?p=openstack/swift.git;a=blob_plain;f=bin/swift;hb=48bfda6e2fdf3886c98bd15649887d54b9a2574e'
+   else
+       swift_auth_server=tempauth
+   fi
+
+   # We do the install of the proxy-server and swift configuration
+   # replacing a few directives to match our configuration.
+   sed "s,%SWIFT_CONFIG_LOCATION%,${SWIFT_CONFIG_LOCATION},;s/%USER%/$USER/;s/%SERVICE_TOKEN%/${SERVICE_TOKEN}/;s/%AUTH_SERVER%/${swift_auth_server}/" \
+       $FILES/swift/proxy-server.conf|sudo tee  ${SWIFT_CONFIG_LOCATION}/proxy-server.conf
+
+   sed -e "s/%SWIFT_HASH%/$SWIFT_HASH/" $FILES/swift/swift.conf > ${SWIFT_CONFIG_LOCATION}/swift.conf
+
+   # We need to generate a object/account/proxy configuration
+   # emulating 4 nodes on different ports we have a little function
+   # that help us doing that.
+   function generate_swift_configuration() {
+       local server_type=$1
+       local bind_port=$2
+       local log_facility=$3
+       local node_number
+
+       for node_number in {1..4};do
+           node_path=${SWIFT_DATA_LOCATION}/${node_number}
+           sed -e "s,%SWIFT_CONFIG_LOCATION%,${SWIFT_CONFIG_LOCATION},;s,%USER%,$USER,;s,%NODE_PATH%,${node_path},;s,%BIND_PORT%,${bind_port},;s,%LOG_FACILITY%,${log_facility}," \
+               $FILES/swift/${server_type}-server.conf > ${SWIFT_CONFIG_LOCATION}/${server_type}-server/${node_number}.conf
+           bind_port=$(( ${bind_port} + 10 ))
+           log_facility=$(( ${log_facility} + 1 ))
+       done
+   }
+   generate_swift_configuration object 6010 2
+   generate_swift_configuration container 6011 2
+   generate_swift_configuration account 6012 2
+
+   # We create two helper scripts :
+   #
+   # - swift-remakerings
+   #   Allow to recreate rings from scratch.
+   # - swift-startmain
+   #   Restart your full cluster.
+   #
+   sed -e "s,%SWIFT_CONFIG_LOCATION%,${SWIFT_CONFIG_LOCATION},;s/%SWIFT_PARTITION_POWER_SIZE%/$SWIFT_PARTITION_POWER_SIZE/" $FILES/swift/swift-remakerings | \
+       sudo tee /usr/local/bin/swift-remakerings
+   sudo install -m755 $FILES/swift/swift-startmain /usr/local/bin/
+   sudo chmod +x /usr/local/bin/swift-*
+
+   # We then can start rsync.
+   sudo /etc/init.d/rsync restart || :
+
+   # Create our ring for the object/container/account.
+   /usr/local/bin/swift-remakerings
+
+   # And now we launch swift-startmain to get our cluster running
+   # ready to be tested.
+   /usr/local/bin/swift-startmain || :
+
+   unset s swift_hash swift_auth_server tmpd
+fi
+
 # Volume Service
 # --------------
 
@@ -593,7 +779,7 @@
     # By default, the backing file is 2G in size, and is stored in /opt/stack.
     #
     if ! sudo vgdisplay | grep -q nova-volumes; then
-        VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-/opt/stack/nova-volumes-backing-file}
+        VOLUME_BACKING_FILE=${VOLUME_BACKING_FILE:-$DEST/nova-volumes-backing-file}
         VOLUME_BACKING_FILE_SIZE=${VOLUME_BACKING_FILE_SIZE:-2052M}
         truncate -s $VOLUME_BACKING_FILE_SIZE $VOLUME_BACKING_FILE
         DEV=`sudo losetup -f --show $VOLUME_BACKING_FILE`
@@ -616,8 +802,16 @@
 add_nova_flag "--allow_admin_api"
 add_nova_flag "--scheduler_driver=$SCHEDULER"
 add_nova_flag "--dhcpbridge_flagfile=$NOVA_DIR/bin/nova.conf"
-add_nova_flag "--network_manager=nova.network.manager.$NET_MAN"
 add_nova_flag "--fixed_range=$FIXED_RANGE"
+if [[ "$ENABLED_SERVICES" =~ "q-svc" ]]; then
+    add_nova_flag "--network_manager=nova.network.quantum.manager.QuantumManager"
+    if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
+        add_nova_flag "--libvirt_vif_type=ethernet"
+        add_nova_flag "--libvirt_vif_driver=nova.virt.libvirt.vif.LibvirtOpenVswitchDriver"
+    fi
+else
+    add_nova_flag "--network_manager=nova.network.manager.$NET_MAN"
+fi
 add_nova_flag "--my_ip=$HOST_IP"
 add_nova_flag "--public_interface=$PUBLIC_INTERFACE"
 add_nova_flag "--vlan_interface=$VLAN_INTERFACE"
@@ -632,15 +826,16 @@
 add_nova_flag "--rabbit_host=$RABBIT_HOST"
 add_nova_flag "--rabbit_password=$RABBIT_PASSWORD"
 add_nova_flag "--glance_api_servers=$GLANCE_HOSTPORT"
+add_nova_flag "--force_dhcp_release"
 if [ -n "$INSTANCES_PATH" ]; then
     add_nova_flag "--instances_path=$INSTANCES_PATH"
 fi
-if [ -n "$MULTI_HOST" ]; then
-    add_nova_flag "--multi_host=$MULTI_HOST"
-    add_nova_flag "--send_arp_for_ha=1"
+if [ "$MULTI_HOST" != "False" ]; then
+    add_nova_flag "--multi_host"
+    add_nova_flag "--send_arp_for_ha"
 fi
 if [ "$SYSLOG" != "False" ]; then
-    add_nova_flag "--use_syslog=1"
+    add_nova_flag "--use_syslog"
 fi
 
 # XenServer
@@ -676,12 +871,6 @@
 
     # (re)create nova database
     $NOVA_DIR/bin/nova-manage db sync
-
-    # create a small network
-    $NOVA_DIR/bin/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE
-
-    # create some floating ips
-    $NOVA_DIR/bin/nova-manage floating create $FLOATING_RANGE
 fi
 
 
@@ -721,16 +910,20 @@
 function screen_it {
     NL=`echo -ne '\015'`
     if [[ "$ENABLED_SERVICES" =~ "$1" ]]; then
-        screen -S nova -X screen -t $1
-        screen -S nova -p $1 -X stuff "$2$NL"
+        screen -S stack -X screen -t $1
+        # sleep to allow bash to be ready to be send the command - we are
+        # creating a new window in screen and then sends characters, so if
+        # bash isn't running by the time we send the command, nothing happens
+        sleep 1
+        screen -S stack -p $1 -X stuff "$2$NL"
     fi
 }
 
 # create a new named screen to run processes in
-screen -d -m -S nova -t nova
+screen -d -m -S stack -t stack
 sleep 1
 
-# launch the glance registery service
+# launch the glance registry service
 if [[ "$ENABLED_SERVICES" =~ "g-reg" ]]; then
     screen_it g-reg "cd $GLANCE_DIR; bin/glance-registry --config-file=etc/glance-registry.conf"
 fi
@@ -739,7 +932,7 @@
 if [[ "$ENABLED_SERVICES" =~ "g-api" ]]; then
     screen_it g-api "cd $GLANCE_DIR; bin/glance-api --config-file=etc/glance-api.conf"
     echo "Waiting for g-api ($GLANCE_HOSTPORT) to start..."
-    if ! timeout 60 sh -c "while ! wget -q -O- http://$GLANCE_HOSTPORT; do sleep 1; done"; then
+    if ! timeout $SERVICE_TIMEOUT sh -c "while ! wget -q -O- http://$GLANCE_HOSTPORT; do sleep 1; done"; then
       echo "g-api did not start"
       exit 1
     fi
@@ -749,7 +942,7 @@
 if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
     screen_it key "cd $KEYSTONE_DIR && $KEYSTONE_DIR/bin/keystone --config-file $KEYSTONE_CONF -d"
     echo "Waiting for keystone to start..."
-    if ! timeout 60 sh -c "while ! wget -q -O- http://127.0.0.1:5000; do sleep 1; done"; then
+    if ! timeout $SERVICE_TIMEOUT sh -c "while ! wget -q -O- http://127.0.0.1:5000; do sleep 1; done"; then
       echo "keystone did not start"
       exit 1
     fi
@@ -759,11 +952,62 @@
 if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then
     screen_it n-api "cd $NOVA_DIR && $NOVA_DIR/bin/nova-api"
     echo "Waiting for nova-api to start..."
-    if ! timeout 60 sh -c "while ! wget -q -O- http://127.0.0.1:8774; do sleep 1; done"; then
+    if ! timeout $SERVICE_TIMEOUT sh -c "while ! wget -q -O- http://127.0.0.1:8774; do sleep 1; done"; then
       echo "nova-api did not start"
       exit 1
     fi
 fi
+
+# Quantum
+if [[ "$ENABLED_SERVICES" =~ "q-svc" ]]; then
+    # Install deps
+    # FIXME add to files/apts/quantum, but don't install if not needed!
+    apt_get install openvswitch-switch openvswitch-datapath-dkms
+
+    # Create database for the plugin/agent
+    if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
+        if [[ "$ENABLED_SERVICES" =~ "mysql" ]]; then
+            mysql -u$MYSQL_USER -p$MYSQL_PASSWORD -e 'CREATE DATABASE IF NOT EXISTS ovs_quantum;'
+        else
+            echo "mysql must be enabled in order to use the $Q_PLUGIN Quantum plugin."
+            exit 1
+        fi
+    fi
+
+    QUANTUM_PLUGIN_INI_FILE=$QUANTUM_DIR/quantum/plugins.ini
+    # Make sure we're using the openvswitch plugin
+    sed -i -e "s/^provider =.*$/provider = quantum.plugins.openvswitch.ovs_quantum_plugin.OVSQuantumPlugin/g" $QUANTUM_PLUGIN_INI_FILE
+    screen_it q-svc "cd $QUANTUM_DIR && export PYTHONPATH=.:$PYTHONPATH; python $QUANTUM_DIR/bin/quantum $QUANTUM_DIR/etc/quantum.conf"
+fi
+
+# Quantum agent (for compute nodes)
+if [[ "$ENABLED_SERVICES" =~ "q-agt" ]]; then
+    if [[ "$Q_PLUGIN" = "openvswitch" ]]; then
+        # Set up integration bridge
+        OVS_BRIDGE=${OVS_BRIDGE:-br-int}
+        sudo ovs-vsctl --no-wait -- --if-exists del-br $OVS_BRIDGE
+        sudo ovs-vsctl --no-wait add-br $OVS_BRIDGE
+        sudo ovs-vsctl --no-wait br-set-external-id $OVS_BRIDGE bridge-id br-int
+    fi
+
+    # Start up the quantum <-> openvswitch agent
+    screen_it q-agt "sleep 4; sudo python $QUANTUM_DIR/quantum/plugins/openvswitch/agent/ovs_quantum_agent.py $QUANTUM_DIR/quantum/plugins/openvswitch/ovs_quantum_plugin.ini -v"
+fi
+
+# If we're using Quantum (i.e. q-svc is enabled), network creation has to
+# happen after we've started the Quantum service.
+if [[ "$ENABLED_SERVICES" =~ "mysql" ]]; then
+    # create a small network
+    $NOVA_DIR/bin/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE
+
+    if [[ "$ENABLED_SERVICES" =~ "q-svc" ]]; then
+        echo "Not creating floating IPs (not supported by QuantumManager)"
+    else
+        # create some floating ips
+        $NOVA_DIR/bin/nova-manage floating create $FLOATING_RANGE
+    fi
+fi
+
 # Launching nova-compute should be as simple as running ``nova-compute`` but
 # have to do a little more than that in our script.  Since we add the group
 # ``libvirtd`` to our user in this script, when nova-compute is run it is
@@ -787,7 +1031,7 @@
 # TTY also uses cloud-init, supporting login via keypair and sending scripts as
 # userdata.  See https://help.ubuntu.com/community/CloudInit for more on cloud-init
 #
-# Override ``IMAGE_URLS`` with a comma-seperated list of uec images.
+# Override ``IMAGE_URLS`` with a comma-separated list of uec images.
 #
 #  * **natty**: http://uec-images.ubuntu.com/natty/current/natty-server-cloudimg-amd64.tar.gz
 #  * **oneiric**: http://uec-images.ubuntu.com/oneiric/current/oneiric-server-cloudimg-amd64.tar.gz
@@ -843,6 +1087,10 @@
 # Using the cloud
 # ===============
 
+echo ""
+echo ""
+echo ""
+
 # If you installed the horizon on this server, then you should be able
 # to access the site using your browser.
 if [[ "$ENABLED_SERVICES" =~ "horizon" ]]; then
diff --git a/stackrc b/stackrc
index 9b110a3..b541711 100644
--- a/stackrc
+++ b/stackrc
@@ -2,6 +2,14 @@
 NOVA_REPO=https://github.com/cloudbuilders/nova.git
 NOVA_BRANCH=diablo
 
+# storage service
+SWIFT_REPO=https://github.com/openstack/swift.git
+SWIFT_BRANCH=stable/diablo
+
+# swift and keystone integration
+SWIFT_KEYSTONE_REPO=https://github.com/cloudbuilders/swift-keystone2.git
+SWIFT_KEYSTONE_BRANCH=master
+
 # image catalog service
 GLANCE_REPO=https://github.com/cloudbuilders/glance.git
 GLANCE_BRANCH=diablo
@@ -27,6 +35,10 @@
 OPENSTACKX_REPO=https://github.com/cloudbuilders/openstackx.git
 OPENSTACKX_BRANCH=diablo
 
+# quantum service
+QUANTUM_REPO=https://github.com/openstack/quantum
+QUANTUM_BRANCH=stable/diablo
+
 # Specify a comma-separated list of uec images to download and install into glance.
 IMAGE_URLS=http://smoser.brickies.net/ubuntu/ttylinux-uec/ttylinux-uec-amd64-11.2_2.6.35-15_1.tar.gz
 
diff --git a/tools/build_kvm.sh b/tools/build_kvm.sh
deleted file mode 100755
index 1b33926..0000000
--- a/tools/build_kvm.sh
+++ /dev/null
@@ -1,402 +0,0 @@
-#!/usr/bin/env bash
-
-# exit on error to stop unexpected errors
-set -o errexit
-
-# Make sure that we have the proper version of ubuntu
-UBUNTU_VERSION=`cat /etc/lsb-release | grep CODENAME | sed 's/.*=//g'`
-if [ ! "oneiric" = "$UBUNTU_VERSION" ]; then
-    if [ ! "natty" = "$UBUNTU_VERSION" ]; then
-        echo "This script only works with oneiric and natty"
-        exit 1
-    fi
-fi
-
-# Echo commands
-set -o xtrace
-
-# Keep track of the current directory
-TOOLS_DIR=$(cd $(dirname "$0") && pwd)
-TOP_DIR=$TOOLS_DIR/..
-
-# Where to store files and instances
-KVMSTACK_DIR=${KVMSTACK_DIR:-/opt/kvmstack}
-
-# Where to store images
-IMAGES_DIR=$KVMSTACK_DIR/images
-
-# Create images dir
-mkdir -p $IMAGES_DIR
-
-# Move to top devstack dir
-cd $TOP_DIR
-
-# Abort if localrc is not set
-if [ ! -e ./localrc ]; then
-    echo "You must have a localrc with ALL necessary passwords defined before proceeding."
-    echo "See stack.sh for required passwords."
-    exit 1
-fi
-
-# Source params
-source ./stackrc
-
-# Configure the root password of the vm to be the same as ``ADMIN_PASSWORD``
-ROOT_PASSWORD=${ADMIN_PASSWORD:-password}
-
-
-# Base image (natty by default)
-DIST_NAME=${DIST_NAME:-natty}
-IMAGE_FNAME=$DIST_NAME.raw
-
-# Name of our instance, used by libvirt
-GUEST_NAME=${GUEST_NAME:-kvmstack}
-
-# Original version of built image
-BASE_IMAGE=$KVMSTACK_DIR/images/$DIST_NAME.raw
-
-# Copy of base image, which we pre-install with tasty treats
-VM_IMAGE=$IMAGES_DIR/$DIST_NAME.$GUEST_NAME.raw
-
-# Mop up after previous runs
-virsh destroy $GUEST_NAME || true
-
-# Where this vm is stored
-VM_DIR=$KVMSTACK_DIR/instances/$GUEST_NAME
-
-# Create vm dir
-mkdir -p $VM_DIR
-
-# Mount point into copied base image
-COPY_DIR=$VM_DIR/copy
-mkdir -p $COPY_DIR
-
-# Create the base image if it does not yet exist
-if [ ! -e $IMAGES_DIR/$IMAGE_FNAME ]; then
-    cd $TOOLS_DIR
-    ./make_image.sh -m -r 5000  $DIST_NAME raw
-    mv $DIST_NAME.raw $BASE_IMAGE
-    cd $TOP_DIR
-fi
-
-# Create a copy of the base image
-if [ ! -e $VM_IMAGE ]; then
-    cp -p $BASE_IMAGE $VM_IMAGE
-fi
-
-# Unmount the copied base image
-function unmount_images() {
-    # unmount the filesystem
-    while df | grep -q $COPY_DIR; do
-        umount $COPY_DIR || echo 'ok'
-        sleep 1
-    done
-}
-
-# Unmount from failed runs
-unmount_images
-
-# Ctrl-c catcher
-function kill_unmount() {
-    unmount_images
-    exit 1
-}
-
-# Install deps if needed
-dpkg -l kvm libvirt-bin kpartx || apt-get install -y --force-yes kvm libvirt-bin kpartx
-
-# Let Ctrl-c kill tail and exit
-trap kill_unmount SIGINT
-
-# Where Openstack code will live in image
-DEST=${DEST:-/opt/stack}
-
-# Mount the file system
-mount -o loop,offset=32256 $VM_IMAGE  $COPY_DIR
-
-# git clone only if directory doesn't exist already.  Since ``DEST`` might not
-# be owned by the installation user, we create the directory and change the
-# ownership to the proper user.
-function git_clone {
-    if [ ! -d $2 ]; then
-        sudo mkdir $2
-        sudo chown `whoami` $2
-        git clone $1 $2
-        cd $2
-        # This checkout syntax works for both branches and tags
-        git checkout $3
-    fi
-}
-
-# Make sure that base requirements are installed
-cp /etc/resolv.conf $COPY_DIR/etc/resolv.conf
-chroot $COPY_DIR apt-get update
-chroot $COPY_DIR apt-get install -y --force-yes `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-chroot $COPY_DIR apt-get install -y --download-only rabbitmq-server libvirt-bin mysql-server
-chroot $COPY_DIR pip install `cat files/pips/*`
-
-# Clean out code repos if directed to do so
-if [ "$CLEAN" = "1" ]; then
-    rm -rf $COPY_DIR/$DEST
-fi
-
-# Cache openstack code
-mkdir -p $COPY_DIR/$DEST
-git_clone $NOVA_REPO $COPY_DIR/$DEST/nova $NOVA_BRANCH
-git_clone $GLANCE_REPO $COPY_DIR/$DEST/glance $GLANCE_BRANCH
-git_clone $KEYSTONE_REPO $COPY_DIR/$DESTkeystone $KEYSTONE_BRANCH
-git_clone $NOVNC_REPO $COPY_DIR/$DEST/noVNC $NOVNC_BRANCH
-git_clone $HORIZON_REPO $COPY_DIR/$DEST/horizon $HORIZON_BRANCH $HORIZON_TAG
-git_clone $NOVACLIENT_REPO $COPY_DIR/$DEST/python-novaclient $NOVACLIENT_BRANCH
-git_clone $OPENSTACKX_REPO $COPY_DIR/$DEST/openstackx $OPENSTACKX_BRANCH
-git_clone $KEYSTONE_REPO $COPY_DIR/$DEST/keystone $KEYSTONE_BRANCH
-git_clone $NOVNC_REPO $COPY_DIR/$DEST/noVNC $NOVNC_BRANCH
-
-# Back to devstack
-cd $TOP_DIR
-
-# Unmount the filesystems
-unmount_images
-
-# Network configuration variables
-BRIDGE=${BRIDGE:-br0}
-GUEST_IP=${GUEST_IP:-192.168.1.50}
-GUEST_CIDR=${GUEST_CIDR:-$GUEST_IP/24}
-GUEST_NETMASK=${GUEST_NETMASK:-255.255.255.0}
-GUEST_GATEWAY=${GUEST_GATEWAY:-192.168.1.1}
-GUEST_MAC=${GUEST_MAC:-"02:16:3e:07:69:`printf '%02X' $(echo $GUEST_IP | sed "s/.*\.//")`"}
-GUEST_RAM=${GUEST_RAM:-1524288}
-GUEST_CORES=${GUEST_CORES:-1}
-
-# libvirt.xml configuration
-LIBVIRT_XML=$VM_DIR/libvirt.xml
-cat > $LIBVIRT_XML <<EOF
-<domain type='kvm'>
-    <name>$GUEST_NAME</name>
-    <memory>$GUEST_RAM</memory>
-    <os>
-        <type>hvm</type>
-        <bootmenu enable='yes'/>
-    </os>
-    <features>
-        <acpi/>
-    </features>
-    <vcpu>$GUEST_CORES</vcpu>
-    <devices>
-        <disk type='file'>
-            <driver type='qcow2'/>
-            <source file='$VM_DIR/disk'/>
-            <target dev='vda' bus='virtio'/>
-        </disk>
-
-        <interface type='bridge'>
-            <source bridge='$BRIDGE'/>
-            <mac address='$GUEST_MAC'/>
-        </interface>
-
-        <!-- The order is significant here.  File must be defined first -->
-        <serial type="file">
-            <source path='$VM_DIR/console.log'/>
-            <target port='1'/>
-        </serial>
-
-        <console type='pty' tty='/dev/pts/2'>
-            <source path='/dev/pts/2'/>
-            <target port='0'/>
-        </console>
-
-        <serial type='pty'>
-            <source path='/dev/pts/2'/>
-            <target port='0'/>
-        </serial>
-
-        <graphics type='vnc' port='-1' autoport='yes' keymap='en-us' listen='0.0.0.0'/>
-    </devices>
-</domain>
-EOF
-
-# Mount point for instance fs
-ROOTFS=$VM_DIR/root
-mkdir -p $ROOTFS
-
-# Make sure we have nbd-ness
-modprobe nbd max_part=63
-
-# Which NBD device to use?
-NBD=${NBD:-/dev/nbd5}
-
-# Clean up from previous runs
-umount $ROOTFS || echo 'ok'
-qemu-nbd -d $NBD || echo 'ok'
-
-# Clean up old runs
-cd $VM_DIR
-rm -f $VM_DIR/disk
-
-# Create our instance fs
-qemu-img create -f qcow2 -b $VM_IMAGE disk
-
-# Connect our nbd and wait till it is mountable
-qemu-nbd -c $NBD disk
-NBD_DEV=`basename $NBD`
-if ! timeout 60 sh -c "while ! [ -e /sys/block/$NBD_DEV/pid ]; do sleep 1; done"; then
-    echo "Couldn't connect $NBD"
-    exit 1
-fi
-
-# Mount the instance
-mount $NBD $ROOTFS -o offset=32256 -t ext4
-
-# Configure instance network
-INTERFACES=$ROOTFS/etc/network/interfaces
-cat > $INTERFACES <<EOF
-auto lo
-iface lo inet loopback
-
-auto eth0
-iface eth0 inet static
-        address $GUEST_IP
-        netmask $GUEST_NETMASK
-        gateway $GUEST_GATEWAY
-EOF
-
-# User configuration for the instance
-chroot $ROOTFS groupadd libvirtd || true
-chroot $ROOTFS useradd stack -s /bin/bash -d $DEST -G libvirtd
-cp -pr $TOOLS_DIR/.. $ROOTFS/$DEST/devstack
-echo "root:$ROOT_PASSWORD" | chroot $ROOTFS chpasswd
-echo "stack:pass" | chroot $ROOTFS chpasswd
-echo "stack ALL=(ALL) NOPASSWD: ALL" >> $ROOTFS/etc/sudoers
-
-# Gracefully cp only if source file/dir exists
-function cp_it {
-    if [ -e $1 ] || [ -d $1 ]; then
-        cp -pRL $1 $2
-    fi
-}
-
-# Copy over your ssh keys and env if desired
-COPYENV=${COPYENV:-1}
-if [ "$COPYENV" = "1" ]; then
-    cp_it ~/.ssh $ROOTFS/$DEST/.ssh
-    cp_it ~/.ssh/id_rsa.pub $ROOTFS/$DEST/.ssh/authorized_keys
-    cp_it ~/.gitconfig $ROOTFS/$DEST/.gitconfig
-    cp_it ~/.vimrc $ROOTFS/$DEST/.vimrc
-    cp_it ~/.bashrc $ROOTFS/$DEST/.bashrc
-fi
-
-# pre-cache uec images
-for image_url in ${IMAGE_URLS//,/ }; do
-    IMAGE_FNAME=`basename "$image_url"`
-    if [ ! -f $IMAGES_DIR/$IMAGE_FNAME ]; then
-        wget -c $image_url -O $IMAGES_DIR/$IMAGE_FNAME
-    fi
-    cp $IMAGES_DIR/$IMAGE_FNAME $ROOTFS/$DEST/devstack/files
-done
-
-# Configure the runner
-RUN_SH=$ROOTFS/$DEST/run.sh
-cat > $RUN_SH <<EOF
-#!/usr/bin/env bash
-
-# Kill any existing screens
-killall screen
-
-# Install and run stack.sh
-sudo apt-get update
-sudo apt-get -y --force-yes install git-core vim-nox sudo
-if [ ! -d "$DEST/devstack" ]; then
-    git clone git://github.com/cloudbuilders/devstack.git $DEST/devstack
-fi
-cd $DEST/devstack && $STACKSH_PARAMS FORCE=yes ./stack.sh > /$DEST/run.sh.log
-echo >> /$DEST/run.sh.log
-echo >> /$DEST/run.sh.log
-echo "All done! Time to start clicking." >> /$DEST/run.sh.log
-cat $DEST/run.sh.log
-EOF
-chmod 755 $RUN_SH
-
-# Make runner launch on boot
-RC_LOCAL=$ROOTFS/etc/init.d/local
-cat > $RC_LOCAL <<EOF
-#!/bin/sh -e
-# Reboot if this is our first run to enable console log on $DIST_NAME :(
-if [ ! -e /root/firstlaunch ]; then
-    touch /root/firstlaunch
-    reboot -f
-    exit 0
-fi
-su -c "$DEST/run.sh" stack
-EOF
-chmod +x $RC_LOCAL
-chroot $ROOTFS sudo update-rc.d local defaults 80
-
-# Make our ip address hostnames look nice at the command prompt
-echo "export PS1='${debian_chroot:+($debian_chroot)}\\u@\\H:\\w\\$ '" >> $ROOTFS/$DEST/.bashrc
-echo "export PS1='${debian_chroot:+($debian_chroot)}\\u@\\H:\\w\\$ '" >> $ROOTFS/etc/profile
-
-# Give stack ownership over $DEST so it may do the work needed
-chroot $ROOTFS chown -R stack $DEST
-
-# Change boot params so that we get a console log
-sudo sed -e "s/quiet splash/splash console=ttyS0 console=ttyS1,19200n8/g" -i $ROOTFS/boot/grub/menu.lst
-sudo sed -e "s/^hiddenmenu//g" -i $ROOTFS/boot/grub/menu.lst
-
-# Set the hostname
-echo $GUEST_NAME > $ROOTFS/etc/hostname
-
-# We need the hostname to resolve for rabbit to launch
-if ! grep -q $GUEST_NAME $ROOTFS/etc/hosts; then
-    echo "$GUEST_IP $GUEST_NAME" >> $ROOTFS/etc/hosts
-fi
-
-# Unmount
-umount $ROOTFS || echo 'ok'
-qemu-nbd -d $NBD
-
-# Create the instance
-cd $VM_DIR && virsh create libvirt.xml
-
-# Tail the console log till we are done
-WAIT_TILL_LAUNCH=${WAIT_TILL_LAUNCH:-1}
-if [ "$WAIT_TILL_LAUNCH" = "1" ]; then
-    # Done creating the container, let's tail the log
-    echo
-    echo "============================================================="
-    echo "                          -- YAY! --"
-    echo "============================================================="
-    echo
-    echo "We're done launching the vm, about to start tailing the"
-    echo "stack.sh log. It will take a second or two to start."
-    echo
-    echo "Just CTRL-C at any time to stop tailing."
-
-    while [ ! -e "$VM_DIR/console.log" ]; do
-      sleep 1
-    done
-
-    tail -F $VM_DIR/console.log &
-
-    TAIL_PID=$!
-
-    function kill_tail() {
-        kill $TAIL_PID
-        exit 1
-    }
-
-    # Let Ctrl-c kill tail and exit
-    trap kill_tail SIGINT
-
-    echo "Waiting stack.sh to finish..."
-    while ! cat $VM_DIR/console.log | grep -q 'All done' ; do
-        sleep 5
-    done
-
-    kill $TAIL_PID
-
-    if grep -q "stack.sh failed" $VM_DIR/console.log; then
-        exit 1
-    fi
-    echo ""
-    echo "Finished - Zip-a-dee Doo-dah!"
-fi
diff --git a/tools/build_libvirt.sh b/tools/build_libvirt.sh
index 5ae8fda..d192879 100755
--- a/tools/build_libvirt.sh
+++ b/tools/build_libvirt.sh
@@ -12,6 +12,27 @@
     fi
 fi
 
+# Clean up any resources that may be in use
+cleanup() {
+    set +o errexit
+    unmount_images
+
+    if [ -n "$ROOTFS" ]; then
+        umount $ROOTFS/dev
+        umount $ROOTFS
+    fi
+
+    # Release NBD devices
+    if [ -n "$NBD" ]; then
+        qemu-nbd -d $NBD
+    fi
+
+    # Kill ourselves to signal any calling process
+    trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
 # Echo commands
 set -o xtrace
 
@@ -100,9 +121,6 @@
 # Install deps if needed
 dpkg -l kvm libvirt-bin kpartx || apt-get install -y --force-yes kvm libvirt-bin kpartx
 
-# Let Ctrl-c kill tail and exit
-trap kill_unmount SIGINT
-
 # Where Openstack code will live in image
 DEST=${DEST:-/opt/stack}
 
@@ -127,8 +145,8 @@
 # Make sure that base requirements are installed
 cp /etc/resolv.conf $COPY_DIR/etc/resolv.conf
 chroot $COPY_DIR apt-get update
-chroot $COPY_DIR apt-get install -y --force-yes `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-chroot $COPY_DIR apt-get install -y --download-only rabbitmq-server libvirt-bin mysql-server
+chroot $COPY_DIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+chroot $COPY_DIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
 chroot $COPY_DIR pip install `cat files/pips/*`
 
 # Clean out code repos if directed to do so
@@ -156,6 +174,7 @@
 
 # Network configuration variables
 GUEST_NETWORK=${GUEST_NETWORK:-1}
+GUEST_RECREATE_NET=${GUEST_RECREATE_NET:-yes}
 
 GUEST_IP=${GUEST_IP:-192.168.$GUEST_NETWORK.50}
 GUEST_CIDR=${GUEST_CIDR:-$GUEST_IP/24}
@@ -176,8 +195,10 @@
 </network>
 EOF
 
-virsh net-destroy devstack-$GUEST_NETWORK || true
-virsh net-create $VM_DIR/net.xml
+if [[ "$GUEST_RECREATE_NET" == "yes" ]]; then
+    virsh net-destroy devstack-$GUEST_NETWORK || true
+    virsh net-create $VM_DIR/net.xml
+fi
 
 # libvirt.xml configuration
 LIBVIRT_XML=$VM_DIR/libvirt.xml
@@ -239,26 +260,35 @@
 # Create our instance fs
 qemu-img create -f qcow2 -b $VM_IMAGE disk
 
+# Finds the next available NBD device
+# Exits script if error connecting or none free
+# map_nbd image
+# returns full nbd device path
+function map_nbd {
+    for i in `seq 0 15`; do
+        if [ ! -e /sys/block/nbd$i/pid ]; then
+            NBD=/dev/nbd$i
+            # Connect to nbd and wait till it is ready
+            qemu-nbd -c $NBD $1
+            if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
+                echo "Couldn't connect $NBD"
+                exit 1
+            fi
+            break
+        fi
+    done
+    if [ -z "$NBD" ]; then
+        echo "No free NBD slots"
+        exit 1
+    fi
+    echo $NBD
+}
+
 # Make sure we have nbd-ness
 modprobe nbd max_part=63
 
 # Set up nbd
-for i in `seq 0 15`; do
-    if [ ! -e /sys/block/nbd$i/pid ]; then
-        NBD=/dev/nbd$i
-        # Connect to nbd and wait till it is ready
-        qemu-nbd -c $NBD disk
-        if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
-            echo "Couldn't connect $NBD"
-            exit 1
-        fi
-        break
-    fi
-done
-if [ -z "$NBD" ]; then
-    echo "No free NBD slots"
-    exit 1
-fi
+NBD=`map_nbd disk`
 NBD_DEV=`basename $NBD`
 
 # Mount the instance
@@ -381,7 +411,9 @@
 
 # Unmount
 umount $ROOTFS || echo 'ok'
+ROOTFS=""
 qemu-nbd -d $NBD
+NBD=""
 
 # Create the instance
 cd $VM_DIR && virsh create libvirt.xml
diff --git a/tools/build_lxc.sh b/tools/build_lxc.sh
index a2c5a22..9d8ce92 100755
--- a/tools/build_lxc.sh
+++ b/tools/build_lxc.sh
@@ -125,8 +125,8 @@
 
 # Make sure that base requirements are installed
 chroot $CACHEDIR apt-get update
-chroot $CACHEDIR apt-get install -y --force-yes `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-chroot $CACHEDIR apt-get install -y --download-only rabbitmq-server libvirt-bin mysql-server
+chroot $CACHEDIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+chroot $CACHEDIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
 chroot $CACHEDIR pip install `cat files/pips/*`
 
 # Clean out code repos if directed to do so
diff --git a/tools/build_nfs.sh b/tools/build_nfs.sh
index 5c591e4..39a2cf0 100755
--- a/tools/build_nfs.sh
+++ b/tools/build_nfs.sh
@@ -32,8 +32,9 @@
 # prime natty with as many apt/pips as we can
 if [ ! -d $CHROOTCACHE/natty-dev ]; then
     rsync -azH $CHROOTCACHE/natty-base/ $CHROOTCACHE/natty-dev/
-    chroot $CHROOTCACHE/natty-dev apt-get install -y `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
-    chroot $CHROOTCACHE/natty-dev pip install `cat files/pips/*`
+    chroot $CHROOTCACHE apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+    chroot $CHROOTCACHE apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
+    chroot $CHROOTCACHE pip install `cat files/pips/*`
 
     # Create a stack user that is a member of the libvirtd group so that stack
     # is able to interact with libvirt.
diff --git a/tools/build_pxe_boot.sh b/tools/build_pxe_boot.sh
index da8bbcc..ab64098 100755
--- a/tools/build_pxe_boot.sh
+++ b/tools/build_pxe_boot.sh
@@ -11,6 +11,22 @@
 OPWD=`pwd`
 PROGDIR=`dirname $0`
 
+# Clean up any resources that may be in use
+cleanup() {
+    set +o errexit
+
+    # Mop up temporary files
+    if [ -n "$MNTDIR" -a -d "$MNTDIR" ]; then
+        umount $MNTDIR
+        rmdir $MNTDIR
+    fi
+
+    # Kill ourselves to signal any calling process
+    trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
 mkdir -p $DEST_DIR/pxelinux.cfg
 cd $DEST_DIR
 for i in memdisk menu.c32 pxelinux.0; do
diff --git a/tools/build_ramdisk.sh b/tools/build_ramdisk.sh
index 43f8999..2c914dc 100755
--- a/tools/build_ramdisk.sh
+++ b/tools/build_ramdisk.sh
@@ -1,60 +1,120 @@
 #!/bin/bash
 # build_ramdisk.sh - Build RAM disk images
 
+# exit on error to stop unexpected errors
+set -o errexit
+
 if [ ! "$#" -eq "1" ]; then
-    echo "$0 builds a gziped natty openstack install"
+    echo "$0 builds a gziped Ubuntu OpenStack install"
     echo "usage: $0 dest"
     exit 1
 fi
 
+# Clean up any resources that may be in use
+cleanup() {
+    set +o errexit
+
+    # Mop up temporary files
+    if [ -n "$MNTDIR" -a -d "$MNTDIR" ]; then
+        umount $MNTDIR
+        rmdir $MNTDIR
+    fi
+    if [ -n "$DEV_FILE_TMP" -a -e "$DEV_FILE_TMP "]; then
+        rm -f $DEV_FILE_TMP
+    fi
+    if [ -n "$IMG_FILE_TMP" -a -e "$IMG_FILE_TMP" ]; then
+        rm -f $IMG_FILE_TMP
+    fi
+
+    # Release NBD devices
+    if [ -n "$NBD" ]; then
+        qemu-nbd -d $NBD
+    fi
+
+    # Kill ourselves to signal any calling process
+    trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
+# Set up nbd
+modprobe nbd max_part=63
+
+# Echo commands
+set -o xtrace
+
 IMG_FILE=$1
 
-PROGDIR=`dirname $0`
-CHROOTCACHE=${CHROOTCACHE:-/var/cache/devstack}
-
-# Source params
-source ./stackrc
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
 
 # Store cwd
 CWD=`pwd`
 
+cd $TOP_DIR
+
+# Source params
+source ./stackrc
+
+CACHEDIR=${CACHEDIR:-/var/cache/devstack}
+
 DEST=${DEST:-/opt/stack}
 
+# Configure the root password of the vm to be the same as ``ADMIN_PASSWORD``
+ROOT_PASSWORD=${ADMIN_PASSWORD:-password}
+
+# Base image (natty by default)
+DIST_NAME=${DIST_NAME:-natty}
+
 # Param string to pass to stack.sh.  Like "EC2_DMZ_HOST=192.168.1.1 MYSQL_USER=nova"
 STACKSH_PARAMS=${STACKSH_PARAMS:-}
 
 # Option to use the version of devstack on which we are currently working
 USE_CURRENT_DEVSTACK=${USE_CURRENT_DEVSTACK:-1}
 
-# Set up nbd
-modprobe nbd max_part=63
-NBD=${NBD:-/dev/nbd9}
-NBD_DEV=`basename $NBD`
-
-# clean install of natty
-if [ ! -r $CHROOTCACHE/natty-base.img ]; then
-    $PROGDIR/get_uec_image.sh natty $CHROOTCACHE/natty-base.img
-#    # copy kernel modules...
-#    # NOTE(ja): is there a better way to do this?
-#    cp -pr /lib/modules/`uname -r` $CHROOTCACHE/natty-base/lib/modules
-#    # a simple password - pass
-#    echo root:pass | chroot $CHROOTCACHE/natty-base chpasswd
+# clean install
+if [ ! -r $CACHEDIR/$DIST_NAME-base.img ]; then
+    $TOOLS_DIR/get_uec_image.sh $DIST_NAME $CACHEDIR/$DIST_NAME-base.img
 fi
 
-# prime natty with as many apt/pips as we can
-if [ ! -r $CHROOTCACHE/natty-dev.img ]; then
-    cp -p $CHROOTCACHE/natty-base.img $CHROOTCACHE/natty-dev.img
-
-    qemu-nbd -c $NBD $CHROOTCACHE/natty-dev.img
-    if ! timeout 60 sh -c "while ! [ -e /sys/block/$NBD_DEV/pid ]; do sleep 1; done"; then
-        echo "Couldn't connect $NBD"
+# Finds the next available NBD device
+# Exits script if error connecting or none free
+# map_nbd image
+# returns full nbd device path
+function map_nbd {
+    for i in `seq 0 15`; do
+        if [ ! -e /sys/block/nbd$i/pid ]; then
+            NBD=/dev/nbd$i
+            # Connect to nbd and wait till it is ready
+            qemu-nbd -c $NBD $1
+            if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
+                echo "Couldn't connect $NBD"
+                exit 1
+            fi
+            break
+        fi
+    done
+    if [ -z "$NBD" ]; then
+        echo "No free NBD slots"
         exit 1
     fi
+    echo $NBD
+}
+
+# prime image with as many apt/pips as we can
+DEV_FILE=$CACHEDIR/$DIST_NAME-dev.img
+DEV_FILE_TMP=`mktemp $DEV_FILE.XXXXXX`
+if [ ! -r $DEV_FILE ]; then
+    cp -p $CACHEDIR/$DIST_NAME-base.img $DEV_FILE_TMP
+
+    NBD=`map_nbd $DEV_FILE_TMP`
     MNTDIR=`mktemp -d --tmpdir mntXXXXXXXX`
     mount -t ext4 ${NBD}p1 $MNTDIR
     cp -p /etc/resolv.conf $MNTDIR/etc/resolv.conf
 
-    chroot $MNTDIR apt-get install -y `cat files/apts/* | cut -d\# -f1 | egrep -v "(rabbitmq|libvirt-bin|mysql-server)"`
+    chroot $MNTDIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+    chroot $MNTDIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1`
     chroot $MNTDIR pip install `cat files/pips/*`
 
     # Create a stack user that is a member of the libvirtd group so that stack
@@ -66,6 +126,7 @@
 
     # a simple password - pass
     echo stack:pass | chroot $MNTDIR chpasswd
+    echo root:$ROOT_PASSWORD | chroot $MNTDIR chpasswd
 
     # and has sudo ability (in the future this should be limited to only what
     # stack requires)
@@ -74,27 +135,31 @@
     umount $MNTDIR
     rmdir $MNTDIR
     qemu-nbd -d $NBD
+    NBD=""
+    mv $DEV_FILE_TMP $DEV_FILE
 fi
+rm -f $DEV_FILE_TMP
 
 # clone git repositories onto the system
 # ======================================
 
+IMG_FILE_TMP=`mktemp $IMG_FILE.XXXXXX`
+
 if [ ! -r $IMG_FILE ]; then
-    qemu-nbd -c $NBD $CHROOTCACHE/natty-dev.img
-    if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
-        echo "Couldn't connect $NBD"
-        exit 1
-    fi
+    NBD=`map_nbd $DEV_FILE`
 
     # Pre-create the image file
     # FIXME(dt): This should really get the partition size to
     #            pre-create the image file
-    dd if=/dev/zero of=$IMG_FILE bs=1 count=1 seek=$((2*1024*1024*1024))
+    dd if=/dev/zero of=$IMG_FILE_TMP bs=1 count=1 seek=$((2*1024*1024*1024))
     # Create filesystem image for RAM disk
-    dd if=${NBD}p1 of=$IMG_FILE bs=1M
+    dd if=${NBD}p1 of=$IMG_FILE_TMP bs=1M
 
     qemu-nbd -d $NBD
+    NBD=""
+    mv $IMG_FILE_TMP $IMG_FILE
 fi
+rm -f $IMG_FILE_TMP
 
 MNTDIR=`mktemp -d --tmpdir mntXXXXXXXX`
 mount -t ext4 -o loop $IMG_FILE $MNTDIR
diff --git a/tools/build_uec.sh b/tools/build_uec.sh
new file mode 100755
index 0000000..d57cb29
--- /dev/null
+++ b/tools/build_uec.sh
@@ -0,0 +1,248 @@
+#!/usr/bin/env bash
+
+# Make sure that we have the proper version of ubuntu (only works on oneiric)
+if ! egrep -q "oneiric" /etc/lsb-release; then
+    echo "This script only works with ubuntu oneiric."
+    exit 1
+fi
+
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+
+cd $TOP_DIR
+
+# Source params
+source ./stackrc
+
+# Ubuntu distro to install
+DIST_NAME=${DIST_NAME:-oneiric}
+
+# Configure how large the VM should be
+GUEST_SIZE=${GUEST_SIZE:-10G}
+
+# exit on error to stop unexpected errors
+set -o errexit
+set -o xtrace
+
+# Abort if localrc is not set
+if [ ! -e $TOP_DIR/localrc ]; then
+    echo "You must have a localrc with ALL necessary passwords defined before proceeding."
+    echo "See stack.sh for required passwords."
+    exit 1
+fi
+
+# Install deps if needed
+DEPS="kvm libvirt-bin kpartx cloud-utils"
+dpkg -l $DEPS || apt-get install -y --force-yes $DEPS
+
+# Where to store files and instances
+WORK_DIR=${WORK_DIR:-/opt/kvmstack}
+
+# Where to store images
+image_dir=$WORK_DIR/images/$DIST_NAME
+mkdir -p $image_dir
+
+# Original version of built image
+uec_url=http://uec-images.ubuntu.com/$DIST_NAME/current/$DIST_NAME-server-cloudimg-amd64.tar.gz
+tarball=$image_dir/$(basename $uec_url)
+
+# download the base uec image if we haven't already
+if [ ! -f $tarball ]; then
+    curl $uec_url -o $tarball
+    (cd $image_dir && tar -Sxvzf $tarball)
+    resize-part-image $image_dir/*.img $GUEST_SIZE $image_dir/disk
+    cp $image_dir/*-vmlinuz-virtual $image_dir/kernel
+fi
+
+
+# Configure the root password of the vm to be the same as ``ADMIN_PASSWORD``
+ROOT_PASSWORD=${ADMIN_PASSWORD:-password}
+
+# Name of our instance, used by libvirt
+GUEST_NAME=${GUEST_NAME:-devstack}
+
+# Mop up after previous runs
+virsh destroy $GUEST_NAME || true
+
+# Where this vm is stored
+vm_dir=$WORK_DIR/instances/$GUEST_NAME
+
+# Create vm dir and remove old disk
+mkdir -p $vm_dir
+rm -f $vm_dir/disk
+
+# Create a copy of the base image
+qemu-img create -f qcow2 -b $image_dir/disk $vm_dir/disk
+
+# Back to devstack
+cd $TOP_DIR
+
+GUEST_NETWORK=${GUEST_NETWORK:-1}
+GUEST_RECREATE_NET=${GUEST_RECREATE_NET:-yes}
+GUEST_IP=${GUEST_IP:-192.168.$GUEST_NETWORK.50}
+GUEST_CIDR=${GUEST_CIDR:-$GUEST_IP/24}
+GUEST_NETMASK=${GUEST_NETMASK:-255.255.255.0}
+GUEST_GATEWAY=${GUEST_GATEWAY:-192.168.$GUEST_NETWORK.1}
+GUEST_MAC=${GUEST_MAC:-"02:16:3e:07:69:`printf '%02X' $GUEST_NETWORK`"}
+GUEST_RAM=${GUEST_RAM:-1524288}
+GUEST_CORES=${GUEST_CORES:-1}
+
+# libvirt.xml configuration
+NET_XML=$vm_dir/net.xml
+cat > $NET_XML <<EOF
+<network>
+  <name>devstack-$GUEST_NETWORK</name>
+  <bridge name="stackbr%d" />
+  <forward/>
+  <ip address="$GUEST_GATEWAY" netmask="$GUEST_NETMASK">
+    <dhcp>
+      <range start='192.168.$GUEST_NETWORK.2' end='192.168.$GUEST_NETWORK.127' />
+    </dhcp>
+  </ip>
+</network>
+EOF
+
+if [[ "$GUEST_RECREATE_NET" == "yes" ]]; then
+    virsh net-destroy devstack-$GUEST_NETWORK || true
+    # destroying the network isn't enough to delete the leases
+    rm -f /var/lib/libvirt/dnsmasq/devstack-$GUEST_NETWORK.leases
+    virsh net-create $vm_dir/net.xml
+fi
+
+# libvirt.xml configuration
+LIBVIRT_XML=$vm_dir/libvirt.xml
+cat > $LIBVIRT_XML <<EOF
+<domain type='kvm'>
+  <name>$GUEST_NAME</name>
+  <memory>$GUEST_RAM</memory>
+  <os>
+    <type>hvm</type>
+    <kernel>$image_dir/kernel</kernel>
+    <cmdline>root=/dev/vda ro console=ttyS0 init=/usr/lib/cloud-init/uncloud-init ds=nocloud-net;s=http://192.168.$GUEST_NETWORK.1:4567/ ubuntu-pass=ubuntu</cmdline>
+  </os>
+  <features>
+    <acpi/>
+  </features>
+  <clock offset='utc'/>
+  <vcpu>$GUEST_CORES</vcpu>
+  <devices>
+    <disk type='file'>
+      <driver type='qcow2'/>
+      <source file='$vm_dir/disk'/>
+      <target dev='vda' bus='virtio'/>
+    </disk>
+
+    <interface type='network'>
+      <source network='devstack-$GUEST_NETWORK'/>
+    </interface>
+        
+    <!-- The order is significant here.  File must be defined first -->
+    <serial type="file">
+      <source path='$vm_dir/console.log'/>
+      <target port='1'/>
+    </serial>
+
+    <console type='pty' tty='/dev/pts/2'>
+      <source path='/dev/pts/2'/>
+      <target port='0'/>
+    </console>
+
+    <serial type='pty'>
+      <source path='/dev/pts/2'/>
+      <target port='0'/>
+    </serial>
+
+    <graphics type='vnc' port='-1' autoport='yes' keymap='en-us' listen='0.0.0.0'/>
+  </devices>
+</domain>
+EOF
+
+
+rm -rf $vm_dir/uec
+cp -r $TOOLS_DIR/uec $vm_dir/uec
+
+# set metadata
+cat > $vm_dir/uec/meta-data<<EOF
+hostname: $GUEST_NAME
+instance-id: i-hop
+instance-type: m1.ignore
+local-hostname: $GUEST_NAME.local
+EOF
+
+# set metadata
+cat > $vm_dir/uec/user-data<<EOF
+#!/bin/bash
+# hostname needs to resolve for rabbit
+sed -i "s/127.0.0.1/127.0.0.1 \`hostname\`/" /etc/hosts
+apt-get update
+apt-get install git sudo -y
+git clone https://github.com/cloudbuilders/devstack.git
+cd devstack
+git remote set-url origin `cd $TOP_DIR; git remote show origin | grep Fetch | awk '{print $3}'`
+git fetch
+git checkout `git rev-parse HEAD`
+cat > localrc <<LOCAL_EOF
+ROOTSLEEP=0
+`cat $TOP_DIR/localrc`
+LOCAL_EOF
+./stack.sh
+EOF
+
+# (re)start a metadata service
+(
+  pid=`lsof -iTCP@192.168.$GUEST_NETWORK.1:4567 -n | awk '{print $2}' | tail -1`
+  [ -z "$pid" ] || kill -9 $pid
+)
+cd $vm_dir/uec
+python meta.py 192.168.$GUEST_NETWORK.1:4567 &
+
+# Create the instance
+virsh create $vm_dir/libvirt.xml
+
+# Tail the console log till we are done
+WAIT_TILL_LAUNCH=${WAIT_TILL_LAUNCH:-1}
+if [ "$WAIT_TILL_LAUNCH" = "1" ]; then
+    set +o xtrace
+    # Done creating the container, let's tail the log
+    echo
+    echo "============================================================="
+    echo "                          -- YAY! --"
+    echo "============================================================="
+    echo
+    echo "We're done launching the vm, about to start tailing the"
+    echo "stack.sh log. It will take a second or two to start."
+    echo
+    echo "Just CTRL-C at any time to stop tailing."
+
+    while [ ! -e "$vm_dir/console.log" ]; do
+      sleep 1
+    done
+
+    tail -F $vm_dir/console.log &
+
+    TAIL_PID=$!
+
+    function kill_tail() {
+        kill $TAIL_PID
+        exit 1
+    }
+
+    # Let Ctrl-c kill tail and exit
+    trap kill_tail SIGINT
+
+    echo "Waiting stack.sh to finish..."
+    while ! egrep -q '^stack.sh (completed|failed)' $vm_dir/console.log ; do
+        sleep 1
+    done
+
+    set -o xtrace
+
+    kill $TAIL_PID
+
+    if ! grep -q "^stack.sh completed in" $vm_dir/console.log; then
+        exit 1
+    fi
+    echo ""
+    echo "Finished - Zip-a-dee Doo-dah!"
+fi
diff --git a/tools/build_usb_boot.sh b/tools/build_usb_boot.sh
index fc5e969..e4dabc0 100755
--- a/tools/build_usb_boot.sh
+++ b/tools/build_usb_boot.sh
@@ -11,6 +11,26 @@
 OPWD=`pwd`
 PROGDIR=`dirname $0`
 
+# Clean up any resources that may be in use
+cleanup() {
+    set +o errexit
+
+    # Mop up temporary files
+    if [ -n "$DEST_DEV" ]; then
+        umount $DEST_DIR
+        rmdir $DEST_DIR
+    fi
+    if [ -n "$MNTDIR" -a -d "$MNTDIR" ]; then
+        umount $MNTDIR
+        rmdir $MNTDIR
+    fi
+
+    # Kill ourselves to signal any calling process
+    trap 2; kill -2 $$
+}
+
+trap cleanup SIGHUP SIGINT SIGTERM
+
 if [ -b $DEST_DIR ]; then
     # We have a block device, install syslinux and mount it
     DEST_DEV=$DEST_DIR
diff --git a/tools/get_uec_image.sh b/tools/get_uec_image.sh
index 935feba..7b95aab 100755
--- a/tools/get_uec_image.sh
+++ b/tools/get_uec_image.sh
@@ -14,6 +14,9 @@
 TOOLS_DIR=$(cd $(dirname "$0") && pwd)
 TOP_DIR=`cd $TOOLS_DIR/..; pwd`
 
+# exit on error to stop unexpected errors
+set -o errexit
+
 usage() {
     echo "Usage: $0 - Prepare Ubuntu images"
     echo ""
@@ -26,6 +29,32 @@
     exit 1
 }
 
+# Clean up any resources that may be in use
+cleanup() {
+    set +o errexit
+
+    # Mop up temporary files
+    if [ -n "$IMG_FILE_TMP" -a -e "$IMG_FILE_TMP" ]; then
+        rm -f $IMG_FILE_TMP
+    fi
+
+    # Release NBD devices
+    if [ -n "$NBD" ]; then
+        qemu-nbd -d $NBD
+    fi
+
+    # Kill ourselves to signal any calling process
+    trap 2; kill -2 $$
+}
+
+# apt-get wrapper to just get arguments set correctly
+function apt_get() {
+    local sudo="sudo"
+    [ "$(id -u)" = "0" ] && sudo="env"
+    $sudo DEBIAN_FRONTEND=noninteractive apt-get \
+        --option "Dpkg::Options::=--force-confold" --assume-yes "$@"
+}
+
 while getopts f:hmr: c; do
     case $c in
         f)  FORMAT=$OPTARG
@@ -89,11 +118,21 @@
                 ;;
 esac
 
+trap cleanup SIGHUP SIGINT SIGTERM SIGQUIT
+
+# Check for dependencies
+
+if [ ! -x "`which qemu-img`" -o ! -x "`which qemu-nbd`" ]; then
+    # Missing KVM?
+    apt_get install qemu-kvm
+fi
+
 # Prepare the base image
 
 # Get the UEC image
 UEC_NAME=$DIST_NAME-server-cloudimg-amd64
 if [ ! -e $CACHEDIR/$UEC_NAME-disk1.img ]; then
+    mkdir -p $CACHEDIR
     (cd $CACHEDIR && wget -N http://uec-images.ubuntu.com/$DIST_NAME/current/$UEC_NAME-disk1.img)
 fi
 
@@ -111,25 +150,33 @@
     qemu-img resize $IMG_FILE_TMP +$((ROOTSIZE - 2000))M
 fi
 
+# Finds the next available NBD device
+# Exits script if error connecting or none free
+# map_nbd image
+# returns full nbd device path
+function map_nbd {
+    for i in `seq 0 15`; do
+        if [ ! -e /sys/block/nbd$i/pid ]; then
+            NBD=/dev/nbd$i
+            # Connect to nbd and wait till it is ready
+            qemu-nbd -c $NBD $1
+            if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
+                echo "Couldn't connect $NBD"
+                exit 1
+            fi
+            break
+        fi
+    done
+    if [ -z "$NBD" ]; then
+        echo "No free NBD slots"
+        exit 1
+    fi
+    echo $NBD
+}
+
 # Set up nbd
 modprobe nbd max_part=63
-for i in `seq 1 15`; do
-    if [ ! -e /sys/block/nbd$i/pid ]; then
-        NBD=/dev/nbd$i
-        # Connect to nbd and wait till it is ready
-        qemu-nbd -c $NBD $IMG_FILE_TMP
-        if ! timeout 60 sh -c "while ! [ -e ${NBD}p1 ]; do sleep 1; done"; then
-            echo "Couldn't connect $NBD"
-            exit 1
-        fi
-        break
-    fi
-done
-if [ -z "$NBD" ]; then
-    echo "No free NBD slots"
-    exit 1
-fi
-NBD_DEV=`basename $NBD`
+NBD=`map_nbd $IMG_FILE_TMP`
 
 # Resize partition 1 to full size of the disk image
 echo "d
@@ -162,5 +209,6 @@
 umount $MNTDIR
 rmdir $MNTDIR
 qemu-nbd -d $NBD
+NBD=""
 
 mv $IMG_FILE_TMP $IMG_FILE
diff --git a/tools/setup_stack_user.sh b/tools/setup_stack_user.sh
new file mode 100755
index 0000000..231a20f
--- /dev/null
+++ b/tools/setup_stack_user.sh
@@ -0,0 +1,74 @@
+#!/usr/bin/env bash
+
+# Echo commands
+set -o xtrace
+
+# Exit on error to stop unexpected errors
+set -o errexit
+
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+
+# Change dir to top of devstack
+cd $TOP_DIR
+
+# Echo usage
+usage() {
+    echo "Add stack user and keys"
+    echo ""
+    echo "Usage: $0 [full path to raw uec base image]"
+}
+
+# Make sure this is a raw image
+if ! qemu-img info $1 | grep -q "file format: raw"; then
+    usage
+    exit 1
+fi
+
+# Mount the image
+DEST=/opt/stack
+STAGING_DIR=/tmp/`echo $1 | sed  "s/\//_/g"`.stage.user
+mkdir -p $STAGING_DIR
+umount $STAGING_DIR || true
+sleep 1
+mount -t ext4 -o loop $1 $STAGING_DIR
+mkdir -p $STAGING_DIR/$DEST
+
+# Create a stack user that is a member of the libvirtd group so that stack
+# is able to interact with libvirt.
+chroot $STAGING_DIR groupadd libvirtd || true
+chroot $STAGING_DIR useradd stack -s /bin/bash -d $DEST -G libvirtd || true
+
+# Add a simple password - pass
+echo stack:pass | chroot $STAGING_DIR chpasswd
+
+# Configure sudo
+grep -q "^#includedir.*/etc/sudoers.d" $STAGING_DIR/etc/sudoers ||
+    echo "#includedir /etc/sudoers.d" | sudo tee -a $STAGING_DIR/etc/sudoers
+cp $TOP_DIR/files/sudo/* $STAGING_DIR/etc/sudoers.d/
+sed -e "s,%USER%,$USER,g" -i $STAGING_DIR/etc/sudoers.d/*
+
+# and has sudo ability (in the future this should be limited to only what
+# stack requires)
+echo "stack ALL=(ALL) NOPASSWD: ALL" >> $STAGING_DIR/etc/sudoers
+
+# Gracefully cp only if source file/dir exists
+function cp_it {
+    if [ -e $1 ] || [ -d $1 ]; then
+        cp -pRL $1 $2
+    fi
+}
+
+# Copy over your ssh keys and env if desired
+cp_it ~/.ssh $STAGING_DIR/$DEST/.ssh
+cp_it ~/.ssh/id_rsa.pub $STAGING_DIR/$DEST/.ssh/authorized_keys
+cp_it ~/.gitconfig $STAGING_DIR/$DEST/.gitconfig
+cp_it ~/.vimrc $STAGING_DIR/$DEST/.vimrc
+cp_it ~/.bashrc $STAGING_DIR/$DEST/.bashrc
+
+# Give stack ownership over $DEST so it may do the work needed
+chroot $STAGING_DIR chown -R stack $DEST
+
+# Unmount
+umount $STAGING_DIR
diff --git a/tools/uec/meta.py b/tools/uec/meta.py
new file mode 100644
index 0000000..5b845d8
--- /dev/null
+++ b/tools/uec/meta.py
@@ -0,0 +1,29 @@
+import sys
+from BaseHTTPServer import HTTPServer, BaseHTTPRequestHandler
+from SimpleHTTPServer import SimpleHTTPRequestHandler
+
+def main(host, port, HandlerClass = SimpleHTTPRequestHandler,
+         ServerClass = HTTPServer, protocol="HTTP/1.0"):
+    """simple http server that listens on a give address:port"""
+
+    server_address = (host, port)
+
+    HandlerClass.protocol_version = protocol
+    httpd = ServerClass(server_address, HandlerClass)
+
+    sa = httpd.socket.getsockname()
+    print "Serving HTTP on", sa[0], "port", sa[1], "..."
+    httpd.serve_forever()
+
+if __name__ == '__main__':
+    if sys.argv[1:]:
+        address = sys.argv[1]
+    else:
+        address = '0.0.0.0'
+    if ':' in address:
+        host, port = address.split(':')
+    else:
+        host = address
+        port = 8080
+
+    main(host, int(port))
diff --git a/tools/warm_apts_and_pips.sh b/tools/warm_apts_and_pips.sh
new file mode 100755
index 0000000..ec7e916
--- /dev/null
+++ b/tools/warm_apts_and_pips.sh
@@ -0,0 +1,53 @@
+#!/usr/bin/env bash
+
+# Echo commands
+set -o xtrace
+
+# Exit on error to stop unexpected errors
+set -o errexit
+
+# Keep track of the current directory
+TOOLS_DIR=$(cd $(dirname "$0") && pwd)
+TOP_DIR=`cd $TOOLS_DIR/..; pwd`
+
+# Change dir to top of devstack
+cd $TOP_DIR
+
+# Echo usage
+usage() {
+    echo "Cache OpenStack dependencies on a uec image to speed up performance."
+    echo ""
+    echo "Usage: $0 [full path to raw uec base image]"
+}
+
+# Make sure this is a raw image
+if ! qemu-img info $1 | grep -q "file format: raw"; then
+    usage
+    exit 1
+fi
+
+# Make sure we are in the correct dir
+if [ ! -d files/apts ]; then
+    echo "Please run this script from devstack/tools/"
+    exit 1
+fi 
+
+# Mount the image
+STAGING_DIR=/tmp/`echo $1 | sed  "s/\//_/g"`.stage
+mkdir -p $STAGING_DIR
+umount $STAGING_DIR || true
+sleep 1
+mount -t ext4 -o loop $1 $STAGING_DIR
+
+# Make sure that base requirements are installed
+cp /etc/resolv.conf $STAGING_DIR/etc/resolv.conf
+
+# Perform caching on the base image to speed up subsequent runs
+chroot $STAGING_DIR apt-get update
+chroot $STAGING_DIR apt-get install -y --download-only `cat files/apts/* | grep NOPRIME | cut -d\# -f1`
+chroot $STAGING_DIR apt-get install -y --force-yes `cat files/apts/* | grep -v NOPRIME | cut -d\# -f1` || true
+mkdir -p $STAGING_DIR/var/cache/pip
+PIP_DOWNLOAD_CACHE=/var/cache/pip chroot $STAGING_DIR pip install `cat files/pips/*` || true
+
+# Unmount
+umount $STAGING_DIR
diff --git a/tools/xen/build_domU.sh b/tools/xen/build_domU.sh
index 8e40225..6362849 100755
--- a/tools/xen/build_domU.sh
+++ b/tools/xen/build_domU.sh
@@ -7,6 +7,12 @@
     exit 1
 fi
 
+# This directory
+TOP_DIR=$(cd $(dirname "$0") && pwd)
+
+# Source params
+cd ../.. && source ./stackrc && cd $TOP_DIR
+
 # Echo commands
 set -o xtrace
 
@@ -41,9 +47,6 @@
 # Size of image
 VDI_MB=${VDI_MB:-2500}
 
-# This directory
-TOP_DIR=$(cd $(dirname "$0") && pwd)
-
 # Make sure we have git
 if ! which git; then
     GITDIR=/tmp/git-1.7.7
@@ -223,15 +226,21 @@
 SR_UUID=`xe sr-list --minimal name-label="Local storage"`
 xe sr-param-set uuid=$SR_UUID other-config:i18n-key=local-storage
 
-# Uninstall previous runs
-xe vm-list --minimal name-label="$LABEL" | xargs ./scripts/uninstall-os-vpx.sh
 
-# Destroy any instances that were launched
-for uuid in `xe vm-list | grep -1 instance | grep uuid | sed "s/.*\: //g"`; do
-    echo "Shutting down nova instance $uuid"
-    xe vm-shutdown uuid=$uuid
-    xe vm-destroy uuid=$uuid
-done
+# Shutdown previous runs
+DO_SHUTDOWN=${DO_SHUTDOWN:-1}
+if [ "$DO_SHUTDOWN" = "1" ]; then
+    # Shutdown all domU's that created previously
+    xe vm-list --minimal name-label="$LABEL" | xargs ./scripts/uninstall-os-vpx.sh
+
+    # Destroy any instances that were launched
+    for uuid in `xe vm-list | grep -1 instance | grep uuid | sed "s/.*\: //g"`; do
+        echo "Shutting down nova instance $uuid"
+        xe vm-unpause uuid=$uuid || true
+        xe vm-shutdown uuid=$uuid
+        xe vm-destroy uuid=$uuid
+    done
+fi
 
 # Path to head xva.  By default keep overwriting the same one to save space
 USE_SEPARATE_XVAS=${USE_SEPARATE_XVAS:-0}
diff --git a/tools/xen/build_domU_multi.sh b/tools/xen/build_domU_multi.sh
new file mode 100755
index 0000000..130bec5
--- /dev/null
+++ b/tools/xen/build_domU_multi.sh
@@ -0,0 +1,35 @@
+#!/usr/bin/env bash
+
+# Echo commands
+set -o xtrace
+
+# Head node host, which runs glance, api, keystone
+HEAD_PUB_IP=${HEAD_PUB_IP:-192.168.1.57}
+HEAD_MGT_IP=${HEAD_MGT_IP:-172.16.100.57}
+
+COMPUTE_PUB_IP=${COMPUTE_PUB_IP:-192.168.1.58}
+COMPUTE_MGT_IP=${COMPUTE_MGT_IP:-172.16.100.58}
+
+# Networking params
+FLOATING_RANGE=${FLOATING_RANGE:-192.168.1.196/30}
+
+# Variables common amongst all hosts in the cluster
+COMMON_VARS="$STACKSH_PARAMS MYSQL_HOST=$HEAD_MGT_IP RABBIT_HOST=$HEAD_MGT_IP GLANCE_HOSTPORT=$HEAD_MGT_IP:9292 FLOATING_RANGE=$FLOATING_RANGE"
+
+# Helper to launch containers
+function build_domU {
+    GUEST_NAME=$1 PUB_IP=$2 MGT_IP=$3 DO_SHUTDOWN=$4 TERMINATE=$TERMINATE STACKSH_PARAMS="$COMMON_VARS $5" ./build_domU.sh
+}
+
+# Launch the head node - headnode uses a non-ip domain name,
+# because rabbit won't launch with an ip addr hostname :(
+build_domU HEADNODE $HEAD_PUB_IP $HEAD_MGT_IP 1 "ENABLED_SERVICES=g-api,g-reg,key,n-api,n-sch,n-vnc,horizon,mysql,rabbit"
+
+# Wait till the head node is up
+while ! curl -L http://$HEAD_PUB_IP | grep -q username; do
+    echo "Waiting for head node ($HEAD_PUB_IP) to start..."
+    sleep 5
+done
+
+# Build the HA compute host
+build_domU $COMPUTE_PUB_IP $COMPUTE_PUB_IP $COMPUTE_MGT_IP 0 "ENABLED_SERVICES=n-cpu,n-net,n-api"