devstack_custom/stack.sh

665 lines
25 KiB
Bash
Raw Normal View History

#!/usr/bin/env bash
2011-10-02 20:36:54 +00:00
# **stack.sh** is an opinionated openstack developer installation.
# This script installs and configures *nova*, *glance*, *dashboard* and *keystone*
2011-10-03 05:08:24 +00:00
# This script allows you to specify configuration options of what git
# repositories to use, enabled services, network configuration and various
# passwords. If you are crafty you can run the script on multiple nodes using
# shared settings for common resources (mysql, rabbitmq) and build a multi-node
# developer install.
2011-10-02 20:53:21 +00:00
2011-09-16 05:54:52 +00:00
# To keep this script simple we assume you are running on an **Ubuntu 11.04
2011-09-16 04:28:23 +00:00
# Natty** machine. It should work in a VM or physical server. Additionally we
# put the list of *apt* and *pip* dependencies and other configuration files in
# this repo. So start by grabbing this script and the dependencies.
2011-10-02 20:36:54 +00:00
# Learn more and get the most recent version at http://devstack.org
2011-09-16 05:19:42 +00:00
# Sanity Check
# ============
2011-10-02 20:53:21 +00:00
# Record the start time. This allows us to print how long this script takes to run.
2011-09-29 00:49:40 +00:00
START_TIME=`python -c "import time; print time.time()"`
2011-09-16 05:54:52 +00:00
# Warn users who aren't on natty, but allow them to override check and attempt
2011-09-16 05:19:42 +00:00
# installation with ``FORCE=yes ./stack``
if ! grep -q natty /etc/lsb-release; then
echo "WARNING: this script has only been tested on natty"
if [[ "$FORCE" != "yes" ]]; then
echo "If you wish to run this script anyway run with FORCE=yes"
exit 1
fi
fi
# stack.sh keeps the list of **apt** and **pip** dependencies in external
2011-09-16 18:27:43 +00:00
# files, along with config templates and other useful files. You can find these
# in the ``files`` directory (next to this script). We will reference this
2011-09-16 18:31:16 +00:00
# directory using the ``FILES`` variable in this script.
FILES=`pwd`/files
if [ ! -d $FILES ]; then
echo "ERROR: missing devstack/files - did you grab more than just stack.sh?"
2011-09-16 05:19:42 +00:00
exit 1
fi
2011-10-02 20:53:21 +00:00
# OpenStack is designed to be run as a regular user (Dashboard will fail to run
# as root, since apache refused to startup serve content from root user). If
# stack.sh is run as root, it automatically creates a stack user with
# sudo privileges and runs as that user.
2011-10-02 21:47:32 +00:00
2011-09-28 21:08:26 +00:00
if [[ $EUID -eq 0 ]]; then
echo "You are running this script as root."
2011-10-02 20:53:21 +00:00
2011-10-02 21:47:32 +00:00
# since this script runs as a normal user, we need to give that user
# ability to run sudo
apt-get update
apt-get install -y sudo
2011-10-02 20:53:21 +00:00
if ! getent passwd | grep -q stack; then
echo "Creating a user called stack"
2011-10-01 07:37:27 +00:00
useradd -U -G sudo -s /bin/bash -m stack
fi
2011-10-02 20:53:21 +00:00
echo "Giving stack user passwordless sudo priviledges"
2011-10-02 15:48:34 +00:00
echo "stack ALL=(ALL) NOPASSWD: ALL" >> /etc/sudoers
2011-10-02 20:53:21 +00:00
echo "Copying files to stack user"
cp -r -f `pwd` /home/stack/
THIS_DIR=$(basename $(dirname $(readlink -f $0)))
2011-10-02 20:53:21 +00:00
chown -R stack /home/stack/$THIS_DIR
echo "Running the script as stack in 3 seconds..."
sleep 3
exec su -c "cd /home/stack/$THIS_DIR/; bash stack.sh; bash" stack
exit 0
2011-09-28 21:08:26 +00:00
fi
2011-10-02 20:53:21 +00:00
# So that errors don't compound we exit on any errors so you see only the
# first error that occured.
set -o errexit
# Print the commands being run so that we can see the command that triggers
# an error. It is also useful for following allowing as the install occurs.
set -o xtrace
2011-09-28 21:08:26 +00:00
2011-09-16 05:19:42 +00:00
# Settings
# ========
2011-09-13 08:24:50 +00:00
# This script is customizable through setting environment variables. If you
# want to override a setting you can either::
#
# export MYSQL_PASS=anothersecret
# ./stack.sh
#
2011-09-16 05:19:42 +00:00
# You can also pass options on a single line ``MYSQL_PASS=simple ./stack.sh``
#
2011-09-29 00:18:23 +00:00
# Additionally, you can put any local variables into a ``localrc`` file, like::
#
# MYSQL_PASS=anothersecret
# MYSQL_USER=hellaroot
#
2011-09-16 05:19:42 +00:00
# We try to have sensible defaults, so you should be able to run ``./stack.sh``
# in most cases.
2011-10-03 05:08:24 +00:00
#
# We our settings from ``stackrc``. This file is distributed with devstack and
# contains locations for what repositories to use. If you want to use other
# repositories and branches, you can add your own settings with another file
# called ``localrc``
#
# If ``localrc`` exists, then ``stackrc`` will load those settings. This is
# useful for changing a branch or repostiory to test other versions. Also you
# can store your other settings like **MYSQL_PASS** or **ADMIN_PASSWORD** instead
# of letting devstack generate random ones for you.
source ./stackrc
2011-09-16 05:19:42 +00:00
# Destination path for installation ``DEST``
2011-09-27 02:50:43 +00:00
DEST=${DEST:-/opt/stack}
sudo mkdir -p $DEST
sudo chown `whoami` $DEST
2011-09-12 18:59:38 +00:00
# Set the destination directories for openstack projects
NOVA_DIR=$DEST/nova
DASH_DIR=$DEST/dash
GLANCE_DIR=$DEST/glance
KEYSTONE_DIR=$DEST/keystone
NOVACLIENT_DIR=$DEST/python-novaclient
OPENSTACKX_DIR=$DEST/openstackx
NOVNC_DIR=$DEST/noVNC
2011-09-14 03:07:44 +00:00
# Specify which services to launch. These generally correspond to screen tabs
2011-10-02 17:01:00 +00:00
ENABLED_SERVICES=${ENABLED_SERVICES:-g-api,g-reg,key,n-api,n-cpu,n-net,n-sch,n-vnc,dash,mysql,rabbit}
2011-10-02 20:53:21 +00:00
# Nova hypervisor configuration. We default to **kvm** but will drop back to
# **qemu** if we are unable to load the kvm module. Stack.sh can also install
# an **LXC** based system.
LIBVIRT_TYPE=${LIBVIRT_TYPE:-kvm}
2011-10-02 21:47:32 +00:00
# nova supports pluggable schedulers. ``SimpleScheduler`` should work in most
# cases unless you are working on multi-zone mode.
2011-10-02 20:53:21 +00:00
SCHEDULER=${SCHEDULER:-nova.scheduler.simple.SimpleScheduler}
2011-09-13 08:24:50 +00:00
# Use the first IP unless an explicit is set by ``HOST_IP`` environment variable
if [ ! -n "$HOST_IP" ]; then
2011-09-13 18:22:14 +00:00
HOST_IP=`LC_ALL=C /sbin/ifconfig | grep -m 1 'inet addr:'| cut -d: -f2 | awk '{print $1}'`
fi
2011-10-02 20:53:21 +00:00
# Nova Network Configuration
# --------------------------
2011-10-03 05:08:24 +00:00
# FIXME: more documentation about why these are important flags. Also
# we should make sure we use the same variable names as the flag names.
2011-09-20 16:39:50 +00:00
PUBLIC_INTERFACE=${PUBLIC_INTERFACE:-eth0}
2011-09-26 19:48:31 +00:00
FIXED_RANGE=${FIXED_RANGE:-10.0.0.0/24}
FIXED_NETWORK_SIZE=${FIXED_NETWORK_SIZE:-256}
2011-10-02 21:47:32 +00:00
FLOATING_RANGE=${FLOATING_RANGE:-172.24.4.1/28}
2011-09-25 20:41:22 +00:00
NET_MAN=${NET_MAN:-FlatDHCPManager}
2011-09-14 03:07:44 +00:00
EC2_DMZ_HOST=${EC2_DMZ_HOST:-$HOST_IP}
2011-09-20 16:39:50 +00:00
FLAT_NETWORK_BRIDGE=${FLAT_NETWORK_BRIDGE:-br100}
2011-10-02 21:47:32 +00:00
VLAN_INTERFACE=${VLAN_INTERFACE:-$PUBLIC_INTERFACE}
# Multi-host is a mode where each compute node runs its own network node. This
# allows network operations and routing for a VM to occur on the server that is
# running the VM - removing a SPOF and bandwidth bottleneck.
MULTI_HOST=${MULTI_HOST:-0}
2011-09-13 08:24:50 +00:00
# If you are using FlatDHCP on multiple hosts, set the ``FLAT_INTERFACE``
# variable but make sure that the interface doesn't already have an
# ip or you risk breaking things.
2011-10-03 05:08:24 +00:00
#
# **DHCP Warning**: If your flat interface device uses DHCP, there will be a
# hiccup while the network is moved from the flat interface to the flat network
# bridge. This will happen when you launch your first instance. Upon launch
# you will lose all connectivity to the node, and the vm launch will probably
# fail.
#
# If you are running on a single node and don't need to access the VMs from
# devices other than that node, you can set the flat interface to the same
# value as ``FLAT_NETWORK_BRIDGE``. This will stop the network hiccup from
# occuring.
FLAT_INTERFACE=${FLAT_INTERFACE:-eth0}
2011-10-02 21:47:32 +00:00
## FIXME(ja): should/can we check that FLAT_INTERFACE is sane?
2011-09-13 08:24:50 +00:00
2011-10-02 20:53:21 +00:00
# MySQL & RabbitMQ
# ----------------
# We configure Nova, Dashboard, Glance and Keystone to use MySQL as their
# database server. While they share a single server, each has their own
# database and tables.
# By default this script will install and configure MySQL. If you want to
# use an existing server, you can pass in the user/password/host parameters.
# You will need to send the same ``MYSQL_PASS`` to every host if you are doing
# a multi-node devstack installation.
2011-09-14 09:39:10 +00:00
MYSQL_USER=${MYSQL_USER:-root}
MYSQL_PASS=${MYSQL_PASS:-`openssl rand -hex 12`}
2011-09-14 03:07:44 +00:00
MYSQL_HOST=${MYSQL_HOST:-localhost}
2011-10-02 20:53:21 +00:00
2011-09-14 03:07:44 +00:00
# don't specify /db in this string, so we can use it for multiple services
2011-09-14 03:11:42 +00:00
BASE_SQL_CONN=${BASE_SQL_CONN:-mysql://$MYSQL_USER:$MYSQL_PASS@$MYSQL_HOST}
2011-09-14 03:07:44 +00:00
# Rabbit connection info
RABBIT_HOST=${RABBIT_HOST:-localhost}
RABBIT_PASSWORD=${RABBIT_PASSWORD:-`openssl rand -hex 12`}
2011-09-14 16:55:31 +00:00
# Glance connection info. Note the port must be specified.
2011-09-20 16:39:50 +00:00
GLANCE_HOSTPORT=${GLANCE_HOSTPORT:-$HOST_IP:9292}
2011-09-14 16:55:31 +00:00
2011-10-02 20:53:21 +00:00
# Keystone
# --------
# Service Token - Openstack components need to have an admin token
# to validate user tokens.
SERVICE_TOKEN=${SERVICE_TOKEN:-`uuidgen`}
ADMIN_PASSWORD=${ADMIN_PASSWORD:-`openssl rand -hex 12`}
# Install Packages
2011-09-13 08:24:50 +00:00
# ================
#
# Openstack uses a fair number of other projects.
2011-09-13 00:09:08 +00:00
# install apt requirements
2011-09-16 18:31:16 +00:00
sudo apt-get install -y -q `cat $FILES/apts/* | cut -d\# -f1 | grep -Ev "mysql-server|rabbitmq-server"`
2011-09-13 00:09:08 +00:00
# install python requirements
2011-09-16 18:31:16 +00:00
sudo PIP_DOWNLOAD_CACHE=/var/cache/pip pip install `cat $FILES/pips/*`
2011-09-13 00:09:08 +00:00
# git clone only if directory doesn't exist already. Since ``DEST`` might not
# be owned by the installation user, we create the directory and change the
# ownership to the proper user.
function git_clone {
if [ ! -d $2 ]; then
sudo mkdir $2
sudo chown `whoami` $2
git clone $1 $2
cd $2
# This checkout syntax works for both branches and tags
git checkout $3
fi
}
2011-09-13 00:09:08 +00:00
# compute service
git_clone $NOVA_REPO $NOVA_DIR $NOVA_BRANCH
2011-09-13 00:09:08 +00:00
# image catalog service
git_clone $GLANCE_REPO $GLANCE_DIR $GLANCE_BRANCH
2011-09-13 00:09:08 +00:00
# unified auth system (manages accounts/tokens)
git_clone $KEYSTONE_REPO $KEYSTONE_DIR $KEYSTONE_BRANCH
2011-09-13 00:09:08 +00:00
# a websockets/html5 or flash powered VNC console for vm instances
git_clone $NOVNC_REPO $NOVNC_DIR $NOVNC_BRANCH
2011-09-13 00:09:08 +00:00
# django powered web control panel for openstack
git_clone $DASH_REPO $DASH_DIR $DASH_BRANCH $DASH_TAG
2011-09-13 00:09:08 +00:00
# python client library to nova that dashboard (and others) use
git_clone $NOVACLIENT_REPO $NOVACLIENT_DIR $NOVACLIENT_BRANCH
# openstackx is a collection of extensions to openstack.compute & nova
2011-09-13 00:09:08 +00:00
# that is *deprecated*. The code is being moved into python-novaclient & nova.
git_clone $OPENSTACKX_REPO $OPENSTACKX_DIR $OPENSTACKX_BRANCH
2011-09-13 00:09:08 +00:00
# Initialization
2011-09-13 08:24:50 +00:00
# ==============
2011-09-16 23:28:13 +00:00
2011-09-13 00:09:08 +00:00
# setup our checkouts so they are installed into python path
2011-09-13 08:24:50 +00:00
# allowing ``import nova`` or ``import glance.client``
2011-09-13 20:37:50 +00:00
cd $NOVA_DIR; sudo python setup.py develop
cd $NOVACLIENT_DIR; sudo python setup.py develop
cd $KEYSTONE_DIR; sudo python setup.py develop
cd $GLANCE_DIR; sudo python setup.py develop
cd $OPENSTACKX_DIR; sudo python setup.py develop
cd $DASH_DIR/django-openstack; sudo python setup.py develop
cd $DASH_DIR/openstack-dashboard; sudo python setup.py develop
2011-09-13 00:09:08 +00:00
2011-09-16 05:19:42 +00:00
# Add a useful screenrc. This isn't required to run openstack but is we do
# it since we are going to run the services in screen for simple
2011-09-16 18:31:16 +00:00
cp $FILES/screenrc ~/.screenrc
2011-09-13 20:17:22 +00:00
2011-09-16 05:19:42 +00:00
## TODO: update current user to allow sudo for all commands in files/sudo/*
2011-09-13 00:09:08 +00:00
# Rabbit
# ---------
2011-10-02 21:47:32 +00:00
if [[ "$ENABLED_SERVICES" =~ "rabbit" ]]; then
# Install and start rabbitmq-server
2011-09-16 06:17:44 +00:00
sudo apt-get install -y -q rabbitmq-server
# change the rabbit password since the default is "guest"
sudo rabbitmqctl change_password guest $RABBIT_PASSWORD
fi
2011-09-16 04:28:23 +00:00
# Mysql
# ---------
2011-10-02 21:47:32 +00:00
2011-09-16 04:28:23 +00:00
if [[ "$ENABLED_SERVICES" =~ "mysql" ]]; then
2011-10-02 21:47:32 +00:00
# Seed configuration with mysql password so that apt-get install doesn't
# prompt us for a password upon install.
cat <<MYSQL_PRESEED | sudo debconf-set-selections
mysql-server-5.1 mysql-server/root_password password $MYSQL_PASS
mysql-server-5.1 mysql-server/root_password_again password $MYSQL_PASS
mysql-server-5.1 mysql-server/start_on_boot boolean true
MYSQL_PRESEED
# Install and start mysql-server
2011-09-16 06:17:44 +00:00
sudo apt-get -y -q install mysql-server
2011-09-16 04:28:23 +00:00
# Update the DB to give user $MYSQL_USER@% full control of the all databases:
sudo mysql -uroot -p$MYSQL_PASS -e "GRANT ALL PRIVILEGES ON *.* TO '$MYSQL_USER'@'%' identified by '$MYSQL_PASS';"
# Edit /etc/mysql/my.cnf to change bind-address from localhost (127.0.0.1) to any (0.0.0.0) and restart the mysql service:
sudo sed -i 's/127.0.0.1/0.0.0.0/g' /etc/mysql/my.cnf
sudo service mysql restart
fi
2011-09-13 08:24:50 +00:00
# Dashboard
# ---------
2011-10-02 21:47:32 +00:00
# Setup the django dashboard application to serve via apache/wsgi
2011-09-13 00:09:08 +00:00
if [[ "$ENABLED_SERVICES" =~ "dash" ]]; then
2011-09-16 04:28:23 +00:00
# Dash currently imports quantum even if you aren't using it. Instead
# of installing quantum we can create a simple module that will pass the
2011-09-16 04:28:23 +00:00
# initial imports
2011-10-02 21:47:32 +00:00
mkdir -p $DASH_DIR/openstack-dashboard/quantum || true
touch $DASH_DIR/openstack-dashboard/quantum/__init__.py
touch $DASH_DIR/openstack-dashboard/quantum/client.py
2011-10-02 21:47:32 +00:00
# ``local_settings.py`` is used to override dashboard default settings.
cp $FILES/dash_settings.py $DASH_DIR/openstack-dashboard/local/local_settings.py
2011-09-16 04:46:20 +00:00
2011-10-02 21:47:32 +00:00
cd $DASH_DIR/openstack-dashboard
dashboard/manage.py syncdb
# create an empty directory that apache uses as docroot
sudo mkdir -p $DASH_DIR/.blackhole
## Configure apache's 000-default to run dashboard
2011-09-16 18:31:16 +00:00
sudo cp $FILES/000-default.template /etc/apache2/sites-enabled/000-default
sudo sed -e "s,%USER%,$USER,g" -i /etc/apache2/sites-enabled/000-default
sudo sed -e "s,%DASH_DIR%,$DASH_DIR,g" -i /etc/apache2/sites-enabled/000-default
fi
2011-09-13 02:29:56 +00:00
2011-09-13 08:24:50 +00:00
# Glance
# ------
if [[ "$ENABLED_SERVICES" =~ "g-reg" ]]; then
2011-09-20 17:38:06 +00:00
GLANCE_IMAGE_DIR=$DEST/glance/images
# Delete existing images
rm -rf $GLANCE_IMAGE_DIR
# Use local glance directories
mkdir -p $GLANCE_IMAGE_DIR
# (re)create glance database
mysql -u$MYSQL_USER -p$MYSQL_PASS -e 'DROP DATABASE IF EXISTS glance;'
mysql -u$MYSQL_USER -p$MYSQL_PASS -e 'CREATE DATABASE glance;'
# Copy over our glance-registry.conf
GLANCE_CONF=$GLANCE_DIR/etc/glance-registry.conf
2011-09-16 18:31:16 +00:00
cp $FILES/glance-registry.conf $GLANCE_CONF
sudo sed -e "s,%SQL_CONN%,$BASE_SQL_CONN/glance,g" -i $GLANCE_CONF
sudo sed -e "s,%SERVICE_TOKEN%,$SERVICE_TOKEN,g" -i $GLANCE_CONF
sudo sed -e "s,%DEST%,$DEST,g" -i $GLANCE_CONF
2011-09-20 07:33:51 +00:00
GLANCE_API_CONF=$GLANCE_DIR/etc/glance-api.conf
cp $FILES/glance-api.conf $GLANCE_API_CONF
sudo sed -e "s,%DEST%,$DEST,g" -i $GLANCE_API_CONF
sudo sed -e "s,%SERVICE_TOKEN%,$SERVICE_TOKEN,g" -i $GLANCE_API_CONF
fi
2011-09-13 00:09:08 +00:00
2011-09-13 08:24:50 +00:00
# Nova
# ----
2011-10-02 21:47:32 +00:00
# We are going to use the sample http middleware configuration from the keystone
# project to launch nova. This paste config adds the configuration required
# for nova to validate keystone tokens - except we need to switch the config
# to use our admin token instead (instead of the token from their sample data).
sudo sed -e "s,999888777666,$SERVICE_TOKEN,g" -i $KEYSTONE_DIR/examples/paste/nova-api-paste.ini
2011-09-13 00:09:08 +00:00
if [[ "$ENABLED_SERVICES" =~ "n-cpu" ]]; then
2011-09-13 20:17:22 +00:00
2011-10-02 21:47:32 +00:00
# Virtualization Configuration
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~
# attempt to load modules: network block device - used to manage qcow images
sudo modprobe nbd || true
2011-10-02 17:11:28 +00:00
2011-10-02 21:47:32 +00:00
# Check for kvm (hardware based virtualization). If unable to load kvm,
# set the libvirt type to qemu. Note: many systems come with hardware
# virtualization disabled in BIOS.
2011-10-02 17:11:28 +00:00
if [[ "$LIBVIRT_TYPE" -eq "kvm" ]]; then
2011-10-02 21:47:32 +00:00
sudo modprobe kvm || true
2011-10-02 17:11:28 +00:00
if [ ! -e /dev/kvm ]; then
2011-10-02 21:47:32 +00:00
echo "WARNING: Switching to QEMU"
2011-10-02 17:11:28 +00:00
LIBVIRT_TYPE=qemu
fi
fi
2011-10-02 21:47:32 +00:00
# Install and configure **LXC** if specified. LXC is another approach to
# splitting a system into many smaller parts. LXC uses cgroups and chroot
# to simulate multiple systems.
2011-10-02 17:11:28 +00:00
if [[ "$LIBVIRT_TYPE" -eq "lxc" ]]; then
2011-10-02 17:17:31 +00:00
sudo apt-get install lxc -y
2011-10-02 21:47:32 +00:00
# lxc requires cgroups to be configured on /cgroup
2011-10-02 17:11:28 +00:00
sudo mkdir -p /cgroup
if ! grep -q cgroup /etc/fstab; then
2011-10-02 17:25:33 +00:00
echo none /cgroup cgroup cpuacct,memory,devices,cpu,freezer,blkio 0 0 | sudo tee -a /etc/fstab
2011-10-02 17:11:28 +00:00
fi
2011-10-02 21:47:32 +00:00
sudo mount /cgroup
2011-09-16 23:28:13 +00:00
fi
2011-10-02 17:11:28 +00:00
2011-09-16 21:54:48 +00:00
# User needs to be member of libvirtd group for nova-compute to use libvirt.
sudo usermod -a -G libvirtd `whoami`
# if kvm wasn't running before we need to restart libvirt to enable it
sudo /etc/init.d/libvirt-bin restart
2011-09-13 20:17:22 +00:00
2011-10-02 21:47:32 +00:00
# Instance Storage
# ~~~~~~~~~~~~~~~~
# Nova stores each instance in its own directory.
mkdir -p $NOVA_DIR/instances
2011-09-13 00:09:08 +00:00
# if there is a partition labeled nova-instances use it (ext filesystems
# can be labeled via e2label)
## FIXME: if already mounted this blows up...
if [ -L /dev/disk/by-label/nova-instances ]; then
sudo mount -L nova-instances $NOVA_DIR/instances
sudo chown -R `whoami` $NOVA_DIR/instances
fi
2011-10-02 21:47:32 +00:00
# Clean out the instances directory.
rm -rf $NOVA_DIR/instances/*
fi
if [[ "$ENABLED_SERVICES" =~ "n-net" ]]; then
# delete traces of nova networks from prior runs
2011-09-20 09:23:54 +00:00
sudo killall dnsmasq || true
rm -rf $NOVA_DIR/networks
mkdir -p $NOVA_DIR/networks
fi
2011-09-16 23:28:13 +00:00
function add_nova_flag {
echo "$1" >> $NOVA_DIR/bin/nova.conf
}
# (re)create nova.conf
rm -f $NOVA_DIR/bin/nova.conf
add_nova_flag "--verbose"
add_nova_flag "--nodaemon"
add_nova_flag "--scheduler_driver=$SCHEDULER"
2011-09-16 23:28:13 +00:00
add_nova_flag "--dhcpbridge_flagfile=$NOVA_DIR/bin/nova.conf"
add_nova_flag "--network_manager=nova.network.manager.$NET_MAN"
add_nova_flag "--my_ip=$HOST_IP"
2011-09-20 16:39:50 +00:00
add_nova_flag "--public_interface=$PUBLIC_INTERFACE"
add_nova_flag "--vlan_interface=$VLAN_INTERFACE"
2011-09-16 23:28:13 +00:00
add_nova_flag "--sql_connection=$BASE_SQL_CONN/nova"
add_nova_flag "--libvirt_type=$LIBVIRT_TYPE"
add_nova_flag "--osapi_extensions_path=$OPENSTACKX_DIR/extensions"
2011-09-16 23:28:13 +00:00
add_nova_flag "--vncproxy_url=http://$HOST_IP:6080"
add_nova_flag "--vncproxy_wwwroot=$NOVNC_DIR/"
add_nova_flag "--api_paste_config=$KEYSTONE_DIR/examples/paste/nova-api-paste.ini"
add_nova_flag "--image_service=nova.image.glance.GlanceImageService"
add_nova_flag "--ec2_dmz_host=$EC2_DMZ_HOST"
add_nova_flag "--rabbit_host=$RABBIT_HOST"
add_nova_flag "--rabbit_password=$RABBIT_PASSWORD"
2011-09-16 23:28:13 +00:00
add_nova_flag "--glance_api_servers=$GLANCE_HOSTPORT"
2011-09-20 16:39:50 +00:00
add_nova_flag "--flat_network_bridge=$FLAT_NETWORK_BRIDGE"
2011-09-16 23:28:13 +00:00
if [ -n "$FLAT_INTERFACE" ]; then
add_nova_flag "--flat_interface=$FLAT_INTERFACE"
fi
if [ -n "$MULTI_HOST" ]; then
add_nova_flag "--multi_host=$MULTI_HOST"
fi
2011-10-02 21:47:32 +00:00
# Nova Database
# ~~~~~~~~~~~~~
# All nova components talk to a central database. We will need to do this step
# only once for an entire cluster.
if [[ "$ENABLED_SERVICES" =~ "mysql" ]]; then
# (re)create nova database
mysql -u$MYSQL_USER -p$MYSQL_PASS -e 'DROP DATABASE IF EXISTS nova;'
mysql -u$MYSQL_USER -p$MYSQL_PASS -e 'CREATE DATABASE nova;'
2011-10-02 21:47:32 +00:00
# (re)create nova database
$NOVA_DIR/bin/nova-manage db sync
# create a small network
2011-09-29 00:18:23 +00:00
$NOVA_DIR/bin/nova-manage network create private $FIXED_RANGE 1 $FIXED_NETWORK_SIZE
# create some floating ips
$NOVA_DIR/bin/nova-manage floating create $FLOATING_RANGE
fi
2011-09-13 22:16:26 +00:00
# Keystone
# --------
if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
# (re)create keystone database
mysql -u$MYSQL_USER -p$MYSQL_PASS -e 'DROP DATABASE IF EXISTS keystone;'
mysql -u$MYSQL_USER -p$MYSQL_PASS -e 'CREATE DATABASE keystone;'
2011-09-13 00:09:08 +00:00
# FIXME (anthony) keystone should use keystone.conf.example
KEYSTONE_CONF=$KEYSTONE_DIR/etc/keystone.conf
2011-09-16 18:31:16 +00:00
cp $FILES/keystone.conf $KEYSTONE_CONF
sudo sed -e "s,%SQL_CONN%,$BASE_SQL_CONN/keystone,g" -i $KEYSTONE_CONF
2011-09-27 02:50:43 +00:00
sudo sed -e "s,%DEST%,$DEST,g" -i $KEYSTONE_CONF
2011-09-13 19:01:45 +00:00
2011-10-02 21:47:32 +00:00
# keystone_data.sh creates our admin user and our ``SERVICE_TOKEN``.
KEYSTONE_DATA=$KEYSTONE_DIR/bin/keystone_data.sh
cp $FILES/keystone_data.sh $KEYSTONE_DATA
sudo sed -e "s,%HOST_IP%,$HOST_IP,g" -i $KEYSTONE_DATA
sudo sed -e "s,%SERVICE_TOKEN%,$SERVICE_TOKEN,g" -i $KEYSTONE_DATA
sudo sed -e "s,%ADMIN_PASSWORD%,$ADMIN_PASSWORD,g" -i $KEYSTONE_DATA
# initialize keystone with default users/endpoints
BIN_DIR=$KEYSTONE_DIR/bin bash $KEYSTONE_DATA
fi
2011-09-13 00:09:08 +00:00
2011-09-13 08:24:50 +00:00
# Launch Services
# ===============
2011-09-13 02:29:56 +00:00
# nova api crashes if we start it with a regular screen command,
# so send the start command by forcing text into the window.
2011-09-13 20:17:22 +00:00
# Only run the services specified in ``ENABLED_SERVICES``
# our screen helper to launch a service in a hidden named screen
2011-09-13 02:29:56 +00:00
function screen_it {
NL=`echo -ne '\015'`
2011-09-13 18:28:56 +00:00
if [[ "$ENABLED_SERVICES" =~ "$1" ]]; then
screen -S nova -X screen -t $1
screen -S nova -p $1 -X stuff "$2$NL"
fi
2011-09-13 02:29:56 +00:00
}
# create a new named screen to run processes in
screen -d -m -S nova -t nova
sleep 1
# launch the glance registery service
if [[ "$ENABLED_SERVICES" =~ "g-reg" ]]; then
screen_it g-reg "cd $GLANCE_DIR; bin/glance-registry --config-file=etc/glance-registry.conf"
fi
# launch the glance api and wait for it to answer before continuing
if [[ "$ENABLED_SERVICES" =~ "g-api" ]]; then
screen_it g-api "cd $GLANCE_DIR; bin/glance-api --config-file=etc/glance-api.conf"
while ! wget -q -O- http://$GLANCE_HOSTPORT; do
echo "Waiting for g-api ($GLANCE_HOSTPORT) to start..."
sleep 1
done
fi
# launch the keystone and wait for it to answer before continuing
if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
screen_it key "cd $KEYSTONE_DIR && $KEYSTONE_DIR/bin/keystone --config-file $KEYSTONE_CONF -d"
while ! wget -q -O- http://127.0.0.1:5000; do
echo "Waiting for keystone to start..."
sleep 1
done
fi
# launch the nova-api and wait for it to answer before continuing
if [[ "$ENABLED_SERVICES" =~ "n-api" ]]; then
screen_it n-api "cd $NOVA_DIR && $NOVA_DIR/bin/nova-api"
while ! wget -q -O- http://127.0.0.1:8774; do
echo "Waiting for nova-api to start..."
sleep 1
done
fi
# Launching nova-compute should be as simple as running ``nova-compute`` but
# have to do a little more than that in our script. Since we add the group
# ``libvirtd`` to our user in this script, when nova-compute is run it is
# within the context of our original shell (so our groups won't be updated).
# We can send the command nova-compute to the ``newgrp`` command to execute
# in a specific context.
screen_it n-cpu "cd $NOVA_DIR && echo $NOVA_DIR/bin/nova-compute | newgrp libvirtd"
screen_it n-net "cd $NOVA_DIR && $NOVA_DIR/bin/nova-network"
screen_it n-sch "cd $NOVA_DIR && $NOVA_DIR/bin/nova-scheduler"
screen_it n-vnc "cd $NOVNC_DIR && ./utils/nova-wsproxy.py 6080 --web ."
screen_it dash "cd $DASH_DIR && sudo /etc/init.d/apache2 restart; sudo tail -f /var/log/apache2/error.log"
2011-09-13 00:09:08 +00:00
2011-09-13 08:24:50 +00:00
# Install Images
# ==============
2011-10-03 04:01:28 +00:00
# Upload a couple images to glance. **TTY** is a simple small image that use the
# lets you login to it with username/password of user/password. TTY is useful
# for basic functionality. We all include an Ubuntu cloud build of **Natty**.
# Natty uses cloud-init, supporting login via keypair and sending scripts as
2011-10-03 05:08:24 +00:00
# userdata.
#
# Read more about cloud-init at https://help.ubuntu.com/community/CloudInit
2011-10-03 03:42:56 +00:00
2011-10-03 04:01:28 +00:00
if [[ "$ENABLED_SERVICES" =~ "g-reg" ]]; then
# create a directory for the downloadedthe images tarballs.
2011-10-03 03:42:56 +00:00
mkdir -p $FILES/images
2011-10-03 05:08:24 +00:00
# Debug Image (TTY)
# -----------------
# Downloads the image (ami/aki/ari style), then extracts it. Upon extraction
# we upload to glance with the glance cli tool. TTY is a stripped down
# version of ubuntu.
2011-09-16 21:16:36 +00:00
if [ ! -f $FILES/tty.tgz ]; then
wget -c http://images.ansolabs.com/tty.tgz -O $FILES/tty.tgz
fi
# extract ami-tty/image, aki-tty/image & ari-tty/image
tar -zxf $FILES/tty.tgz -C $FILES/images
2011-10-03 05:08:24 +00:00
# Use glance client to add the kernel, ramdisk and finally the root
# filesystem. We parse the results of the uploads to get glance IDs of the
# ramdisk and kernel and use them for the root filesystem.
RVAL=`glance add -A $SERVICE_TOKEN name="tty-kernel" is_public=true container_format=aki disk_format=aki < $FILES/images/aki-tty/image`
KERNEL_ID=`echo $RVAL | cut -d":" -f2 | tr -d " "`
RVAL=`glance add -A $SERVICE_TOKEN name="tty-ramdisk" is_public=true container_format=ari disk_format=ari < $FILES/images/ari-tty/image`
RAMDISK_ID=`echo $RVAL | cut -d":" -f2 | tr -d " "`
2011-10-02 23:23:22 +00:00
glance add -A $SERVICE_TOKEN name="tty" is_public=true container_format=ami disk_format=ami kernel_id=$KERNEL_ID ramdisk_id=$RAMDISK_ID < $FILES/images/ami-tty/image
2011-10-03 03:42:56 +00:00
2011-10-03 05:08:24 +00:00
# Ubuntu 11.04 aka Natty
# ----------------------
2011-10-03 03:42:56 +00:00
2011-10-03 05:08:24 +00:00
# Downloaded from ubuntu enterprise cloud images. This
2011-10-03 03:42:56 +00:00
# image doesn't use the ramdisk functionality
if [ ! -f $FILES/natty.tgz ]; then
wget -c http://uec-images.ubuntu.com/natty/current/natty-server-cloudimg-amd64.tar.gz -O $FILES/natty.tgz
fi
tar -zxf $FILES/natty.tgz -C $FILES/images
RVAL=`glance add -A $SERVICE_TOKEN name="uec-natty-kernel" is_public=true container_format=aki disk_format=aki < $FILES/images/natty-server-cloudimg-amd64-vmlinuz-virtual`
KERNEL_ID=`echo $RVAL | cut -d":" -f2 | tr -d " "`
glance add -A $SERVICE_TOKEN name="uec-natty" is_public=true container_format=ami disk_format=ami kernel_id=$KERNEL_ID < $FILES/images/natty-server-cloudimg-amd64.img
fi
2011-09-16 04:28:23 +00:00
# Using the cloud
# ===============
# If you installed the dashboard on this server, then you should be able
# to access the site using your browser.
2011-09-16 04:28:23 +00:00
if [[ "$ENABLED_SERVICES" =~ "dash" ]]; then
echo "dashboard is now available at http://$HOST_IP/"
fi
# If keystone is present, you can point nova cli to this server
if [[ "$ENABLED_SERVICES" =~ "key" ]]; then
echo "keystone is serving at http://$HOST_IP:5000/v2.0/"
echo "examples on using novaclient command line is in exercise.sh"
echo "the default users are: admin and demo"
echo "the password: $ADMIN_PASSWORD"
2011-09-16 04:28:23 +00:00
fi
2011-09-29 00:49:40 +00:00
2011-10-03 05:08:24 +00:00
# Fin
# ===
2011-09-29 00:49:40 +00:00
# End our timer and give a timing summary
END_TIME=`python -c "import time; print time.time()"`
ELAPSED=`python -c "print $END_TIME - $START_TIME"`
echo "stack.sh completed in $ELAPSED seconds."