| # Copyright 2012 OpenStack Foundation |
| # All Rights Reserved. |
| # |
| # Licensed under the Apache License, Version 2.0 (the "License"); you may |
| # not use this file except in compliance with the License. You may obtain |
| # a copy of the License at |
| # |
| # http://www.apache.org/licenses/LICENSE-2.0 |
| # |
| # Unless required by applicable law or agreed to in writing, software |
| # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT |
| # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the |
| # License for the specific language governing permissions and limitations |
| # under the License. |
| |
| from boto import exception |
| |
| from tempest.common.utils import data_utils |
| from tempest.common.utils.linux.remote_client import RemoteClient |
| from tempest import config |
| from tempest import exceptions |
| from tempest.openstack.common import log as logging |
| from tempest.test import attr |
| from tempest.test import skip_because |
| from tempest.thirdparty.boto.test import BotoTestCase |
| from tempest.thirdparty.boto.utils.s3 import s3_upload_dir |
| from tempest.thirdparty.boto.utils.wait import re_search_wait |
| from tempest.thirdparty.boto.utils.wait import state_wait |
| |
| CONF = config.CONF |
| |
| LOG = logging.getLogger(__name__) |
| |
| |
| class InstanceRunTest(BotoTestCase): |
| |
| @classmethod |
| def setUpClass(cls): |
| super(InstanceRunTest, cls).setUpClass() |
| if not cls.conclusion['A_I_IMAGES_READY']: |
| raise cls.skipException("".join(("EC2 ", cls.__name__, |
| ": requires ami/aki/ari manifest"))) |
| cls.s3_client = cls.os.s3_client |
| cls.ec2_client = cls.os.ec2api_client |
| cls.zone = cls.ec2_client.get_good_zone() |
| cls.materials_path = CONF.boto.s3_materials_path |
| ami_manifest = CONF.boto.ami_manifest |
| aki_manifest = CONF.boto.aki_manifest |
| ari_manifest = CONF.boto.ari_manifest |
| cls.instance_type = CONF.boto.instance_type |
| cls.bucket_name = data_utils.rand_name("s3bucket-") |
| cls.keypair_name = data_utils.rand_name("keypair-") |
| cls.keypair = cls.ec2_client.create_key_pair(cls.keypair_name) |
| cls.addResourceCleanUp(cls.ec2_client.delete_key_pair, |
| cls.keypair_name) |
| bucket = cls.s3_client.create_bucket(cls.bucket_name) |
| cls.addResourceCleanUp(cls.destroy_bucket, |
| cls.s3_client.connection_data, |
| cls.bucket_name) |
| s3_upload_dir(bucket, cls.materials_path) |
| cls.images = {"ami": |
| {"name": data_utils.rand_name("ami-name-"), |
| "location": cls.bucket_name + "/" + ami_manifest}, |
| "aki": |
| {"name": data_utils.rand_name("aki-name-"), |
| "location": cls.bucket_name + "/" + aki_manifest}, |
| "ari": |
| {"name": data_utils.rand_name("ari-name-"), |
| "location": cls.bucket_name + "/" + ari_manifest}} |
| for image in cls.images.itervalues(): |
| image["image_id"] = cls.ec2_client.register_image( |
| name=image["name"], |
| image_location=image["location"]) |
| cls.addResourceCleanUp(cls.ec2_client.deregister_image, |
| image["image_id"]) |
| |
| for image in cls.images.itervalues(): |
| def _state(): |
| retr = cls.ec2_client.get_image(image["image_id"]) |
| return retr.state |
| state = state_wait(_state, "available") |
| if state != "available": |
| for _image in cls.images.itervalues(): |
| cls.ec2_client.deregister_image(_image["image_id"]) |
| raise exceptions.EC2RegisterImageException(image_id= |
| image["image_id"]) |
| |
| @attr(type='smoke') |
| def test_run_idempotent_instances(self): |
| # EC2 run instances idempotently |
| |
| def _run_instance(client_token): |
| reservation = self.ec2_client.run_instances( |
| image_id=self.images["ami"]["image_id"], |
| kernel_id=self.images["aki"]["image_id"], |
| ramdisk_id=self.images["ari"]["image_id"], |
| instance_type=self.instance_type, |
| client_token=client_token) |
| rcuk = self.addResourceCleanUp(self.destroy_reservation, |
| reservation) |
| return (reservation, rcuk) |
| |
| def _terminate_reservation(reservation, rcuk): |
| for instance in reservation.instances: |
| instance.terminate() |
| self.cancelResourceCleanUp(rcuk) |
| |
| reservation_1, rcuk_1 = _run_instance('token_1') |
| reservation_2, rcuk_2 = _run_instance('token_2') |
| reservation_1a, rcuk_1a = _run_instance('token_1') |
| |
| self.assertIsNotNone(reservation_1) |
| self.assertIsNotNone(reservation_2) |
| self.assertIsNotNone(reservation_1a) |
| |
| # same reservation for token_1 |
| self.assertEqual(reservation_1.id, reservation_1a.id) |
| |
| # Cancel cleanup -- since it's a duplicate, it's |
| # handled by rcuk1 |
| self.cancelResourceCleanUp(rcuk_1a) |
| |
| _terminate_reservation(reservation_1, rcuk_1) |
| _terminate_reservation(reservation_2, rcuk_2) |
| |
| reservation_3, rcuk_3 = _run_instance('token_1') |
| self.assertIsNotNone(reservation_3) |
| |
| # make sure we don't get the old reservation back |
| self.assertNotEqual(reservation_1.id, reservation_3.id) |
| |
| # clean up |
| _terminate_reservation(reservation_3, rcuk_3) |
| |
| @attr(type='smoke') |
| def test_run_stop_terminate_instance(self): |
| # EC2 run, stop and terminate instance |
| image_ami = self.ec2_client.get_image(self.images["ami"] |
| ["image_id"]) |
| reservation = image_ami.run(kernel_id=self.images["aki"]["image_id"], |
| ramdisk_id=self.images["ari"]["image_id"], |
| instance_type=self.instance_type) |
| rcuk = self.addResourceCleanUp(self.destroy_reservation, reservation) |
| |
| for instance in reservation.instances: |
| LOG.info("state: %s", instance.state) |
| if instance.state != "running": |
| self.assertInstanceStateWait(instance, "running") |
| |
| for instance in reservation.instances: |
| instance.stop() |
| LOG.info("state: %s", instance.state) |
| if instance.state != "stopped": |
| self.assertInstanceStateWait(instance, "stopped") |
| |
| for instance in reservation.instances: |
| instance.terminate() |
| self.cancelResourceCleanUp(rcuk) |
| |
| @attr(type='smoke') |
| def test_run_stop_terminate_instance_with_tags(self): |
| # EC2 run, stop and terminate instance with tags |
| image_ami = self.ec2_client.get_image(self.images["ami"] |
| ["image_id"]) |
| reservation = image_ami.run(kernel_id=self.images["aki"]["image_id"], |
| ramdisk_id=self.images["ari"]["image_id"], |
| instance_type=self.instance_type) |
| rcuk = self.addResourceCleanUp(self.destroy_reservation, reservation) |
| |
| for instance in reservation.instances: |
| LOG.info("state: %s", instance.state) |
| if instance.state != "running": |
| self.assertInstanceStateWait(instance, "running") |
| instance.add_tag('key1', value='value1') |
| |
| tags = self.ec2_client.get_all_tags() |
| self.assertEqual(tags[0].name, 'key1') |
| self.assertEqual(tags[0].value, 'value1') |
| |
| tags = self.ec2_client.get_all_tags(filters={'key': 'key1'}) |
| self.assertEqual(tags[0].name, 'key1') |
| self.assertEqual(tags[0].value, 'value1') |
| |
| tags = self.ec2_client.get_all_tags(filters={'value': 'value1'}) |
| self.assertEqual(tags[0].name, 'key1') |
| self.assertEqual(tags[0].value, 'value1') |
| |
| tags = self.ec2_client.get_all_tags(filters={'key': 'value2'}) |
| self.assertEqual(len(tags), 0, str(tags)) |
| |
| for instance in reservation.instances: |
| instance.remove_tag('key1', value='value1') |
| |
| tags = self.ec2_client.get_all_tags() |
| self.assertEqual(len(tags), 0, str(tags)) |
| |
| for instance in reservation.instances: |
| instance.stop() |
| LOG.info("state: %s", instance.state) |
| if instance.state != "stopped": |
| self.assertInstanceStateWait(instance, "stopped") |
| |
| for instance in reservation.instances: |
| instance.terminate() |
| self.cancelResourceCleanUp(rcuk) |
| |
| @skip_because(bug="1098891") |
| @attr(type='smoke') |
| def test_run_terminate_instance(self): |
| # EC2 run, terminate immediately |
| image_ami = self.ec2_client.get_image(self.images["ami"] |
| ["image_id"]) |
| reservation = image_ami.run(kernel_id=self.images["aki"]["image_id"], |
| ramdisk_id=self.images["ari"]["image_id"], |
| instance_type=self.instance_type) |
| |
| for instance in reservation.instances: |
| instance.terminate() |
| try: |
| instance.update(validate=True) |
| except ValueError: |
| pass |
| except exception.EC2ResponseError as exc: |
| if self.ec2_error_code.\ |
| client.InvalidInstanceID.NotFound.match(exc): |
| pass |
| else: |
| raise |
| else: |
| self.assertNotEqual(instance.state, "running") |
| |
| # NOTE(afazekas): doctored test case, |
| # with normal validation it would fail |
| @skip_because(bug="1182679") |
| @attr(type='smoke') |
| def test_integration_1(self): |
| # EC2 1. integration test (not strict) |
| image_ami = self.ec2_client.get_image(self.images["ami"]["image_id"]) |
| sec_group_name = data_utils.rand_name("securitygroup-") |
| group_desc = sec_group_name + " security group description " |
| security_group = self.ec2_client.create_security_group(sec_group_name, |
| group_desc) |
| self.addResourceCleanUp(self.destroy_security_group_wait, |
| security_group) |
| self.assertTrue( |
| self.ec2_client.authorize_security_group( |
| sec_group_name, |
| ip_protocol="icmp", |
| cidr_ip="0.0.0.0/0", |
| from_port=-1, |
| to_port=-1)) |
| self.assertTrue( |
| self.ec2_client.authorize_security_group( |
| sec_group_name, |
| ip_protocol="tcp", |
| cidr_ip="0.0.0.0/0", |
| from_port=22, |
| to_port=22)) |
| reservation = image_ami.run(kernel_id=self.images["aki"]["image_id"], |
| ramdisk_id=self.images["ari"]["image_id"], |
| instance_type=self.instance_type, |
| key_name=self.keypair_name, |
| security_groups=(sec_group_name,)) |
| self.addResourceCleanUp(self.destroy_reservation, |
| reservation) |
| volume = self.ec2_client.create_volume(1, self.zone) |
| self.addResourceCleanUp(self.destroy_volume_wait, volume) |
| instance = reservation.instances[0] |
| LOG.info("state: %s", instance.state) |
| if instance.state != "running": |
| self.assertInstanceStateWait(instance, "running") |
| |
| address = self.ec2_client.allocate_address() |
| rcuk_a = self.addResourceCleanUp(address.delete) |
| self.assertTrue(address.associate(instance.id)) |
| |
| rcuk_da = self.addResourceCleanUp(address.disassociate) |
| # TODO(afazekas): ping test. dependecy/permission ? |
| |
| self.assertVolumeStatusWait(volume, "available") |
| # NOTE(afazekas): it may be reports availble before it is available |
| |
| ssh = RemoteClient(address.public_ip, |
| CONF.compute.ssh_user, |
| pkey=self.keypair.material) |
| text = data_utils.rand_name("Pattern text for console output -") |
| resp = ssh.write_to_console(text) |
| self.assertFalse(resp) |
| |
| def _output(): |
| output = instance.get_console_output() |
| return output.output |
| |
| re_search_wait(_output, text) |
| part_lines = ssh.get_partitions().split('\n') |
| volume.attach(instance.id, "/dev/vdh") |
| |
| def _volume_state(): |
| volume.update(validate=True) |
| return volume.status |
| |
| self.assertVolumeStatusWait(_volume_state, "in-use") |
| re_search_wait(_volume_state, "in-use") |
| |
| # NOTE(afazekas): Different Hypervisor backends names |
| # differently the devices, |
| # now we just test is the partition number increased/decrised |
| |
| def _part_state(): |
| current = ssh.get_partitions().split('\n') |
| if current > part_lines: |
| return 'INCREASE' |
| if current < part_lines: |
| return 'DECREASE' |
| return 'EQUAL' |
| |
| state_wait(_part_state, 'INCREASE') |
| part_lines = ssh.get_partitions().split('\n') |
| |
| # TODO(afazekas): Resource compare to the flavor settings |
| |
| volume.detach() |
| |
| self.assertVolumeStatusWait(_volume_state, "available") |
| re_search_wait(_volume_state, "available") |
| LOG.info("Volume %s state: %s", volume.id, volume.status) |
| |
| state_wait(_part_state, 'DECREASE') |
| |
| instance.stop() |
| address.disassociate() |
| self.assertAddressDissasociatedWait(address) |
| self.cancelResourceCleanUp(rcuk_da) |
| address.release() |
| self.assertAddressReleasedWait(address) |
| self.cancelResourceCleanUp(rcuk_a) |
| |
| LOG.info("state: %s", instance.state) |
| if instance.state != "stopped": |
| self.assertInstanceStateWait(instance, "stopped") |
| # TODO(afazekas): move steps from teardown to the test case |
| |
| |
| # TODO(afazekas): Snapshot/volume read/write test case |