Set up Ceph-Radosgw in the same model as an external S3 backend for
Glance:
- create an S3 user in Radosgw
- create an S3 bucket for Glance
- inject those parameters into Glance charm configurations as an
external S3 backend
- upload an image to the S3 backend via Glance, and check the size of
the image in the S3 bucket directly
Co-authored-by: Connor Chamberlain <connor.chamberlain@canonical.com>
Launchpad: #1911011
Gerrit review: https://review.opendev.org/c/openstack/charm-nova-compute/+/783371
* Functional tests for `instance-count` action on nova-compute charm.
Test for `remove_from_cloud_actions` waits at the end until
the nova-compute service is available again. However there's a few
seconds lag until that host shows up also in `hypervisor-list` which
can cause race conditions with other tests.
Add test for the new nova_cc charm action syncing AZs
The new action is meant to sync the availability zones from the
underlying Juju provider with the OpenStack availability zones.
It is useful within MAAS environments to map the MAAZ AZs to
OpenStack AZs.
Also fix output check logic for sync az test.
The validation logic for the sync az test checks the specific output of
the command text, but this is dependent on ordering. The code also
validates that the hypervisors are added to the appropriate host
aggregates, so this changes the test to verify the action was successful
and has some output.
Also use the availability-zone based on the configured settings.
In the attempt to sync the az test, the availability zone is assumed
to be the JUJU_AVAILABILITY_ZONE environment setting. In various
scenarios, this is not what is used for the availability zone of a
compute node.
This change mimics the logic that's contained within the nova-compute
charm to determine which availability zone should be used for the
compute node.
Signed-off-by: Billy Olsen <billy.olsen@gmail.com>
Co-authored-by: Ionut Balutoiu <ibalutoiu@cloudbasesolutions.com>
In newer versions of Ubuntu and the Cloud Archive (currently Hirsute,
Wallaby, and above), there is an openstack-release package that, if
installed, specifies the current OpenStack release in
/etc/openstack-release.
Also adds wallaby definitions to non-version definitions in
os_versions.py.
Fixes#560
Currently the new partition test_110 is breaking on master due to the
test landing before the charm PR which implemented the change [0] has
landed.
Move the partition test to its own class which can be added as a target
in the charm's tests.yaml. This will not affect master and can be
trivially updated in [0].
[0] https://review.opendev.org/c/openstack/charm-mysql-innodb-cluster/+/779427
Do not run the deferred hooks action for neutron gateway as it is
not needed and will end in an action error.
Tested locally here: https://paste.ubuntu.com/p/cSmgxmrTr4/
This closes issue #553
If there are multiple charms of a given application then a hook
like leader-settings-changed can trigger a deferred hook. So a
race can occur where the deferred hooks of unit1 are cleared
then in the process of clearing unit2 a deferred hook is
triggerred in unit1. To fix this do not block on the wl status
message after the hook test completes.
Also, make sure that an action failure is fatal to surface any
issues.
* Add ovn-chassis test
* Yet another refactor to reduce the amount that implementations need to override.
* Add ovn dedicated chassis support
* Fix race with checking wlm
The ServiceTest.test_start_specific under the ceph osd tests stops all
services using ceph-osd.target. This will stop all the OSDs, but the
restarting of any single OSD will restart all the OSDs. This behavior
causes the charm test to fail.
Unfortunately, this wasn't caught in the initial review because the test
is skipped when there are fewer the 2 OSDs on a unit and the final test
run skipped the test due to this condition.
Fixes#542
Signed-off-by: Billy Olsen <billy.olsen@gmail.com>