- Feb 28, 2019
-
-
Guillaume Abrioux authored
in addition to 15812970f033206b8680cc68351952d49cc18314 Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com> (cherry picked from commit d5be83e5042a5e22ace6250234ccd81acaffb0a2)
-
- Feb 20, 2019
-
-
Guillaume Abrioux authored
introduce two new variables to make the check that 'wait for all osd to be up' configurable. It's possible that for some deployments, OSDs can take longer to be seen as UP and IN. Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1676763 Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com> (cherry picked from commit 21e5db8982afd6e075541e7fc88620d59a1df498)
-
- Feb 19, 2019
-
-
David Waiting authored
The existing task checks that the number of OSDs is equal to the number of up OSDs before continuing. The problem is that if none of the OSDs have been discovered yet, the task will exit immediately and subsequent pool creation will fail (num_osds = 0, num_up_osds = 0). This is related to Bugzilla 1578086. In this change, we also check that at least one OSD is present. In our testing, this results in the task correctly waiting for all OSDs to come up before continuing. Signed-off-by:
David Waiting <david_waiting@comcast.com> (cherry picked from commit 3930791cb7d2872e3388d33713171d7a0c1951e8)
-
- Feb 06, 2019
-
-
Sébastien Han authored
In order to be able to retrieve udev information, we must expose its socket. As per, https://github.com/ceph/ceph/pull/25201 ceph-volume will start consuming udev output. Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit 997667a8734eddaa616fe642e57f6378408736a9)
-
Guillaume Abrioux authored
without this, the command `ceph-volume lvm list --format json` hangs and takes a very long time to complete. Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com> (cherry picked from commit 7ade0328072896e99817b070b6a82448024bfb84)
-
- Jan 16, 2019
-
-
Noah Watkins authored
the python3 fix merged by: https://github.com/ceph/ceph-ansible/pull/3346 was reintroduced a few days later by: https://github.com/ceph/ceph-ansible/commit/82a6b5adec4d72eb4b7219147f2225b7b2904460 and this patch fixes it again :) Signed-off-by:
Noah Watkins <nwatkins@redhat.com> (cherry picked from commit 3cf5fd2c3ee1fc342ac8dc3365ed82d863c7127e)
-
- Dec 20, 2018
-
-
Kai Wembacher authored
Signed-off-by:
Kai Wembacher <kai@ktwe.de> (cherry picked from commit a273ed7f6038b51d3ddb5198d4f3ab57d45bc328)
-
- Dec 04, 2018
-
-
Sébastien Han authored
Applying and passing the OSD_BLUESTORE/FILESTORE on the fly is wrong for existing clusters as their config will be changed. Typically, if an OSD was prepared with ceph-disk on filestore and we change the default objectstore to bluestore, the activation will fail. The flag osd_objectstore should only be used for the preparation, not activation. The activate in this case detects the osd objecstore which prevents failures like the one described above. Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit 4c5113019893c92c4d75c9fc457b04158b86398b)
-
Sébastien Han authored
If an existing cluster runs this config, and has ceph-disk OSD, the `expose_partitions` won't be expected by jinja since it's inside the 'old' if. We need it as part of the osd_scenario != 'lvm' condition. Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1640273 Signed-off-by:
Sébastien Han <seb@redhat.com> (cherry picked from commit bef522627e1e9827b86710c7a54f35a0cd596fbb)
-
- Nov 29, 2018
-
-
Sébastien Han authored
The code is now able (again) to start osds that where configured with ceph-disk on a non-container scenario. Closes: https://github.com/ceph/ceph-ansible/issues/3388 Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Guillaume Abrioux authored
Add real default value for osd pool size customization. Ceph itself has an `osd_pool_default_size` default value to `3`. If users don't specify a pool size in various pools definition within ceph-ansible, we should default to `3`. By the way, this kind of condition isn't really clear: ``` when: - rbd_pool_size | default ("") ``` we should try to get the customized value then default to what is in `osd_pool_default_size` (which has its default value pointing to `ceph_osd_pool_default_size` (`3`) as well) and compare it to `ceph_osd_pool_default_size`. Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com> (cherry picked from commit 7774069d45477df9f37c98bc414b3bf38cf41feb)
-
Guillaume Abrioux authored
`osd_pool_default_pg_num` parameter is set in `ceph-mon`. When using ceph-ansible with `--limit` on a specifc group of nodes, it will fail when trying to access this variables since it wouldn't be defined. Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1518696 Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com> (cherry picked from commit d4c0960f04342e995db2453b50940aa9933ceb09)
-
- Nov 28, 2018
-
-
Sébastien Han authored
This commit https://github.com/ceph/ceph-ansible/commit/4cc1506303739f13bb7a6e1022646ef90e004c90#diff-51bbe3572e46e3b219ad726da44b64ebL13 accidentally removed this check. This is a must have for ceph-disk based containerized OSDs. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Guillaume Abrioux authored
since `ceph-volume` introduction, there is no need to split those tasks. Let's refact this part of the code so it's clearer. By the way, this was breaking rolling_update.yml when `openstack_config: true` playbook because nothing ensured OSDs were started in ceph-osd role (In `openstack_config.yml` there is a check ensuring all OSD are UP which was obviously failing) and resulted with OSDs on the last OSD node not started anyway. Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com> (cherry picked from commit f7fcc012e9a5b5d37bcffd39f3062adbc2886006)
-
- Oct 29, 2018
-
-
Guillaume Abrioux authored
append 'm' suffix to specify the unit size used in all `*_docker_memory_limit`. Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com>
-
Neha Ojha authored
Since we do not have enough data to put valid upper bounds for the memory usage of these daemons, do not put artificial limits by default. This will help us avoid failures like OOM kills due to low default values. Whenever required, these limits can be manually enforced by the user. More details in https://bugzilla.redhat.com/show_bug.cgi?id=1638148 Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1638148 Signed-off-by:
Neha Ojha <nojha@redhat.com>
-
- Oct 22, 2018
-
-
Rishabh Dave authored
Fixes: https://bugzilla.redhat.com/show_bug.cgi?id=1596339 Signed-off-by:
Rishabh Dave <ridave@redhat.com>
-
- Oct 17, 2018
-
-
Sébastien Han authored
The playbook has various improvements: * run ceph-validate role before doing anything * run ceph-fetch-keys only on the first monitor of the inventory list * set noup flag so PGs get distributed once all the new OSDs have been added to the cluster and unset it when they are up and running Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1624962 Signed-off-by:
Sébastien Han <seb@redhat.com>
-
- Oct 12, 2018
-
-
Guillaume Abrioux authored
As of now, we should no longer support Jewel in ceph-ansible. The latest ceph-ansible release supporting Jewel is `stable-3.1`. Signed-off-by:
Guillaume Abrioux <gabrioux@redhat.com>
-
- Oct 10, 2018
-
-
Sébastien Han authored
This commit does a couple of things: * Avoid code duplication * Clarify the code * add more unit tests * add myself to the author of the module Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
This task was created for ceph-disk based deployments so it's not needed when osd are prepared with ceph-volume. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
We don't need to pass the device and discover the OSD ID. We have a task that gathers all the OSD ID present on that machine, so we simply re-use them and activate them. This also handles the situation when you have multiple OSDs running on the same device. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
We don't need to pass the hostname on the container name but we can keep it simple and just call it ceph-osd-$id. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
expose_partitions is only needed on ceph-disk OSDs so we don't need to activate this code when running lvm prepared OSDs. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
The batch option got recently added, while rebasing this patch it was necessary to implement it. So now, the batch option can work on containerized environments. Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1630977 Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
If we run on a containerized deployment we pass an env variable which contains the container image. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Sébastien Han authored
Signed-off-by:
Sébastien Han <seb@redhat.com>
-
Noah Watkins authored
Fixes the deprecation warning: [DEPRECATION WARNING]: Using tests as filters is deprecated. Instead of using `result|search` use `result is search`. Signed-off-by:
Noah Watkins <nwatkins@redhat.com>
-
- Oct 09, 2018
-
-
Andrew Schoen authored
Signed-off-by:
Andrew Schoen <aschoen@redhat.com>
-
- Oct 04, 2018
-
-
Rishabh Dave authored
Instead used "import_tasks" and "include_tasks" to tell whether tasks must be included statically or dynamically. Fixes: https://github.com/ceph/ceph-ansible/issues/2998 Signed-off-by:
Rishabh Dave <ridave@redhat.com>
-
- Sep 27, 2018
-
-
Rishabh Dave authored
Use "import_tasks" or "include_tasks" instead. Signed-off-by:
Rishabh Dave <ridave@redhat.com>
-
- Sep 12, 2018
-
-
Andrew Schoen authored
If this is set to anything other than the default value of 1 then the --osds-per-device flag will be used by the batch command to define how many osds will be created per device. Signed-off-by:
Andrew Schoen <aschoen@redhat.com>
-
- Aug 28, 2018
-
-
Sébastien Han authored
As promised, these will go unsupported for 3.1 so let's actually remove them :). Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1622729 Signed-off-by:
Sébastien Han <seb@redhat.com>
-
- Aug 20, 2018
-
-
Sébastien Han authored
We need ceph_release in the condition, not ceph_stable_release Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1619255 Signed-off-by:
Sébastien Han <seb@redhat.com>
-
- Aug 16, 2018
-
-
Sébastien Han authored
This reverts commit e84f11e99ef42057cd1c3fbfab41ef66cda27302. This commit was giving a new failure later during the rolling_update process. Basically, this was modifying the list of devices and started impacting the ceph-osd itself. The modification to accomodate the osd_auto_discovery parameter should happen outside of the ceph-osd. Also we are trying to not play ceph-osd role during the rolling_update process so we can speed up the upgrade. Signed-off-by:
Sébastien Han <seb@redhat.com>
-
- Aug 10, 2018
-
-
Andrew Schoen authored
devices and lvm_volumes will always be defined, so we need to instead check it's length before deciding to run the scenario. This fixes the failure here: https://2.jenkins.ceph.com/job/ceph-ansible-prs-luminous-bluestore_lvm_osds/86/consoleFull#1667273050b5dd38fa-a56e-4233-a5ca-584604e56e3a Signed-off-by:
Andrew Schoen <aschoen@redhat.com>
-
Sébastien Han authored
rolling_update relies on the list of devices when performing the restart of the OSDs. The task that is builind the devices list out of the ansible_devices dict only runs when there are no partitions on the drives. However during an upgrade the OSD are already configured, they have been prepared and have partitions so this task won't run and thus the devices list will be empty, skipping the restart during rolling_update. We now run the same task under different requirements when rolling_update is true and build a list when: * osd_auto_discovery is true * rolling_update is true * ansible_devices exists * no dm/lv are part of the discovery * the device is not removable * the device has more than 1 sector Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1613626 Signed-off-by:
Sébastien Han <seb@redhat.com>
-
- Aug 09, 2018
-
-
Andrew Schoen authored
This is used with the lvm osd scenario. When using devices you need the option to set the crush device class for all of the OSDs that are created from those devices. Signed-off-by:
Andrew Schoen <aschoen@redhat.com>
-
Andrew Schoen authored
This adds the action 'batch' to the ceph-volume module so that we can run the new 'ceph-volume lvm batch' subcommand. A functional test is also included. If devices is defind and osd_scenario is lvm then the 'ceph-volume lvm batch' command will be used to create the OSDs. Signed-off-by:
Andrew Schoen <aschoen@redhat.com>
-
- Jul 30, 2018
-
-
Sébastien Han authored
The container runs with --rm which means it will be deleted by Docker when exiting. Also 'docker rm -f' is not idempotent and returns 1 if the container does not exist. Closes: https://bugzilla.redhat.com/show_bug.cgi?id=1609007 Signed-off-by:
Sébastien Han <seb@redhat.com>
-