different ceph-osd use the same journals on SSD
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Fuel for OpenStack |
Fix Released
|
High
|
Ryan Moe |
Bug Description
{"ostf_sha": "83ada35fec2664
Deployment ceph-osd node with 2 SSD (for journal).
After the successful deployment we have this situation
root@node-25:~# find /var/lib/ceph/osd/ -name journal -exec ls -la {} \;
lrwxrwxrwx 1 root root 9 Feb 15 11:15 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:15 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:15 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:15 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:15 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
lrwxrwxrwx 1 root root 9 Feb 15 11:14 /var/lib/
Every SSD partition are used twice by different ceph-osd processes.
That's another command
root@node-25:~# grep 'Running command' /root/ceph.log
2014-02-15 11:14:13,211 [node-25][INFO ] Running command: udevadm trigger --subsystem-
2014-02-15 11:14:13,290 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdd2 /dev/sda2
2014-02-15 11:14:15,696 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sde2 /dev/sda2
2014-02-15 11:14:19,596 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdf2 /dev/sda3
2014-02-15 11:14:23,102 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdg2 /dev/sda3
2014-02-15 11:14:26,725 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdh2 /dev/sda4
2014-02-15 11:14:30,207 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdi2 /dev/sda4
2014-02-15 11:14:32,951 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdj2 /dev/sda5
2014-02-15 11:14:35,935 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdk2 /dev/sda5
2014-02-15 11:14:38,422 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdl2 /dev/sda6
2014-02-15 11:14:42,011 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdm2 /dev/sda6
2014-02-15 11:14:43,990 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdn2 /dev/sda7
2014-02-15 11:14:45,742 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdo2 /dev/sda7
2014-02-15 11:14:48,628 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdp2 /dev/sda8
2014-02-15 11:14:50,352 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdq2 /dev/sda8
2014-02-15 11:14:52,977 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdr2 /dev/sda9
2014-02-15 11:14:55,607 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sds2 /dev/sda9
2014-02-15 11:14:58,955 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdt2 /dev/sdb2
2014-02-15 11:15:02,629 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdu2 /dev/sdb2
2014-02-15 11:15:04,643 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdv2 /dev/sdb3
2014-02-15 11:15:07,295 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdw2 /dev/sdb3
2014-02-15 11:15:09,395 [node-25][INFO ] Running command: ceph-disk-prepare --fs-type xfs --cluster ceph -- /dev/sdx2 /dev/sdb4
Also number of journal partitions is not enough. We have 22 ceph-osd and only 21 journal partition
root@node-25:~# parted /dev/sda print
Model: HP LOGICAL VOLUME (scsi)
Disk /dev/sda: 400GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt
Number Start End Size File system Name Flags
1 17.4kB 25.2MB 25.1MB primary bios_grub
2 25.2MB 10.8GB 10.7GB primary
3 10.8GB 21.5GB 10.7GB primary
4 21.5GB 32.2GB 10.7GB primary
5 32.2GB 43.0GB 10.7GB primary
6 43.0GB 53.7GB 10.7GB primary
7 53.7GB 64.4GB 10.7GB primary
8 64.4GB 75.2GB 10.7GB primary
9 75.2GB 85.9GB 10.7GB primary
10 85.9GB 96.7GB 10.7GB primary
11 96.7GB 107GB 10.7GB primary
12 107GB 118GB 10.7GB primary
root@node-25:~# parted /dev/sdb print
Model: HP LOGICAL VOLUME (scsi)
Disk /dev/sdb: 400GB
Sector size (logical/physical): 512B/512B
Partition Table: gpt
Number Start End Size File system Name Flags
1 17.4kB 25.2MB 25.1MB primary bios_grub
2 25.2MB 10.8GB 10.7GB primary
3 10.8GB 21.5GB 10.7GB primary
4 21.5GB 32.2GB 10.7GB primary
5 32.2GB 43.0GB 10.7GB primary
6 43.0GB 53.7GB 10.7GB primary
7 53.7GB 64.4GB 10.7GB primary
8 64.4GB 75.2GB 10.7GB primary
9 75.2GB 85.9GB 10.7GB primary
10 85.9GB 96.7GB 10.7GB primary
11 96.7GB 107GB 10.7GB primary
Changed in fuel: | |
importance: | Undecided → Critical |
assignee: | nobody → Dmitry Borodaenko (dborodaenko) |
milestone: | none → 4.1 |
description: | updated |
tags: | added: ceph customer-found |
Changed in fuel: | |
milestone: | 5.0 → 4.1 |
importance: | Medium → High |
assignee: | Fuel Python Team (fuel-python) → Ryan Moe (rmoe) |
tags: | added: release-notes |
Changed in fuel: | |
status: | Fix Committed → Fix Released |
I also have 3 nodes with 2 equal SSDs per server and 20 OSDs per server.
Andrey, does this actions are right before your script?
http:// ceph.com/ docs/master/ rados/troublesh ooting/ troubleshooting -osd/#stopping- w-out-rebalanci ng
My guess:
node1# ceph osd set noout dangerous_ script_ beware_ of_dragons. sh /dev/sdb /dev/sdw <--- SSDs
node1# for i in id1 id2 id3...; do ceph osd stop osd.$i; done
node1# ./ultimate_
node1# for i in id1 id2 id3...; do ceph osd start osd.$i; done
node1# ceph osd unset noout
And same for node2, node3.