Please upgrade to a non-outdated version
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
mdadm (Ubuntu) |
Fix Released
|
Undecided
|
Dimitri John Ledkov | ||
Lucid |
Won't Fix
|
Undecided
|
Dimitri John Ledkov |
Bug Description
Binary package hint: mdadm
Hi,
The current upstream version of mdadm is 3.1.1. It would be great to have this version in lucid, which is an LTS release. It contains many nice, advanced features since the current version in lucid, which is 2.6.7. These include e.g.
- conversion between RAID levels
- changing of chunk size
- user space metadata updates
Thanks.
tags: | added: upgrade |
causticsoda (glenn-thomasfamily) wrote : | #1 |
description: | updated |
s34gull (s34gull) wrote : | #2 |
Another vote for adoption of 3.1.x.
tags: | added: lucid |
stephen howell (showell) wrote : | #3 |
Please upgrade to 3.1.x, the re-shape features are much improved and the defaults improve performance and data safety.
Ilya Barygin (randomaction) wrote : | #4 |
3.1.1 is in Debian unstable.
Vikram Dhillon (dhillon-v10) wrote : | #5 |
Seems like we might have to wait for some time: http://
Changed in mdadm (Ubuntu): | |
assignee: | nobody → Vikram Dhillon (dhillon-v10) |
status: | New → Confirmed |
Erik Reuter (misc71) wrote : | #6 |
Version 3.0.3 is in testing. If you cannot use the 3.1.1 in unstable, can't Ubuntu at least use the 3.0.3 in testing? Ubuntu is the only major linux distribution that is shipping with such an outdated version of mdadm.
Jools Wills (jools) wrote : | #7 |
Due to the state/age of mdadm and raid on ubuntu, I've been seriously considering switching back to debian. Currently I am running my own package of 3.1.2 based on ubuntu's package (with the hotplug\udev way of building arrays). I basically grabbed the ubuntu source, uupdated, and manually fixed up any rejected patches, and reversed any I thought were no longer needed.
Jools Wills (jools) wrote : | #8 |
(and i changed the path mdadm uses to store its map file when constructing arrays on boot from initramfs), although the better fix is to add the missing folder /var/run/mdadm to the mdadm initramfs script i guess.
Hi Jools,
wow, you know to roll packages, and fixed some of the longstanding bugs allready. Could you upload them or have them build in your Personal Package Archive (PPA) https:/
I think maybe the initramfs creation hook (/debian/
Jools Wills (jools) wrote : | #10 |
I wouldn't say I'm good at building packages, but I know how to update existing ones a little.. Didn't try getting the map files across from the initramfs to the root fs. Does mdadm require it once the array is already constructed ?
All right, maybe you could try if that ppa facility can build your modified source package into one that is installable with a ppa sources.list entry. I think mdadm --incremental needs the map file in the system to be successfull with later hot-plugging of raid devices. At least what I remember from when I last tried is that hot-plugging worked in initramfs, but not in the running system. I don't know, maybe initramfs already does some copying of /var/run and it will just work if you enable mdadm to create the mapfile there and it can can also find it there later.
Jools Wills (jools) wrote : | #12 |
I haven't sorted a ppa yet, but I stuck my package up here http://
My changes:
3.1.2 wants to put its map file by default at /var/run/map so
init-top script to make a /var/run folder for the map file (would be
better in initramfs init - but I only want to maintain one package here
and not mess with the important stuff)
init-bottom script to copy the map file to .initramfs in /dev (temporary
as root is read only at this point).
change in the init script that starts the monitor service to copy the
map file to /var/run/map
change in package postinst, to not try and run a non
existent /dev/MAKEDEV
If you try this out, please make sure you can recover your system and
install the old mdadm back. However, it works for me.
Jools packages http://
tags: | added: patch |
thometal (thometal) wrote : | #14 |
it should be at least updated to version 3.1.2 because of an annoying reshape/grow bug in 3.1.1
- Link to Jools temporary package site. Edit (110 bytes, text/plain)
Installed the 3.1.2 package on a machine and ...
it was the first time mdadm --incremental actually succeeded in re-adding a drive when it was attached to an ubuntu machine after booting. Thumbs up!
You wrote about /var/run/map, but in the running system I see mdadm created /var/run/mdadm.map and /var/run/mdadm only contains monitor.pid, is that intended?
I think instead of recreating the dir on every boot, it could be added to the initramfs staging area with the mdadm's /debian/
What I haven't checked is if udev is stopped when the initramfs map file is copied so no new devices show up and may be written to the old location until the real root fs is set up and used.
Jools Wills (jools) wrote : | #16 |
The first version copied the map file to /var/run/map. I then changed my mind as I saw ubuntu had a mdadm folder there, and I changed the location to /var/run/mdadm/map. I'm running the 3.1.2-0ubuntu2 version and the map file is working correctly from the mdadm folder.
3.1.2-0ubuntu2 here too, and if you ls /var/run/md* in the running system you don't have the mdadm.map file?
From the comments I am not sure if 3.1.2 is supposed to already have that unplug support or not, at least for me it does not yet unbind unplugged disks automatically. https:/
Jools Wills (jools) wrote : | #18 |
I have jools@aero:~$ ls -l /var/run/mdadm/map
-rw------- 1 root root 54 2010-04-21 22:27 /var/run/mdadm/map
Btw I did try making the initramfs folder from the hook, but it never worked. I wonder if it ignores empty folders, which is why i went for the init-top script instead.
$ mdadm --version
mdadm - v3.1.2 - 10th March 2010
$ls -l /var/run/md*
-rw------- 1 root root 108 2010-04-23 20:51 /var/run/mdadm.map
/var/run/mdadm:
-rw-r--r-- 1 root root 5 2010-04-23 20:51 monitor.pid
Jools Wills (jools) wrote : | #20 |
I'm slightly confused then since my scripts specifically copy to /var/run/mdadm/map. mdadm.map is the third choice of mdadm should it not be able to make files in the first location that is set in the makefile config. however even if it did make a map file there in the initramfs, my script wouldnt copy it.
my initramfs init-bottom has "cp /var/run/mdadm/map /dev/.initramfs
and my /etc/init.d/mdadm has
MAPFILE=
if [ -f $MAPFILE ]; then
mv $MAPFILE /var/run/mdadm/map
fi
so I don't know how you get the mdadm.map file, unless its some odd thing with mdadm reading my earlier map file and rewriting that one ?
ceg (ceg) wrote : Re: [Bug 495370] Re: Please upgrade to 3.1.x for lucid | #21 |
Maybe udev rules fire before /etc/init.d/mdadm is processed.
But it also looks pretty empty under /dev/.initramfs on this 9.10
machine, so nothing is copied.
#ls -R /dev/.initramfs
/dev/.initramfs:
varrun
/dev/.initramfs
sendsigs.omit
Whats the best way to boot into a initramfs shell to see whats there?
ceg (ceg) wrote : Re: Please upgrade to 3.1.x for lucid | #22 |
Just found the --rebuild-map option in the new version. That may even be cleaner, to just rebuild the map in the main system before udev is restarted.
Upstream actually ships with udev hotplug rules by now, maybe those should be used instead of maintaining ubuntu ones.
With the successor to 3.1.2 mdadm seems to also ship unplug rules.
Debian's package repository has 3.1.2 and ships the upstream udev rules.
http://
Philip Muškovac (yofel) wrote : | #25 |
The attachement isn't a patch. Removing patch flag and tag and unsubscribing reviewers.
tags: | removed: patch |
Max Waterman (davidmaxwaterman) wrote : | #26 |
I've upgraded to lucid and I get a message complaining about a missing /var/run/mdadm/map when I boot up (blank screen with the message at the top).
$ sudo /sbin/mdadm --rebuild-map /dev/md2
$ ls /var/run/mdadm/map
ls: cannot access /var/run/mdadm/map: No such file or directory
$ ls -l /var/run/mdadm/
total 4
-rw-r--r-- 1 root root 5 2010-05-29 08:29 monitor.pid
$ ls -l /var/run/mdadm.map
-rw------- 1 root root 54 2010-05-29 08:29 /var/run/mdadm.map
Is there something I can do to get rid of this message?
Max.
ii mdadm 3.1.2-0ubuntu2
Vikram Dhillon (dhillon-v10) wrote : Re: [Bug 495370] Re: Please upgrade to 3.1.x for lucid | #27 |
On Sat, May 29, 2010 at 05:48:32AM -0000, Max Waterman wrote:
> I've upgraded to lucid and I get a message complaining about a missing
> /var/run/mdadm/map when I boot up (blank screen with the message at the
> top).
>
> $ sudo /sbin/mdadm --rebuild-map /dev/md2
> $ ls /var/run/mdadm/map
> ls: cannot access /var/run/mdadm/map: No such file or directory
> $ ls -l /var/run/mdadm/
> total 4
> -rw-r--r-- 1 root root 5 2010-05-29 08:29 monitor.pid
> $ ls -l /var/run/mdadm.map
> -rw------- 1 root root 54 2010-05-29 08:29 /var/run/mdadm.map
>
> Is there something I can do to get rid of this message?
>
> Max.
> ii mdadm 3.1.2-0ubuntu2
>
> --
> Please upgrade to 3.1.x for lucid
> https:/
> You received this bug notification because you are a bug assignee.
>
> Status in “mdadm” package in Ubuntu: Confirmed
>
> Bug description:
> Binary package hint: mdadm
>
> Hi,
>
> The current upstream version of mdadm is 3.1.1. It would be great to have this version in lucid, which is an LTS release. It contains many nice, advanced features since the current version in lucid, which is 2.6.7. These include e.g.
>
> - conversion between RAID levels
> - changing of chunk size
> - user space metadata updates
>
> Thanks.
>
>
>
>
I apologize for not being able to get this bug done in time, if someone else wants to take over, feel free to do so. Thanks.
assignee nobody
--
Regards,
Vikram Dhillon
Changed in mdadm (Ubuntu): | |
assignee: | Vikram Dhillon (dhillon-v10) → nobody |
ceg (ceg) wrote : Re: Please upgrade to 3.1.x for lucid | #28 |
Something is wrong with the non-existing maintenance of a basic system component like mdadm in ubuntu. Not even the updated debian packages get synced.
summary: |
- Please upgrade to 3.1.x for lucid + Please upgrade to a non-outdated version |
Simon Eisenmann (longsleep) wrote : | #29 |
Err .. this ticket needs some love.
Its a shame that lucid still has mdadm 2.6
What can be done to get the most recent version available to lucid users through ppa or backport and into the next ubuntu release?
Rune K. Svendsen (runeks) wrote : | #30 |
Would it be completely crazy to try to install mdadm 3.0.3-2 from Debian squeeze in Lucid?:
http://
Well I just tried it. It did spit out an error:
rune@
Selecting previously deselected package mdadm.
(Reading database ... 197151 files and directories currently installed.)
Unpacking mdadm (from mdadm_3.
Setting up mdadm (3.0.3-2) ...
Generating array device nodes... /var/lib/
failed.
Generating mdadm.conf... done.
update-
* Starting MD monitoring service mdadm --monitor [ OK ]
* Assembling MD arrays... [ OK ]
Processing triggers for ureadahead ...
Processing triggers for man-db ...
Processing triggers for doc-base ...
Processing 6 added doc-base file(s)...
Registering documents with scrollkeeper...
Processing triggers for initramfs-tools ...
update-
W: mdadm: /etc/mdadm/
W: mdadm: no arrays defined in configuration file.
Is this insanely dangerous to do? Or just somewhat dangerous?
I haven't had time to test it yet, but I will when I finish copying some files over to my new HDD.
Jools Wills (jools) wrote : | #31 |
yes its dangerous. debian most likely has completely difference initramfs scripts. you could leave your machine unbootable. If you want a newer version you can try my packages
Rune K. Svendsen (runeks) wrote : Re: [Bug 495370] Re: Please upgrade to a non-outdated version | #32 |
I see. Thanks for the heads up.
How would you rate the reliability of your package? Would you recommend
that I just stick with the version in the repos unless I'm feeling
adventurous, or would you say it's relatively safe to use your package?
I guess you'd be the one to know if you've touched some sensitive areas
in the package, in order to adapt it to Lucid...
Max Waterman describes an issue in comment #26, do you reckon that this
is something that is easily fixed?
> yes its dangerous. debian most likely has completely difference
> initramfs scripts. you could leave your machine unbootable. If you want
> a newer version you can try my packages
>
> http://
>
Jools Wills (jools) wrote : | #33 |
I have not adapted a debian package to lucid, rather upgraded the ubuntu package to a newer version and fixed some bugs. The package works fine for me at least. Not sure about issue 26. I'd have to have a look. I'm recently considering to "upgrade" to debian though on my system, as ubuntu is letting these core components stagnate.
Rune K. Svendsen (runeks) wrote : | #34 |
Cool. I'll give it a whirl as well then. I have backups of my data on a
non-RAID partition anyway.
I must admit that I was surprised as well, when I saw that Lucid shipped
with a 20 month old build of this component. Wouldn't this mean that the
Server edition uses this version as well? If that's the case, then it
must mean that very few people actually use Ubuntu for server use, or,
at least, that the ones who do, do not use RAID.
It would be nice to get some attention from the developers, to hear why
this hasn't progressed further.
On the other hand, only 14 people have marked this bug as affecting
them, so I guess it's not entirely inappropriate for Canonical to not
prioritize this more highly than they do.
> I have not adapted a debian package to lucid, rather upgraded the ubuntu
> package to a newer version and fixed some bugs. The package works fine
> for me at least. Not sure about issue 26. I'd have to have a look. I'm
> recently considering to "upgrade" to debian though on my system, as
> ubuntu is letting these core components stagnate.
>
Rune K. Svendsen (runeks) wrote : | #35 |
I just tried to install your package. I received an error, which was this:
cp: cannot stat `/lib/udev/
at the very end of the installation process. Not sure whether it has any significance.
Jools Wills (jools) wrote : | #36 |
Shouldn't happen. Might be safest right now to use the standard package until I have time to take a look.
Rune K. Svendsen (runeks) wrote : | #37 |
Woops... Too late :). It's in the process of syncing right now...
But as mentioned earlier, the data I'm going to put on the RAID
partition resides on a second, non-RAID partition. So as far as I can
tell I'm not risking anything.
But I _would_ appreciate you taking a look at it though...
The only "rules"-file that contain "mdadm" is the following file:
2>/dev/null
> Shouldn't happen. Might be safest right now to use the standard package
> until I have time to take a look.
>
Rune K. Svendsen (runeks) wrote : | #38 |
It seems that the file "65-mdadm.
I'm guessing your package is based on the Karmic version of mdadm?
Installing your package, this file is called "65-vol_id.rules", and is identical to both the file installed with Lucid's version of mdadm (2.6.7.
It seems the above error comes from the initramfs hook script that your package places in "/usr/share/
# copy the udev rules
for rules in 65-mdadm.
cp -p /lib/udev/
done
When the installation script runs "update-initramfs" (and when update-initramfs is run subsequently, independantly of the installation script), the error appears:
rune@
update-
cp: cannot stat `/lib/udev/
In "/usr/share/
So unless something more is lurking beneath the surface - I don't know enough about the inner workings of either mdadm of Linux in general to assess that - the fix seems to be fairly simple. Changing the above line in "/usr/share/
Jools Wills (jools) wrote : | #39 |
Yes it is a karmic package. Ii get a chance I could upgrade it to lucid. Your fix is probably fine though. What we need though is someone as part of ubuntu to maintain this though. I will never put ubuntu on a software raid machine again :/
RnSC (webclark) wrote : | #40 |
I plan on running 10.04 until a few months after the *next* LTS is released. I don't know how to interpret what I see here. Will the process result in a newer/better stable mdadm being released for 10.04 at some point? If so or if maybe, how far out is that likely to be? Thanks.
Jools Wills (jools) wrote : | #41 |
I finally got around to doing some upgrades on my fileserver. Thought about switching back to debian, but I thought for now, I'll just get my mdadm package working again properly on lucid.
http://
binaries for x86 are there.
you can build from source with dpkg-source -x mdadm_3.1.4-1.dsc then cd to the folder and dpkg-buildpackage
I have only tested this on a virtualbox setup, as I've not yet finished my upgrade. I include changes from lucid mdadm (2.6.7-1ubuntu15). the map file is solved now in mdadm itself as by default it chooses /dev/.mdadm as a location to put it which is available throughout the boot process.
No warranty with this of course, but I hope it is of some use.
Jools Wills (jools) wrote : | #42 |
Just to follow up, I am now running my 3.1.4 package on my fileserver. Would be interested to know if anyone else is using it, whilst we wait for an official mdadm update.
Amos Hayes (ahayes-polkaroo) wrote : | #43 |
Hi Jools. Thanks very much for the packages. I wanted to migrate a RAID 5 to RAID 6 on 10.04 and you saved the day. I too would like to see this picked up for a 10.04 backport or the like. :)
I was thrown off a bit when the device was renamed after installing/
My experience looked like:
# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdf[3] sde[2] sdd[1] sdc[0]
5860543488 blocks level 5, 128k chunk, algorithm 2 [4/4] [UUUU]
unused devices: <none>
# dpkg -i mdadm_3.
(Reading database ... 43484 files and directories currently installed.)
Preparing to replace mdadm 2.6.7.1-1ubuntu15 (using mdadm_3.
* Stopping MD monitoring service mdadm --monitor
...done.
Unpacking replacement mdadm ...
Setting up mdadm (3.1.4-1) ...
Generating array device nodes... Removing any system startup links for /etc/init.
update-initramfs: deferring update (trigger activated)
update-rc.d: warning: mdadm start runlevel arguments (2 3 4 5) do not match LSB Default-Start values (S)
update-rc.d: warning: mdadm stop runlevel arguments (0 1 6) do not match LSB Default-Stop values (0 6)
* Starting MD monitoring service mdadm --monitor
...done.
Processing triggers for man-db ...
Processing triggers for initramfs-tools ...
update-initramfs: Generating /boot/initrd.
# reboot
# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md127 : inactive sde[2] sdf[3] sdd[1] sdc[0]
7814057984 blocks
unused devices: <none>
# vgchange -an
# mdadm --stop /dev/md127
# mdadm --assemble /dev/md0 /dev/sdc /dev/sdd /dev/sde /dev/sdf
# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdc[0] sdf[3] sde[2] sdd[1]
5860543488 blocks level 5, 128k chunk, algorithm 2 [4/4] [UUUU]
unused devices: <none>
# mdadm --add /dev/md0 /dev/sdb
mdadm: added /dev/sdb
# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid5 sdb[4](S) sdc[0] sdf[3] sde[2] sdd[1]
5860543488 blocks level 5, 128k chunk, algorithm 2 [4/4] [UUUU]
unused devices: <none>
# mdadm --grow /dev/md0 --level=6 --backup-
mdadm level of /dev/md0 changed to raid6
# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5] [raid4] [raid10]
md0 : active raid6 sdb[4] sdc[0] sdf[3] sde[2] sdd[1]
5860543488 blocks super 0.91 level 6, 128k chunk, algorithm 18 [5/4] [UUUU_]
[
unused devices: <none>
Jools Wills (jools) wrote : | #44 |
Glad it's working for you. In regards to the device renaming, it's because it didn't recognise the array as belonging to the machine so it assembled it as md127 to not conflict with any other arrays. Please see
http://
You probably want to update your mdadm.conf and update initramfs again. I have
# definitions of existing MD arrays
ARRAY /dev/md/1 metadata=1.2 UUID=875d492f:
without the name=aero:1 (the line was generated with mdadm --detail --scan), I think my array came back as md127 also.
Rune K. Svendsen (runeks) wrote : | #45 |
- Output of "sudo dpkg -i mdadm_3.1.4-1_i386.deb" Edit (1.3 KiB, text/plain)
Hi Jools
Thanks for updating the package for Lucid. I have installed it without any issues. I attach a log of the output of dpkg.
Cristian Calin (cristi-calin) wrote : | #46 |
Hi Jools,
Do you think it is safe to build and install you package also on maverick ? I'm trying to ditch dmraid in favor of mdadm 3 which supports Intel Matrix metadata, and so far ubuntu isn't much help with it's outdated package.
Thanks,
Cristian
Jools Wills (jools) wrote : | #47 |
Funnily enough I just updated a machine to maverick with it to test, and it worked fine. btw i put it on a PPA https:/
Cristian Calin (cristi-calin) wrote : | #48 |
Thanks, I'll try it in a test system to see if I can migrate from dmraid to mdadm before I break my main system. I tried the same scenario with debian and it failed to initialize the matrix. After some tinkering I got it to assemble the matrix but when trying to mount the volume I ended up with a nice kernel panic in md.c. Hope the ubuntu kernel survives this one. I'll get back with a report.
Alf Gaida (agaida) wrote : | #49 |
I'm used the mdadm packages for i386/amd64 directly from debian/squeeze without problems in 10.04 and 10.10. I think it's time for Canonical to move. But this is not a problem for me anymore. I changend my distribution to debian/squeeze after that.
causticsoda (glenn-thomasfamily) wrote : | #50 |
Is there much chance of a updated version of mdadm for 11.04? I am going to want to migrate my RAID5 to RAID6 soon, and I am thinking of just switching to Debian, as I am not technically competent enough to install a newer version into Ubuntu myself, and I would rather use a supported version.
Cristian Calin (cristi-calin) wrote : | #51 |
I'm sorry to report that I experienced the same kernel panic with the ubuntu kernel, both generic and server flavors. Guess I'm stuck with dmraid for now ... or just move to SuSE (which works with mdadm 3.0.3).
RnSC (webclark) wrote : | #52 |
Jools, I think I about to take the plunge, but would like to ask a few questions first. I have been reading all of the mdadm wiki doc, manpages, scripts, rules, etc. until I am just beginning to *think* that I understand the problem somewhat, and am coming back around to the fact that you have done all of this. This is making me more comfortable with your package.
What versions is your PPA good for? Is this still built on the karmic or is it built on the lucid package now? (I want to run 10.04 until next LTS).
It seems like virtually all of the complexity is in init.d scripts and udev rules to have things happen automatically. If you were willing to just do an assemble manually and mount, or in a very few line script at startup, this could be trivially simple IF (big IF) you are not booting off of it. Do I have the right idea?
Further, it seems like the complexity for booting off of root is really not very complex, the difficulty is knowing where the few tentacles are that need to be twiddled in a fairly trivial way. Again, is this the right idea?
If you have not changed any of the 3.1.4 C source, just tracked down the script and initfs stuff and fixed it, then I would conclude that there should be very little risk in using your package... that you have done what needs to be done and we can rely on it. Anything residual problems would be in the little scripts and references automating things, not in the fundamental md/mdadm stuff, and that could be fixed manually with a few mdadm commands if I knew what I was doing better.
Would you please comment on all of this? I think we, thanks to you, may be on the verge of getting where we want to go or are there already.
Thanks. And even more, thanks for your work!
Jools Wills (jools) wrote : | #53 |
>What versions is your PPA good for?
It is for lucid. will also work on maverick. Don't use it on Karmic.
>It seems like virtually all of the complexity is in init.d scripts and udev rules to have things happen automatically. If you were willing to just do an assemble manually and mount, or in a very few line script at startup, this could be trivially simple IF (big IF) you are not booting off of it. Do I have the right idea?
the complexity is mostly down to handling hotplugging events correctly during bootup and so on.
>If you have not changed any of the 3.1.4 C source, just tracked down the script and initfs stuff and fixed it,
Originally I took the karmic ubuntu package, and applied the ubuntu diffs over the 3.1.2 source. fixing up anything that didn't apply cleanly. I then fixed a couple of issues regarding handling of the map file (adding some additional script and some minor .c modifications), and some other things. I then upgraded again to 3.1.4 and applied new patch from ubuntu lucid, and removed all my map handling stuff as it was no longer needed due to the new location for storing the map file. so now its basically the same as the lucid/maverick package, cept its up to date, and with a working map file that makes it from initramfs stage to root stage, as well as a fix for the MAKEDEV install error. only missing one patch they have put in very recently for copying mdadm.conf, which was "untested" which I thought I might want to test before i add it.
I was running my package on lucid, and now im running it on maverick. everything is working fine.
raas (andras-horvath-gmail) wrote : | #54 |
On Mon, Oct 18, 2010 at 4:40 AM, Jools Wills <email address hidden> wrote:
>
> the complexity is mostly down to handling hotplugging events correctly
> during bootup and so on.
Indeed. I have a box with 48 disks (on 6 controllers) in a number of
MD arrays and stock 10.04 can't assemble them during bootup, although
it manages just fine with hardware having only a few drives. :(
Andras
RnSC (webclark) wrote : | #55 |
Jools,
I am trying to sort this out, and another similar thread came to life today - 603582. It looks like you have both done the same work. Are you aware of his work, what is the difference? Is this duplicated effort?
It seems like it would be better for our cause (and admittedly any pre-release users like me) if the knowledge of both of your were pooled to form a single version for people to use in the short term, and to be rolled into a future Ubuntu release in the long term.
Thanks again.
--Ray
Michael Vogt (mvo) wrote : | #56 |
I'm testing the changes now and will upload after that. It would be great if we can start a conversation with debian if they can take some of our patches as it seems to be a pretty large delta and e.g. the degrated support is something that they probably want as well.
fermulator (fermulator) wrote : | #57 |
I'm re-opening this defect. It was marked as a duplicate of bug #603582, that bug was resolved, but it didn't back-port to 10.04 (lucid!) Still running 10.04.04 LTS
fermulator (fermulator) wrote : | #58 |
fermulator@
Ubuntu 10.04.4 LTS \n \l
fermulator@
Linux fermmy-server 2.6.32-
fermulator@
mdadm - v2.6.7.1 - 15th October 2008
Dimitri John Ledkov (xnox) wrote : | #59 |
You can request backports at https:/
Unfortunately such a major package upgrade does not qualify for a Stable Release Update.
Please do not reopen this bug, but instead follow the Backports request procedure as linked above.
Changed in mdadm (Ubuntu Lucid): | |
status: | New → Won't Fix |
assignee: | nobody → Dmitrijs Ledkovs (xnox) |
Changed in mdadm (Ubuntu): | |
status: | Confirmed → Fix Released |
assignee: | nobody → Dmitrijs Ledkovs (xnox) |
I agree it would be excellent to have a newer version of mdadm for lucid. I note that debian has 3.0.3 in "squeeze" http:// packages. debian. org/squeeze/ mdadm