Crash in _nv026629rm+0x44/0x176 [nvidia]

Bug #696729 reported by Nils Gladitz
28
This bug affects 5 people
Affects Status Importance Assigned to Milestone
binutils
Invalid
Undecided
Unassigned
nvidia-graphics-drivers (Ubuntu)
Triaged
High
Unassigned

Bug Description

After a 2.6.35 kernel update my laptop seemed to hang when X was about to start.
I'm using nvidia-current (260.19.06-0ubuntu1) [Ubuntu 10.10 maverick].
I could remote login via ssh and got the following dmesg output:

[ 113.240956] BUG: unable to handle kernel paging request at ffffffffa1bad8aa
[ 113.240961] IP: [<ffffffffa0fdbdf8>] _nv026629rm+0x44/0x176 [nvidia]
[ 113.241116] PGD 1a2c067 PUD 1a30063 PMD 0
[ 113.241118] Oops: 0000 [#1] SMP
[ 113.241120] last sysfs file: /sys/bus/acpi/drivers/NVIDIA ACPI Video Driver/uevent
[ 113.241122] CPU 1
[ 113.241123] Modules linked in: aesni_intel cryptd aes_x86_64 aes_generic nvidia(P) dm_crypt snd_hda_codec_nvhdmi pata_pcmcia snd_hda_codec_idt tpm_infineon snd_hda_intel snd_hda_codec snd_hwdep snd_pcm snd_seq_midi snd_rawmidi snd_seq_midi_event snd_seq snd_timer snd_seq_device arc4 joydev pcmcia iwlagn tpm_tis uvcvideo iwlcore tpm r852 btusb ppdev tpm_bios sm_common hp_accel lis3lv02d input_polldev videodev parport_pc v4l1_compat mac80211 v4l2_compat_ioctl32 nand nand_ids nand_ecc hp_wmi bluetooth mtd snd yenta_socket pcmcia_rsrc pcmcia_core intel_ips psmouse serio_raw cfg80211 xhci_hcd soundcore snd_page_alloc lp parport firewire_ohci firewire_core sdhci_pci sdhci e1000e ahci led_class crc_itu_t libahci intel_agp video output [last unloaded: nvidia]
[ 113.241158]
[ 113.241160] Pid: 1758, comm: Xorg Tainted: P 2.6.35-24-generic #42-Ubuntu 1521/HP EliteBook 8540w
[ 113.241162] RIP: 0010:[<ffffffffa0fdbdf8>] [<ffffffffa0fdbdf8>] _nv026629rm+0x44/0x176 [nvidia]
[ 113.241259] RSP: 0018:ffff88021cefbaa8 EFLAGS: 00010282
[ 113.241260] RAX: ffff88021bc24000 RBX: ffff88022b998000 RCX: 0000000000000001
[ 113.241262] RDX: ffff88021bc24000 RSI: 0000000000000016 RDI: ffff880232dda000
[ 113.241263] RBP: ffff88021934df68 R08: ffff88021d668000 R09: ffff88021a843000
[ 113.241265] R10: 00000000ffffffff R11: 0000000000000077 R12: ffff88022b99c000
[ 113.241266] R13: ffff880232dda000 R14: ffff88021bc24000 R15: ffff88022ba89400
[ 113.241268] FS: 00007f648c931840(0000) GS:ffff880001e40000(0000) knlGS:0000000000000000
[ 113.241269] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
[ 113.241270] CR2: ffffffffa1bad8aa CR3: 00000002190b9000 CR4: 00000000000006e0
[ 113.241272] DR0: 0000000000000000 DR1: 0000000000000000 DR2: 0000000000000000
[ 113.241273] DR3: 0000000000000000 DR6: 00000000ffff0ff0 DR7: 0000000000000400
[ 113.241275] Process Xorg (pid: 1758, threadinfo ffff88021cefa000, task ffff88021b670000)
[ 113.241276] Stack:
[ 113.241277] ffffc90011a80000 ffff880232dda000 ffff88021934dfe8 ffff88022b99c000
[ 113.241279] <0> ffffc90011a80000 ffffffffa133fe7c ffff88021bc24000 ffff88022ba89400
[ 113.241281] <0> ffff880232dda000 ffffc90011a80000 ffff88022fdcc800 ffffffffa1340e36
[ 113.241284] Call Trace:
[ 113.241396] [<ffffffffa133fe7c>] ? _nv002152rm+0x213/0x256 [nvidia]
[ 113.241502] [<ffffffffa1340e36>] ? _nv002145rm+0x406/0x65a [nvidia]
[ 113.241609] [<ffffffffa13468d9>] ? rm_init_adapter+0x89/0xfd [nvidia]
[ 113.241715] [<ffffffffa1363fde>] ? nv_kern_open+0x5ae/0x760 [nvidia]
[ 113.241720] [<ffffffff81156daa>] ? chrdev_open+0x10a/0x200
[ 113.241722] [<ffffffff81156ca0>] ? chrdev_open+0x0/0x200
[ 113.241726] [<ffffffff81151205>] ? __dentry_open+0xe5/0x330
[ 113.241729] [<ffffffff81260c3f>] ? security_inode_permission+0x1f/0x30
[ 113.241731] [<ffffffff81151564>] ? nameidata_to_filp+0x54/0x70
[ 113.241734] [<ffffffff8115e1a8>] ? finish_open+0xe8/0x1d0
[ 113.241737] [<ffffffff81166e2f>] ? dput+0xdf/0x1b0
[ 113.241739] [<ffffffff8115f606>] ? do_last+0x86/0x460
[ 113.241741] [<ffffffff8116193b>] ? do_filp_open+0x21b/0x660
[ 113.241745] [<ffffffff8116d14a>] ? alloc_fd+0x10a/0x150
[ 113.241747] [<ffffffff81150fa9>] ? do_sys_open+0x69/0x170
[ 113.241749] [<ffffffff811510f0>] ? sys_open+0x20/0x30
[ 113.241753] [<ffffffff8100a0f2>] ? system_call_fastpath+0x16/0x1b
[ 113.241754] Code: 00 ba 00 00 00 00 be 3d 00 00 00 41 ff 55 20 48 89 c3 b9 01 00 00 00 ba 00 00 00 00 be 16 00 00 00 4c 89 ef 41 ff 55 20 49 89 c6 <48> 8b 05 ab 1a bd 00 48 89 45 10 8b 05 a9 1a bd 00 89 45 18 0f
[ 113.241771] RIP [<ffffffffa0fdbdf8>] _nv026629rm+0x44/0x176 [nvidia]
[ 113.241868] RSP <ffff88021cefbaa8>
[ 113.241869] CR2: ffffffffa1bad8aa
[ 113.241871] ---[ end trace 44086d2298d53c5b ]---

Nils Gladitz (gladitz)
description: updated
Revision history for this message
Nils Gladitz (gladitz) wrote :

This seems to be related: http://www.nvnews.net/vbulletin/showthread.php?p=2355003
Apparently the 4 users that currently complain about this problem are also using maverick so it might be ubuntu specific.

Revision history for this message
Magnus Hoff (maghoff) wrote :

This bug affects me as well. Is there anything I should do? Should I dump my dmesg here as well?

Revision history for this message
Nils Gladitz (gladitz) wrote :

sudo apt-get purge binutils-gold
sudo apt-get install nvidia-current --reinstall

Fixed this for me (tried on two affected systems ).

Revision history for this message
Magnus Hoff (maghoff) wrote :

Fix confirmed.

In other news, my nouveau drivers also stopped working, so I was convinced I had a hardware problem. I haven't investigated the nouveau issue any further.

Thank you. This is a huge load off my mind!

Revision history for this message
Paul Tagliamonte (paultag) wrote :

Fix found in the wild, this solution fixed it. Someone should get around to fixing this up. I'm going to triage it.

Changed in nvidia-graphics-drivers (Ubuntu):
status: New → Triaged
Changed in binutils:
status: New → Confirmed
status: Confirmed → New
Changed in nvidia-graphics-drivers (Ubuntu):
importance: Undecided → High
Revision history for this message
Paul Tagliamonte (paultag) wrote :

Rational for High: Causes the system to become unbootable.

Revision history for this message
Paul Tagliamonte (paultag) wrote :

I doubt the issue is binutils. I'm going to strip that. Someone add it if it's needed

Changed in binutils:
status: New → Invalid
Revision history for this message
Paul Tagliamonte (paultag) wrote :
Revision history for this message
Nils Gladitz (gladitz) wrote :

To clarify the reason why I added binutils to the bug report is that the problem only seems to show up when using the (experimental) gold linker to build the kernel module. The problem does not show up when using the "old" binutils ld.

Revision history for this message
Paul Tagliamonte (paultag) wrote :

Aye! That actually makes sense, but what binutils-gold does is reduce what's being brought into the link, so let's say with LD (old), if you include "A", it will auto-include "B" and "C". With binutils gold, it may not -- so you need to state you need "B" and "C" as well.

Since binutils is doing it's job "correctly", it's technically an issue with nvidia's build chain :)

Revision history for this message
Nils Gladitz (gladitz) wrote :

I'm not too sure but isn't what you mean shared library related and kernel modules don't have those?

To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.