[Ocfs2-users] OCFS2 1.4.1 DLM unhandled error

Saul Gabay SaulG at herbalife.com
Tue Jun 16 17:39:45 PDT 2009


Confirmation 

________________________________

The following Service Request has been created:

 

SR Number

  7561997.993

Priority

  1

SR Submitted Date

  16-Jun-2009 17:38:17 GMT






This SR will be assigned to a support analyst during normal business
hours in your country

 

 

________________________________

From: Herbert van den Bergh [mailto:herbert.van.den.bergh at oracle.com] 
Sent: Tuesday, June 16, 2009 2:43 PM
To: Saul Gabay
Cc: Sunil Mushran; Server Ops_Linux; ocfs2-users at oss.oracle.com
Subject: Re: [Ocfs2-users] OCFS2 1.4.1 DLM unhandled error

 


Hello Saul,

Please log a Support Request via Metalink using your Oracle CSI.

Thanks,
Herbert,


Saul Gabay wrote: 

I reported this incident as a new BUG #1130.
 
Please treat this as urgent is affecting our production environment
repetitively.
 
Let me know if more information is needed.
 
Thank you
 
Saul
 
-----Original Message-----
From: Sunil Mushran [mailto:sunil.mushran at oracle.com] 
Sent: Tuesday, June 16, 2009 11:58 AM
To: Saul Gabay
Cc: ocfs2-users at oss.oracle.com; Server Ops_Linux
Subject: Re: [Ocfs2-users] OCFS2 1.4.1 DLM unhandled error
 
Please file a bugzilla in oss.oracle.com/bugzilla.
 
Saul Gabay wrote:
  

	We have a 2 node OCFS2 cluster running Oracle 10g, both nodes
crashed.
	 
	 
	 
	Node 1 because it panic running IOSTAT, the second node crashed
with 
	this error message you can see below.
	 
	 
	 
	I was hoping to see a newer version of OCFS2 so I could proceed
with 
	the upgrade if necessary.
	 
	 
	 
	Have you seen this problem, anybody has resolved this?
	 
	 
	 
	Node 1 reboots
	 
	reboot   system boot  2.6.18-92.el5    Tue Jun 16 09:26
	    

(02:14)
  

	 
	 
	Node 2 reboots
	 
	reboot   system boot  2.6.18-92.el5    Tue Jun 16 09:29
	    

(02:10)
  

	 
	 
	Running Kernel
	 
	Linux uscosprdvrtxdb02 2.6.18-92.el5 #1 SMP Tue Apr 29 13:16:15
EDT 
	2008 x86_64 x86_64 x86_64 GNU/Linux
	 
	 
	 
	OCFS2 Version installed
	 
	ocfs2console-1.4.1-1.el5
	 
	ocfs2-tools-1.4.1-1.el5
	 
	ocfs2-2.6.18-92.el5-1.4.1-1.el5
	 
	 
	 
	*_Crash analysis:_*
	 
	 
	 
	KERNEL: /usr/lib/debug/lib/modules/2.6.18-92.el5/vmlinux
	 
	    DUMPFILE: vmcore  [PARTIAL DUMP]
	 
	        CPUS: 8
	 
	        DATE: Tue Jun 16 09:15:26 2009
	 
	      UPTIME: 2 days, 02:17:01
	 
	LOAD AVERAGE: 0.22, 0.31, 0.21
	 
	       TASKS: 570
	 
	    NODENAME: uscosprdvrtxdb02
	 
	     RELEASE: 2.6.18-92.el5
	 
	     VERSION: #1 SMP Tue Apr 29 13:16:15 EDT 2008
	 
	     MACHINE: x86_64  (2666 Mhz)
	 
	      MEMORY: 11.8 GB
	 
	       *PANIC: ""*
	 
	*         PID: 28123*
	 
	*     COMMAND: "oracle"*
	 
	        TASK: ffff8102e25e97e0  [THREAD_INFO: ffff8102cf0ba000]
	 
	         CPU: 3
	 
	       *STATE: TASK_RUNNING (PANIC)*
	 
	 
	 
	*_ _*
	 
	*_Kernel messages:_*
	 
	*o2net: connection to node uscosprdvrtxdb01 (num 0) at 
	192.168.5.1:7000 has been idle for 60.0 seconds, shutting it
down.*
	 
	*(0,0):o2net_idle_timer:1476 here are some times that might help
debug
	    

 
  

	the situation: (tmr 1245143657.942607 now 1245143717.944198 dr 
	1245143657.942600 adv 12*
	 
	*45143657.942608:1245143657.942609 func (5010bc9a:505) 
	1245128670.144972:1245128670.144981)*
	 
	*o2net: no longer connected to node uscosprdvrtxdb01 (num 0) at 
	192.168.5.1:7000*
	 
	*(28123,3):dlm_do_master_request:1330 ERROR: unhandled 
	error!----------- [cut here ] --------- [please bite here ]
---------*
	 
	*Kernel BUG at
	    

...mushran/BUILD/ocfs2-1.4.1/fs/ocfs2/dlm/dlmmaster.c:1331*
  

	*invalid opcode: 0000 [1] SMP*
	 
	last sysfs file: 
	 
	    

/devices/pci0000:00/0000:00:05.0/0000:10:00.0/0000:11:01.0/0000:14:00.0/
0000:15:00.0/irq
  

	CPU 3
	 
	Modules linked in: nfs lockd fscache nfs_acl mptctl mptbase
ipmi_si(U)
	    

 
  

	ipmi_devintf(U) ipmi_msghandler(U) autofs4 hidp l2cap bluetooth 
	ocfs2(U) ocfs2_dlmfs(U
	 
	) ocfs2_dlm(U) ocfs2_nodemanager(U) configfs sunrpc hp_ilo(U)
bonding 
	ipv6 xfrm_nalgo crypto_api emcpdm(PU) emcpgpx(PU) emcpmpx(PU)
emcp(PU)
	    

 
  

	dm_mirror dm_mul
	 
	tipath dm_mod video sbs backlight i2c_ec i2c_core button battery

	asus_acpi acpi_memhotplug ac parport_pc lp parport i5000_edac
edac_mc 
	bnx2 sg serio_raw shpc
	 
	hp pcspkr usb_storage lpfc scsi_transport_fc cciss(U) sd_mod
scsi_mod 
	ext3 jbd uhci_hcd ohci_hcd ehci_hcd
	 
	*Pid: 28123, comm: oracle Tainted: P      2.6.18-92.el5 #1*
	 
	RIP: 0010:[<ffffffff88652f8a>]  [<ffffffff88652f8a>] 
	:ocfs2_dlm:dlm_do_master_request+0x2f1/0x61c
	 
	RSP: 0018:ffff8102cf0bba38  EFLAGS: 00010286
	 
	RAX: 000000000000003f RBX: 00000000fffffe00 RCX:
ffffffff802ec9a8
	 
	RDX: ffffffff802ec9a8 RSI: 0000000000000000 RDI:
ffffffff802ec9a0
	 
	RBP: ffff8101b98d3e40 R08: ffffffff802ec9a8 R09:
0000000000000046
	 
	R10: 0000000000000000 R11: 0000000000000080 R12:
0000000000000000
	 
	R13: ffff810316df5c00 R14: ffff810316df5c00 R15:
ffff8101bc0625c0
	 
	FS:  00002b806dfccc40(0000) GS:ffff81032ff24640(0000) 
	knlGS:0000000000000000
	 
	CS:  0010 DS: 0000 ES: 0000 CR0: 000000008005003b
	 
	CR2: 00000000086229b8 CR3: 00000002cf119000 CR4:
00000000000006e0
	 
	Process oracle (pid: 28123, threadinfo ffff8102cf0ba000, task 
	ffff8102e25e97e0)
	 
	Stack:  0000000000001f01 3030303030303057 3030303030303030 
	3435303061323030
	 
	 0061316437626364 0000000000000000 0000000000000000
0000000000000000
	 
	 0000000000000000 000000008865344a 0000000116df5c00
0000000000000000
	 
	Call Trace:
	 
	 [<ffffffff88658669>]
:ocfs2_dlm:dlm_get_lock_resource+0xa5e/0x1913
	 
	 [<ffffffff8005be70>] cache_alloc_refill+0x106/0x186
	 
	 [<ffffffff8865dde5>]
:ocfs2_dlm:dlm_wait_for_recovery+0xa1/0x116
	 
	 [<ffffffff88650c46>] :ocfs2_dlm:dlmlock+0x731/0x11f9
	 
	 [<ffffffff886a5ad0>] :ocfs2:ocfs2_cluster_unlock+0x240/0x2ad
	 
	 [<ffffffff80009523>] __d_lookup+0xb0/0xff
	 
	 [<ffffffff886a17d8>] :ocfs2:ocfs2_dentry_revalidate+0x111/0x259
	 
	 [<ffffffff886a69c1>] :ocfs2:ocfs2_init_mask_waiter+0x24/0x3d
	 
	 [<ffffffff8000cb46>] do_lookup+0x65/0x1d4
	 
	 [<ffffffff886a7e00>] :ocfs2:ocfs2_cluster_lock+0x354/0x7eb
	 
	 [<ffffffff886a9a5c>] :ocfs2:ocfs2_locking_ast+0x0/0x486
	 
	 [<ffffffff886acfd2>] :ocfs2:ocfs2_blocking_ast+0x0/0x2c1
	 
	 [<ffffffff801458b9>] snprintf+0x44/0x4c
	 
	 [<ffffffff886ac242>] :ocfs2:ocfs2_rw_lock+0x10f/0x1d6
	 
	 [<ffffffff886b0159>] :ocfs2:ocfs2_file_aio_read+0x128/0x394
	 
	 [<ffffffff886a75eb>]
:ocfs2:ocfs2_add_lockres_tracking+0x73/0x81
	 
	 [<ffffffff8000caa4>] do_sync_read+0xc7/0x104
	 
	 [<ffffffff886aedcc>] :ocfs2:ocfs2_init_file_private+0x4d/0x5a
	 
	 [<ffffffff8001e35e>] __dentry_open+0x101/0x1dc
	 
	 [<ffffffff8009dde2>] autoremove_wake_function+0x0/0x2e
	 
	 [<ffffffff80027338>] do_filp_open+0x2a/0x38
	 
	 [<ffffffff8000b337>] vfs_read+0xcb/0x171
	 
	 [<ffffffff800130a3>] sys_pread64+0x50/0x70
	 
	 [<ffffffff8005d229>] tracesys+0x71/0xe0
	 
	 [<ffffffff8005d28d>] tracesys+0xd5/0xe0
	 
	 
	 
	 
	 
	Code: 0f 0b 68 de 85 66 88 c2 33 05 48 b8 00 09 00 00 01 00 00
00
	 
	RIP  [<ffffffff88652f8a>]
:ocfs2_dlm:dlm_do_master_request+0x2f1/0x61c
	 
	 RSP <ffff8102cf0bba38>
	 
	* *
	 
	 
	 
	 
	 
	/*/Saul J. Gabay/*/**
	 
	//Sr. Linux Engineer//////
	 
	//IT Infrastructure & Operations////
	 
	//Herbalife International Inc.////
	 
	//310-410-9600 x24341//
	 
	//saulg at herbalife.com//
	 
	 
	 
	 
	    

------------------------------------------------------------------------
  

	_______________________________________________
	Ocfs2-users mailing list
	Ocfs2-users at oss.oracle.com
	http://oss.oracle.com/mailman/listinfo/ocfs2-users
	    

 
 
_______________________________________________
Ocfs2-users mailing list
Ocfs2-users at oss.oracle.com
http://oss.oracle.com/mailman/listinfo/ocfs2-users
  
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20090616/6ee2a348/attachment-0001.html 


More information about the Ocfs2-users mailing list