[Ocfs2-users] input / out error on some nodes

gjprabu gjprabu at zohocorp.com
Thu Oct 22 06:00:29 PDT 2015


Hi Eric,



Thanks for your reply, Still we are facing same issue. we found this dmesg logs and this is known logs because our self made down node1 and made up this is showing in logs and other then we didn't found error message. Even we do have problem while unmounting. umount process goes to "D" stat and  fsck through fsck.ocfs2: I/O error. If required to run any other command pls let me know. 



ocfs2 version
debugfs.ocfs2 1.8.0



# cat /etc/sysconfig/o2cb

#

# This is a configuration file for automatic startup of the O2CB

# driver.  It is generated by running /etc/init.d/o2cb configure.

# On Debian based systems the preferred method is running

# 'dpkg-reconfigure ocfs2-tools'.

#



# O2CB_STACK: The name of the cluster stack backing O2CB.

O2CB_STACK=o2cb



# O2CB_BOOTCLUSTER: If not empty, the name of a cluster to start.

O2CB_BOOTCLUSTER=ocfs2



# O2CB_HEARTBEAT_THRESHOLD: Iterations before a node is considered dead.

O2CB_HEARTBEAT_THRESHOLD=31



# O2CB_IDLE_TIMEOUT_MS: Time in ms before a network connection is considered dead.

O2CB_IDLE_TIMEOUT_MS=30000



# O2CB_KEEPALIVE_DELAY_MS: Max time in ms before a keepalive packet is sent

O2CB_KEEPALIVE_DELAY_MS=2000



# O2CB_RECONNECT_DELAY_MS: Min time in ms between connection attempts

O2CB_RECONNECT_DELAY_MS=2000



# fsck.ocfs2 -fy /home/build/downloads/

fsck.ocfs2 1.8.0

fsck.ocfs2: I/O error on channel while opening "/zoho/build/downloads/"



dmesg logs
       

[ 4229.886284] o2dlm: Joining domain A895BC216BE641A8A7E20AA89D57E051 ( 5 ) 1 nodes

[ 4251.437451] o2dlm: Node 3 joins domain A895BC216BE641A8A7E20AA89D57E051 ( 3 5 ) 2 nodes

[ 4267.836392] o2dlm: Node 1 joins domain A895BC216BE641A8A7E20AA89D57E051 ( 1 3 5 ) 3 nodes

[ 4292.755589] o2dlm: Node 2 joins domain A895BC216BE641A8A7E20AA89D57E051 ( 1 2 3 5 ) 4 nodes

[ 4306.262165] o2dlm: Node 4 joins domain A895BC216BE641A8A7E20AA89D57E051 ( 1 2 3 4 5 ) 5 nodes

[316476.505401] (kworker/u192:0,95923,0):dlm_do_assert_master:1717 ERROR: Error -112 when sending message 502 (key 0xc3460ae7) to node 1

[316476.505470] o2cb: o2dlm has evicted node 1 from domain A895BC216BE641A8A7E20AA89D57E051

[316480.437231] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316480.442389] o2cb: o2dlm has evicted node 1 from domain A895BC216BE641A8A7E20AA89D57E051

[316480.442412] (kworker/u192:0,95923,20):dlm_begin_reco_handler:2765 A895BC216BE641A8A7E20AA89D57E051: dead_node previously set to 1, node 3 changing it to 1

[316480.541237] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316480.541241] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316485.542733] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316485.542740] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316485.542742] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316490.544535] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316490.544538] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316490.544539] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316495.546356] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316495.546362] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316495.546364] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316500.548135] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316500.548139] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316500.548140] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316505.549947] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316505.549951] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316505.549952] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316510.551734] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316510.551739] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316510.551740] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316515.553543] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316515.553547] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316515.553548] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316520.555337] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316520.555341] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316520.555343] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316525.557131] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316525.557136] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316525.557153] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316530.558952] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316530.558955] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316530.558957] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[316535.560781] o2dlm: Begin recovery on domain A895BC216BE641A8A7E20AA89D57E051 for node 1

[316535.560789] o2dlm: Node 3 (he) is the Recovery Master for the dead node 1 in domain A895BC216BE641A8A7E20AA89D57E051

[316535.560792] o2dlm: End recovery on domain A895BC216BE641A8A7E20AA89D57E051

[319419.525609] o2dlm: Node 1 joins domain A895BC216BE641A8A7E20AA89D57E051 ( 1 2 3 4 5 ) 5 nodes







ps -auxxxxx | grep umount

root     32083 21.8  0.0 125620  2828 pts/14   D+   19:37   0:18 umount /home/build/repository

root     32196  0.0  0.0 112652  2264 pts/8    S+   19:38   0:00 grep --color=auto umount





cat /proc/32083/stack 

[<ffffffff8132ad7d>] o2net_send_message_vec+0x71d/0xb00

[<ffffffff81352148>] dlm_send_remote_unlock_request.isra.2+0x128/0x410

[<ffffffff813527db>] dlmunlock_common+0x3ab/0x9e0

[<ffffffff81353088>] dlmunlock+0x278/0x800

[<ffffffff8131f765>] o2cb_dlm_unlock+0x35/0x50

[<ffffffff8131ecfe>] ocfs2_dlm_unlock+0x1e/0x30

[<ffffffff812a8776>] ocfs2_drop_lock.isra.29.part.30+0x1f6/0x700

[<ffffffff812ae40d>] ocfs2_simple_drop_lockres+0x2d/0x40

[<ffffffff8129b43c>] ocfs2_dentry_lock_put+0x5c/0x80

[<ffffffff8129b4a2>] ocfs2_dentry_iput+0x42/0x1d0

[<ffffffff81204dc2>] __dentry_kill+0x102/0x1f0

[<ffffffff81205294>] shrink_dentry_list+0xe4/0x2a0

[<ffffffff81205aa8>] shrink_dcache_parent+0x38/0x90

[<ffffffff81205b16>] do_one_tree+0x16/0x50

[<ffffffff81206e9f>] shrink_dcache_for_umount+0x2f/0x90

[<ffffffff811efb15>] generic_shutdown_super+0x25/0x100

[<ffffffff811eff57>] kill_block_super+0x27/0x70

[<ffffffff811f02a9>] deactivate_locked_super+0x49/0x60

[<ffffffff811f089e>] deactivate_super+0x4e/0x70

[<ffffffff8120da83>] cleanup_mnt+0x43/0x90

[<ffffffff8120db22>] __cleanup_mnt+0x12/0x20

[<ffffffff81093ba4>] task_work_run+0xc4/0xe0

[<ffffffff81013c67>] do_notify_resume+0x97/0xb0

[<ffffffff817d2ee7>] int_signal+0x12/0x17

[<ffffffffffffffff>] 0xffffffffffffffff



Regards

Prabu









 ---- On Wed, 21 Oct 2015 08:32:15 +0530 Eric Ren <zren at suse.com> wrote ----




Hi Prabu,

 

 I guess others like me are not familiar with this case that combine CEPH RBD and OCFS2.

 

 We'd really like to help you. But I think ocfs2 developers cannot get any info about what happened

 to ocfs2 from your descriptions. 

 

 So, I'm wondering if you can reproduce and tell us the steps. Once developers can reproduce it,

 it's likely be resolved;-) BTW, any dmesg log about ocfs2 especially the initial error message and stack

 back trace will be helpful!

 

 Thanks,

 Eric

 

On 10/20/15 17:29, gjprabu wrote:




Hi 



        We are looking forward to your input on this.



Regads

Prabu




--- On Fri, 09 Oct 2015 12:08:19 +0530 gjprabu <gjprabu at zohocorp.com> wrote ----
















Hi All,



         Anybody pls help me on this issue.



Regards

Prabu










---- On Thu, 08 Oct 2015 12:33:57 +0530 gjprabu <gjprabu at zohocorp.com> wrote ----











Hi All, 



       We have CEPH  RBD with OCFS2 mounted servers. we are facing i/o errors simultaneously while move the data's in the same disk (Copying is not having any problem). Temporary we remount the partition and the issue get resolved but after sometime problem again reproduced. If anybody faced same issue. Please help us.



Note : We have total 5 Nodes, here two nodes working fine other nodes are showing like below input/output error.



ls -althr 

ls: cannot access LITE_3_0_M4_1_TEST: Input/output error 

ls: cannot access LITE_3_0_M4_1_OLD: Input/output error 

total 0 

d????????? ? ? ? ? ? LITE_3_0_M4_1_TEST 

d????????? ? ? ? ? ? LITE_3_0_M4_1_OLD 



cluster:

       node_count=5

       heartbeat_mode = local

       name=ocfs2



node:

        ip_port = 7777

        ip_address = 192.168.113.42

        number = 1

        name = integ-hm9

        cluster = ocfs2



node:

        ip_port = 7777

        ip_address = 192.168.112.115

        number = 2

        name = integ-hm2

        cluster = ocfs2



node:

        ip_port = 7777

        ip_address = 192.168.113.43

        number = 3

        name = integ-ci-1

        cluster = ocfs2

node:

        ip_port = 7777

        ip_address = 192.168.112.217

        number = 4

        name = integ-hm8

        cluster = ocfs2

node:

        ip_port = 7777

        ip_address = 192.168.112.192

        number = 5

        name = integ-hm5

        cluster = ocfs2





Regards

Prabu









_______________________________________________ 

Ocfs2-users mailing list 

Ocfs2-users at oss.oracle.com 

https://oss.oracle.com/mailman/listinfo/ocfs2-users








 

_______________________________________________ Ocfs2-users mailing list Ocfs2-users at oss.oracle.com https://oss.oracle.com/mailman/listinfo/ocfs2-users




-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20151022/461b9261/attachment-0001.html 


More information about the Ocfs2-users mailing list