[Ocfs2-users] input / out error on some nodes

Eric Ren zren at suse.com
Tue Oct 20 20:02:15 PDT 2015


Hi Prabu,

I guess others like me are not familiar with this case that combine CEPH 
RBD and OCFS2.

We'd really like to help you. But I think ocfs2 developers cannot get 
any info about what happened
to ocfs2 from your descriptions.

So, I'm wondering if you can reproduce and tell us the steps. Once 
developers can reproduce it,
it's likely be resolved;-) BTW, any dmesg log about ocfs2 especially the 
initial error message and stack
back trace will be helpful!

Thanks,
Eric

On 10/20/15 17:29, gjprabu wrote:
> Hi
>
>         We are looking forward to your input on this.
>
> Regads
> Prabu
>
> --- On Fri, 09 Oct 2015 12:08:19 +0530 *gjprabu 
> <gjprabu at zohocorp.com>* wrote ----
>
>
>
>
>
>         Hi All,
>
>                  Anybody pls help me on this issue.
>
>         Regards
>         Prabu
>
>
>
>
>         ---- On Thu, 08 Oct 2015 12:33:57 +0530 *gjprabu
>         <gjprabu at zohocorp.com <mailto:gjprabu at zohocorp.com>>* wrote ----
>
>
>
>             Hi All,
>
>                    We have CEPH  RBD with OCFS2 mounted servers. we
>             are facing i/o errors simultaneously while move the data's
>             in the same disk (Copying is not having any problem).
>             Temporary we remount the partition and the issue get
>             resolved but after sometime problem again reproduced. If
>             anybody faced same issue. Please help us.
>
>             Note : We have total 5 Nodes, here two nodes working fine
>             other nodes are showing like below input/output error.
>
>             ls -althr
>             ls: cannot access MICKEYLITE_3_0_M4_1_TEST: Input/output
>             error
>             ls: cannot access MICKEYLITE_3_0_M4_1_OLD: Input/output error
>             total 0
>             d????????? ? ? ? ? ? MICKEYLITE_3_0_M4_1_TEST
>             d????????? ? ? ? ? ? MICKEYLITE_3_0_M4_1_OLD
>
>             cluster:
>                    node_count=5
>                    heartbeat_mode = local
>                    name=ocfs2
>
>             node:
>                     ip_port = 7777
>                     ip_address = 192.168.113.42
>                     number = 1
>                     name = integ-hm9
>                     cluster = ocfs2
>
>             node:
>                     ip_port = 7777
>                     ip_address = 192.168.112.115
>                     number = 2
>                     name = integ-hm2
>                     cluster = ocfs2
>
>             node:
>                     ip_port = 7777
>                     ip_address = 192.168.113.43
>                     number = 3
>                     name = integ-ci-1
>                     cluster = ocfs2
>             node:
>                     ip_port = 7777
>                     ip_address = 192.168.112.217
>                     number = 4
>                     name = integ-hm8
>                     cluster = ocfs2
>             node:
>                     ip_port = 7777
>                     ip_address = 192.168.112.192
>                     number = 5
>                     name = integ-hm5
>                     cluster = ocfs2
>
>
>             Regards
>             Prabu
>
>
>
>             _______________________________________________
>             Ocfs2-users mailing list
>             Ocfs2-users at oss.oracle.com
>             <mailto:Ocfs2-users at oss.oracle.com>
>             https://oss.oracle.com/mailman/listinfo/ocfs2-users
>
>
>
>
> _______________________________________________
> Ocfs2-users mailing list
> Ocfs2-users at oss.oracle.com
> https://oss.oracle.com/mailman/listinfo/ocfs2-users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://oss.oracle.com/pipermail/ocfs2-users/attachments/20151021/d1e95bfc/attachment.html 


More information about the Ocfs2-users mailing list