<html>
  <head>
    <meta content="text/html; charset=ISO-8859-1"
      http-equiv="Content-Type">
  </head>
  <body bgcolor="#FFFFFF" text="#000000">
    crash looks similar to what patch
    <a class="moz-txt-link-freetext" href="https://oss.oracle.com/pipermail/ocfs2-devel/2012-January/008469.html">https://oss.oracle.com/pipermail/ocfs2-devel/2012-January/008469.html</a>
    trying to address. The fix is not yet accepted because as explained
    in the patch description we need to fix the master node to skip
    sending BAST after receiving unlock message.<br>
    <br>
    regarding ERROR: status = -17 what storage do you use? could be due
    to stale data.<br>
    <br>
    On 8/22/2012 2:25 AM, Pawel wrote:
    <blockquote cite="mid:5034A59D.7060205@mp.pl" type="cite">
      <meta content="text/html; charset=ISO-8859-1"
        http-equiv="Content-Type">
      <div class="moz-cite-prefix">It was done multiple times,<br>
        even more: system was recreated&nbsp; by mkfs.<br>
        Still the same behavior...<br>
        &nbsp;<br>
        <br>
        Pawel<br>
        &nbsp;<br>
        On 2012-08-22 04:21, Sunil Mushran wrote:<br>
      </div>
      <blockquote
cite="mid:CAEeiSHU6rPsrtfZLemJzUPSeNgq5q=0=SCgiQnCmMBPN0Bpnbg@mail.gmail.com"
        type="cite">You may want to run a full fsck on the fs.
        <div><br>
        </div>
        <div>fsck.ocfs2 -fy /dev/xxxx<br>
          <br>
          <div class="gmail_quote">On Tue, Aug 21, 2012 at 12:49 AM,
            Pawel <span dir="ltr">&lt;<a moz-do-not-send="true"
                href="mailto:pzlist@mp.pl" target="_blank">pzlist@mp.pl</a>&gt;</span>
            wrote:<br>
            <blockquote class="gmail_quote" style="margin:0 0 0
              .8ex;border-left:1px #ccc solid;padding-left:1ex">Hi,<br>
              After upgrading ocfs2 my cluster is instable.<br>
              <br>
              At least ones per week I can see:<br>
              kernel panic: Null pointer dereference &nbsp;at 00048<br>
              o2dlm_blocking_ast_wrapper + 0x8/0x20 [ocfs2_stack_o2cb]<br>
              stack:<br>
              dlm_do_local_bast [ocfs2_dlm]<br>
              dlm_lookup_lockers [ocfs2_dlm]<br>
              dlm_proxy_ast_handler<br>
              add_timer<br>
              ..<br>
              <br>
              After that sometimes deadlock happens on another nodes.
              Entire cluster<br>
              restart solve the issue.<br>
              I see in log:<br>
              (dlm_thread,7227,3):dlm_send_proxy_ast_msg:484 ERROR:<br>
              ECB9442E19A94EAC896641BFADD55E4B: res
              M0000000000000001f411c900000000,<br>
              error -107 send AST to node 4<br>
              (dlm_thread,7227,3):dlm_flush_asts:605 ERROR: status =
              -107<br>
              o2net: No connection established with node 4 after 10.0
              seconds, giving up.<br>
              o2net: No connection established with node 4 after 10.0
              seconds, giving up.<br>
              o2net: No connection established with node 4 after 10.0
              seconds, giving up.<br>
              (dlm_thread,7227,4):dlm_send_proxy_ast_msg:484 ERROR:<br>
              ECB9442E19A94EAC896641BFADD55E4B: res
              M0000000000000001f411c900000000,<br>
              error -107 send AST to node 4<br>
              (dlm_thread,7227,4):dlm_flush_asts:605 ERROR: status =
              -107<br>
              o2cb: o2dlm has evicted node 4 from domain
              ECB9442E19A94EAC896641BFADD55E4B<br>
              o2cb: o2dlm has evicted node 4 from domain
              ECB9442E19A94EAC896641BFADD55E4B<br>
              o2dlm: Begin recovery on domain
              ECB9442E19A94EAC896641BFADD55E4B for node 4<br>
              o2dlm: Node 5 (he) is the Recovery Master for the dead
              node 4 in domain<br>
              ECB9442E19A94EAC896641BFADD55E4B<br>
              o2dlm: End recovery on domain
              ECB9442E19A94EAC896641BFADD55E4B<br>
              <br>
              <br>
              Additionaly ~4 times per day I see:<br>
              <br>
              ocfs2_check_dir_for_entry:2119 ERROR: status = -17<br>
              ocfs2_mknod:459 ERROR: status = -17<br>
              ocfs2_create:629 ERROR: status = -17<br>
              <br>
              <br>
              I currently use kernel 3.4.2<br>
              my filesystem has been created with:<br>
              -N 8-b 4096 -C 32768 --fs-features<br>
backup-super,strict-journal-super,sparse,extended-slotmap,inline-data,metaecc,xattr,indexed-dirs,refcount,discontig-bg,unwritten,usrquota,grpquota<br>
              <br>
              Could you tell me what could make my system instable?
              Which feature ?<br>
              <br>
              Thanks for any &nbsp;help<br>
              <br>
              Pawel<br>
              <br>
              <br>
              _______________________________________________<br>
              Ocfs2-users mailing list<br>
              <a moz-do-not-send="true"
                href="mailto:Ocfs2-users@oss.oracle.com">Ocfs2-users@oss.oracle.com</a><br>
              <a moz-do-not-send="true"
                href="https://oss.oracle.com/mailman/listinfo/ocfs2-users"
                target="_blank">https://oss.oracle.com/mailman/listinfo/ocfs2-users</a><br>
            </blockquote>
          </div>
          <br>
        </div>
      </blockquote>
      <br>
      <br>
      <fieldset class="mimeAttachmentHeader"></fieldset>
      <br>
      <pre wrap="">_______________________________________________
Ocfs2-users mailing list
<a class="moz-txt-link-abbreviated" href="mailto:Ocfs2-users@oss.oracle.com">Ocfs2-users@oss.oracle.com</a>
<a class="moz-txt-link-freetext" href="https://oss.oracle.com/mailman/listinfo/ocfs2-users">https://oss.oracle.com/mailman/listinfo/ocfs2-users</a></pre>
    </blockquote>
  </body>
</html>