<div dir="ltr"><div>I haven&#39;t done anything more than peer thus far, so I&#39;m a bit confused as to how the volume info fits in, can you expand on this a bit?<br></div><div><br></div><div>Failed commits? Is this split brain on the replica volumes? I don&#39;t get any return from &#39;gluster volume heal &lt;volname&gt; info&#39; on all the replica volumes, but if I try a gluster volume heal &lt;volname&gt; full I get: &#39;Launching heal operation to perform full self heal on volume &lt;volname&gt; has been unsuccessful&#39;.</div><div><br></div><div>I have 5 volumes total.</div><div><br></div><div>&#39;Replica 3&#39; volumes running on gluster01/02/03:</div><div>vm-storage</div><div>iso-storage</div><div>export-domain-storage</div><div>env-modules</div><div><br></div><div>And one distributed only volume &#39;storage&#39; info shown below:<br></div><div><div><br></div><div><b>From existing host gluster01/02:</b></div><div><div>type=0</div><div>count=4</div><div>status=1</div><div>sub_count=0</div><div>stripe_count=1</div><div>replica_count=1</div><div>disperse_count=0</div><div>redundancy_count=0</div><div>version=25</div><div>transport-type=0</div><div>volume-id=26d355cb-c486-481f-ac16-e25390e73775</div><div>username=eb9e2063-6ba8-4d16-a54f-2c7cf7740c4c</div><div>password=</div><div>op-version=3</div><div>client-op-version=3</div><div>quota-version=1</div><div>parent_volname=N/A</div><div>restored_from_snap=00000000-0000-0000-0000-000000000000</div><div>snap-max-hard-limit=256</div><div>features.quota-deem-statfs=on</div><div>features.inode-quota=on</div><div>diagnostics.brick-log-level=WARNING</div><div>features.quota=on</div><div>performance.readdir-ahead=on</div><div>performance.cache-size=1GB</div><div>performance.stat-prefetch=on</div><div>brick-0=10.0.231.50:-mnt-raid6-storage-storage</div><div>brick-1=10.0.231.51:-mnt-raid6-storage-storage</div><div>brick-2=10.0.231.52:-mnt-raid6-storage-storage</div><div>brick-3=10.0.231.53:-mnt-raid6-storage-storage</div></div><div><br></div><div><div><b>From existing host gluster03/04:</b><br></div><div><div>type=0</div><div>count=4</div><div>status=1</div><div>sub_count=0</div><div>stripe_count=1</div><div>replica_count=1</div><div>disperse_count=0</div><div>redundancy_count=0</div><div>version=25</div><div>transport-type=0</div><div>volume-id=26d355cb-c486-481f-ac16-e25390e73775</div><div>username=eb9e2063-6ba8-4d16-a54f-2c7cf7740c4c</div><div>password=</div><div>op-version=3</div><div>client-op-version=3</div><div>quota-version=1</div><div>parent_volname=N/A</div><div>restored_from_snap=00000000-0000-0000-0000-000000000000</div><div>snap-max-hard-limit=256</div><div>features.quota-deem-statfs=on</div><div>features.inode-quota=on</div><div>performance.stat-prefetch=on</div><div>performance.cache-size=1GB</div><div>performance.readdir-ahead=on</div><div>features.quota=on</div><div>diagnostics.brick-log-level=WARNING</div><div>brick-0=10.0.231.50:-mnt-raid6-storage-storage</div><div>brick-1=10.0.231.51:-mnt-raid6-storage-storage</div><div>brick-2=10.0.231.52:-mnt-raid6-storage-storage</div><div>brick-3=10.0.231.53:-mnt-raid6-storage-storage</div></div><div><br></div><div>So far between gluster01/02 and gluster03/04 the configs are the same, although the ordering is different for some of the features.</div><div><br></div><div>On gluster05/06 the ordering is different again, and the quota-version=0 instead of 1.</div><div><br></div><div><b>From new hosts gluster05/gluster06:</b></div><div>type=0</div><div>count=4</div><div>status=1</div><div>sub_count=0</div><div>stripe_count=1</div><div>replica_count=1</div><div>disperse_count=0</div><div>redundancy_count=0</div><div>version=25</div><div>transport-type=0</div><div>volume-id=26d355cb-c486-481f-ac16-e25390e73775</div><div>username=eb9e2063-6ba8-4d16-a54f-2c7cf7740c4c</div><div>password=</div><div>op-version=3</div><div>client-op-version=3</div><div>quota-version=0</div><div>parent_volname=N/A</div><div>restored_from_snap=00000000-0000-0000-0000-000000000000</div><div>snap-max-hard-limit=256</div><div>performance.stat-prefetch=on</div><div>performance.cache-size=1GB</div><div>performance.readdir-ahead=on</div><div>features.quota=on</div><div>diagnostics.brick-log-level=WARNING</div><div>features.inode-quota=on</div><div>features.quota-deem-statfs=on</div><div>brick-0=10.0.231.50:-mnt-raid6-storage-storage</div><div>brick-1=10.0.231.51:-mnt-raid6-storage-storage</div><div>brick-2=10.0.231.52:-mnt-raid6-storage-storage</div><div>brick-3=10.0.231.53:-mnt-raid6-storage-storage</div></div><div><br></div></div><div>Also, I forgot to mention that when I initially peer&#39;d the two new hosts, glusterd crashed on gluster03 and had to be restarted (log attached) but has been fine since.</div><div><br></div><div>Thanks,</div><div>Steve</div></div><div class="gmail_extra"><br><div class="gmail_quote">On Thu, Feb 25, 2016 at 11:27 AM, Mohammed Rafi K C <span dir="ltr">&lt;<a href="mailto:rkavunga@redhat.com" target="_blank">rkavunga@redhat.com</a>&gt;</span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
  
    
  
  <div bgcolor="#FFFFFF" text="#000000"><span class="">
    <br>
    <br>
    <div>On 02/25/2016 11:45 PM, Steve Dainard
      wrote:<br>
    </div>
    <blockquote type="cite">
      <div dir="ltr">Hello,<br>
        <br>
        I upgraded from 3.6.6 to 3.7.6 a couple weeks ago. I just peered
        2 new nodes to a 4 node cluster and gluster peer status is:<br>
        <br>
        # gluster peer status <b>&lt;-- from node gluster01</b><br>
        Number of Peers: 5<br>
        <br>
        Hostname: 10.0.231.51<br>
        Uuid: b01de59a-4428-486b-af49-cb486ab44a07<br>
        State: Peer in Cluster (Connected)<br>
        <br>
        Hostname: 10.0.231.52<br>
        Uuid: 75143760-52a3-4583-82bb-a9920b283dac<br>
        State: Peer in Cluster (Connected)<br>
        <br>
        Hostname: 10.0.231.53<br>
        Uuid: 2c0b8bb6-825a-4ddd-9958-d8b46e9a2411<br>
        State: Peer in Cluster (Connected)<br>
        <br>
        Hostname: 10.0.231.54 <b>&lt;-- new node gluster05</b><br>
        Uuid: 408d88d6-0448-41e8-94a3-bf9f98255d9c<br>
        <b>State: Peer Rejected (Connected)</b><br>
        <br>
        Hostname: 10.0.231.55 <b>&lt;-- new node gluster06</b><br>
        Uuid: 9c155c8e-2cd1-4cfc-83af-47129b582fd3<br>
        <b>State: Peer Rejected (Connected)</b><br>
      </div>
    </blockquote>
    <br></span>
    Looks like your configuration files are mismatching, ie the checksum
    calculation differs on this two node than the others,<br>
    <br>
    Did you had any failed commit ?<br>
    <br>
    Compare your /var/lib/glusterd/&lt;volname&gt;/info of the failed
    node against good one, mostly you could see some difference.<br>
    <br>
    can you paste the /var/lib/glusterd/&lt;volname&gt;/info ?<br>
    <br>
    Regards<br>
    Rafi KC<br>
    <br>
    <br>
    <blockquote type="cite"><span class="">
      <div dir="ltr">
        <div><b><br>
          </b></div>
        <div>I followed the write-up here: <a href="http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected" target="_blank">http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected</a>
          and the two new nodes peer&#39;d properly but after a reboot of
          the two new nodes I&#39;m seeing the same Peer Rejected
          (Connected) State.</div>
        <div><br>
        </div>
        <div>I&#39;ve attached logs from an existing node, and the two new
          nodes.</div>
        <div><br>
        </div>
        <div>Thanks for any suggestions,</div>
        <div>Steve</div>
        <div><br>
        </div>
        <div>
          <div><br>
          </div>
        </div>
      </div>
      <br>
      <fieldset></fieldset>
      <br>
      </span><pre>_______________________________________________
Gluster-users mailing list
<a href="mailto:Gluster-users@gluster.org" target="_blank">Gluster-users@gluster.org</a>
<a href="http://www.gluster.org/mailman/listinfo/gluster-users" target="_blank">http://www.gluster.org/mailman/listinfo/gluster-users</a></pre>
    </blockquote>
    <br>
  </div>

</blockquote></div><br></div>