<div dir="ltr"><div class="gmail_extra"><div><div class="gmail_signature" data-smartmail="gmail_signature"><div dir="ltr"><br></div></div></div><div class="gmail_quote">On 15 June 2016 at 08:55, Arif Ali <span dir="ltr"><<a href="mailto:mail@arif-ali.co.uk" target="_blank">mail@arif-ali.co.uk</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left-width:1px;border-left-color:rgb(204,204,204);border-left-style:solid;padding-left:1ex"><div dir="ltr"><br><div class="gmail_extra"><div class="gmail_quote"><div><div class="h5">On 15 June 2016 at 08:09, Atin Mukherjee <span dir="ltr"><<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0px 0px 0px 0.8ex;border-left-width:1px;border-left-color:rgb(204,204,204);border-left-style:solid;padding-left:1ex"><span><br>
<br>
On 06/15/2016 12:14 PM, Arif Ali wrote:<br>
><br>
> On 15 June 2016 at 06:48, Atin Mukherjee <<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a><br>
</span><span>> <mailto:<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a>>> wrote:<br>
><br>
><br>
><br>
> On 06/15/2016 11:06 AM, Gandalf Corvotempesta wrote:<br>
> > Il 15 giu 2016 07:09, "Atin Mukherjee" <<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a> <mailto:<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a>><br>
</span>> > <mailto:<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a> <mailto:<a href="mailto:amukherj@redhat.com" target="_blank">amukherj@redhat.com</a>>>> ha scritto:<br>
<span>> >> To get rid of this situation you'd need to stop all the running glusterd<br>
> >> instances and go into /var/lib/glusterd/peers folder on all the nodes<br>
> >> and manually correct the UUID file names and their content if required.<br>
> ><br>
> > If i understood properly the only way to fix this is by bringing the<br>
> > whole cluster down? "you'd need to stop all the running glusterd instances"<br>
> ><br>
> > I hope you are referring to all instances on the failed node...<br>
><br>
> No, since the configuration are synced across all the nodes, any<br>
> incorrect data gets replicated through out. So in this case to be on the<br>
> safer side and validate the correctness all glusterd instances on *all*<br>
> the nodes should be brought down. Having said that, this doesn't impact<br>
> I/O as the management path is different than I/O.<br>
><br>
><br>
> As a sanity, one of the things I did last night, was to reboot the whole<br>
> gluster system, when I had downtime arranged. I thought this is<br>
> something would be asked, as I had seen similar requests on the mailing<br>
> list previously<br>
><br>
> Unfortunately though, it didn't fix the problem.<br>
<br>
</span>Only reboot is not going to solve the problem. You'd need to correct the<br>
configuration as I explained earlier in this thread. If it doesn't<br>
please send the me the content of /var/lib/glusterd/peers/ &<br>
/var/lib/glusterd/<a href="http://glusterd.info" rel="noreferrer" target="_blank">glusterd.info</a> file from all the nodes where glusterd<br>
instances are running. I'll take a look and correct them and send it<br>
back to you.<br></blockquote><div><br></div></div></div><div>Thanks Atin,</div><div><br></div><div>Apologies, I missed your mail, as I was travelling</div><div><br></div><div>I have checked the relevant files you have mentioned, and they seem to look correct to me, but I have attached it for sanity, maybe you can spot something, that I have not seen</div></div></div></div>
</blockquote></div><br></div><div class="gmail_extra">I have been discussing the issue with Atin on IRC, and we have resolved the problem. Thanks Atin, it was much appreciated<br>
<br>
For the purpose of this list. I had the UUID file matching the host in
/var/lib/glusterd/peers for the host itself. This was not required. Once
I removed the UUID based on the node where glusterd was running, the
node was able function correctly<br></div></div>