<html>
<head>
<meta content="text/html; charset=utf-8" http-equiv="Content-Type">
</head>
<body bgcolor="#FFFFFF" text="#000000">
<br>
On 06/13/2015 04:50 PM, Atin Mukherjee wrote:<br>
<span style="white-space: pre;">> <br>
> Sent from Samsung Galaxy S4 On 13 Jun 2015 14:42, "Anand
Nekkunti"<br>
> <<a class="moz-txt-link-abbreviated" href="mailto:anekkunt@redhat.com">anekkunt@redhat.com</a>
<a class="moz-txt-link-rfc2396E" href="mailto:anekkunt@redhat.com"><mailto:anekkunt@redhat.com></a>> wrote:<br>
>> <br>
>> <br>
>> On 06/13/2015 02:27 PM, Atin Mukherjee wrote:<br>
>>> <br>
>>> Sent from Samsung Galaxy S4 On 13 Jun 2015 13:15,
"Raghavendra<br>
>>> Talur" <<a class="moz-txt-link-abbreviated" href="mailto:raghavendra.talur@gmail.com">raghavendra.talur@gmail.com</a><br>
>>> <a class="moz-txt-link-rfc2396E" href="mailto:raghavendra.talur@gmail.com"><mailto:raghavendra.talur@gmail.com></a>> wrote:<br>
>>>> <br>
>>>> <br>
>>>> <br>
>>>> On Sat, Jun 13, 2015 at 1:00 PM, Atin Mukherjee<br>
>>>> <<a class="moz-txt-link-abbreviated" href="mailto:atin.mukherjee83@gmail.com">atin.mukherjee83@gmail.com</a><br>
>>>> <a class="moz-txt-link-rfc2396E" href="mailto:atin.mukherjee83@gmail.com"><mailto:atin.mukherjee83@gmail.com></a>>
wrote:<br>
>>>>> <br>
>>>>> Sent from Samsung Galaxy S4 On 13 Jun 2015
12:58, "Anand<br>
>>>>> Nekkunti" <<a class="moz-txt-link-abbreviated" href="mailto:anekkunt@redhat.com">anekkunt@redhat.com</a>
<a class="moz-txt-link-rfc2396E" href="mailto:anekkunt@redhat.com"><mailto:anekkunt@redhat.com></a>><br>
>>>>> wrote:<br>
>>>>>> <br>
>>>>>> Hi All Rebalance is not working in single
node cluster<br>
>>>>>> environment ( current test frame work ).
I am getting<br>
>>>>>> error in below test , it seems re-balance
is not migrated<br>
>>>>>> to current cluster test framework.<br>
>>>>> Could you pin point which test case fails and
what log do you<br>
>>>>> see?<br>
>>>>>> <br>
>>>>>> cleanup; TEST launch_cluster 2; TEST
$CLI_1 peer probe<br>
>>>>>> $H2;<br>
>>>>>> <br>
>>>>>> EXPECT_WITHIN $PROBE_TIMEOUT 1
check_peers<br>
>>>>>> <br>
>>>>>> $CLI_1 volume create $V0 $H1:$B1/$V0
$H2:$B2/$V0 EXPECT<br>
>>>>>> 'Created' volinfo_field $V0 'Status';<br>
>>>>>> <br>
>>>>>> $CLI_1 volume start $V0 EXPECT 'Started'
volinfo_field $V0<br>
>>>>>> 'Status';<br>
>>>>>> <br>
>>>>>> #Mount FUSE TEST glusterfs -s $H1
--volfile-id=$V0 $M0;<br>
>>>>>> <br>
>>>>>> TEST mkdir $M0/dir{1..4}; TEST touch<br>
>>>>>> $M0/dir{1..4}/files{1..4};<br>
>>>>>> <br>
>>>>>> TEST $CLI_1 volume add-brick $V0
$H1:$B1/${V0}1<br>
>>>>>> $H2:$B2/${V0}1<br>
>>>>>> <br>
>>>>>> TEST $CLI_1 volume rebalance $V0 start<br>
>>>>>> <br>
>>>>>> EXPECT_WITHIN 60 "completed"
CLI_1_rebalance_status_field<br>
>>>>>> $V0<br>
>>>>>> <br>
>>>>>> $CLI_2 volume status $V0 EXPECT 'Started'
volinfo_field $V0<br>
>>>>>> 'Status';<br>
>>>>>> <br>
>>>>>> cleanup;<br>
>>>>>> <br>
>>>>>> Regards Anand.N<br>
>>>>>> <br>
>>>>>> <br>
>>>>>> <br>
>>>>>>
_______________________________________________ <br>
>>>>>> Gluster-devel mailing list
<a class="moz-txt-link-abbreviated" href="mailto:Gluster-devel@gluster.org">Gluster-devel@gluster.org</a><br>
>>>>>> <a class="moz-txt-link-rfc2396E" href="mailto:Gluster-devel@gluster.org"><mailto:Gluster-devel@gluster.org></a>
<br>
>>>>>>
<a class="moz-txt-link-freetext" href="http://www.gluster.org/mailman/listinfo/gluster-devel">http://www.gluster.org/mailman/listinfo/gluster-devel</a><br>
>>>>>> <br>
>>>>> <br>
>>>>> <br>
>>>>>
_______________________________________________ Gluster-devel<br>
>>>>> mailing list <a class="moz-txt-link-abbreviated" href="mailto:Gluster-devel@gluster.org">Gluster-devel@gluster.org</a><br>
>>>>> <a class="moz-txt-link-rfc2396E" href="mailto:Gluster-devel@gluster.org"><mailto:Gluster-devel@gluster.org></a> <br>
>>>>>
<a class="moz-txt-link-freetext" href="http://www.gluster.org/mailman/listinfo/gluster-devel">http://www.gluster.org/mailman/listinfo/gluster-devel</a><br>
>>>>> <br>
>>>> <br>
>>>> If it is a crash of glusterd when you do
rebalance start, it is<br>
>>>> because of FORTIFY_FAIL in libc. Here is the
patch that Susant<br>
>>>> has already sent:
<a class="moz-txt-link-freetext" href="http://review.gluster.org/#/c/11090/">http://review.gluster.org/#/c/11090/</a><br>
>>>> <br>
>>>> You can verify that it is the same crash by
checking the core<br>
>>>> in gdb; a SIGABRT would be raised after strncpy.<br>
>> <br>
>> <br>
>> glusterd is not crashing, but I am getting rebalance
status as<br>
>> fail in my test case. It is happening in test frame work
( any<br>
>> simulated cluster environment in same node ) only. RCA:
1. we are<br>
>> passing always "localhost" as volfile server for
rebalance xlator<br>
>> . 2.Rebalance processes are overwriting unix socket and
log files<br>
>> each other (All rebalance processes are creating socket
with same<br>
>> name) .<br>
>> <br>
>> I will send patch for this<br>
> I thought we were already in an agreement for this yesterday.
IIRC,<br>
> the same is true for all other daemons. As of now we dont
have any<br>
> tests which invoke daemons using cluster.rc<br>
> </span><br>
ya .. yesterday we found that volfile server is the problem ,
I modified volfile server but still i was getting rebalance status
fail . Initially I thought some problem in rebalance process, later
I found that rebalance not able send respond to to glusterd after
completing rebalance due to unix socket file corruption and all
rebalance daemons are writing log into same log file .<br>
I think there is no issue with other daemons which are are using
SVC framwork work.<br>
<br>
patch: <a class="moz-txt-link-freetext" href="http://review.gluster.org/#/c/11210/">http://review.gluster.org/#/c/11210/</a> - this patch enable the
writing test cases for rebalance in cluster environment. <br>
<br>
<br>
<span style="white-space: pre;">>> <br>
>> Regards Anand.N<br>
>>> <br>
>>>> <br>
>>> AFAIR Anand tried it in mainline and that fix was
already in<br>
>>> place. I think this is something different.<br>
>>>> -- Raghavendra Talur<br>
>>>> <br>
>> <br>
>> <br>
> </span><br>
<br>
<br>
</body>
</html>