[Gluster-users] uninterruptible processes writing to glusterfsshare

Markus Fröhlich markus.froehlich at xidras.com
Tue Jun 7 13:52:58 UTC 2011


hi!

there ist no relavant output from dmesg.
no entries in the server log - only the one line in the client-server 
log, I already posted.

the glusterfs version on the server had been updated to gfs 3.2.0 more 
than a month ago.
because of the troubles on the backup server, I deleted the whole backup 
share and started from scratch.


I looked for a update of "fuse" and upgraded from 2.7.2-61.18.1 to 
2.8.5-41.1
maybe this helps.

here is the changelog info:

Authors:
--------
     Miklos Szeredi <miklos at szeredi.hu>
Distribution: systemsmanagement:baracus / SLE_11_SP1
* Tue Mar 29 2011 dbahi at novell.com
- remove the --no-canonicalize usage for suse_version <= 11.3

* Mon Mar 21 2011 coolo at novell.com
- licenses package is about to die

* Thu Feb 17 2011 mszeredi at suse.cz
- In case of failure to add to /etc/mtab don't umount. [bnc#668820]
   [CVE-2011-0541]

* Tue Nov 16 2010 mszeredi at suse.cz
- Fix symlink attack for mount and umount [bnc#651598]

* Wed Oct 27 2010 mszeredi at suse.cz
- Remove /etc/init.d/boot.fuse [bnc#648843]

* Tue Sep 28 2010 mszeredi at suse.cz
- update to 2.8.5
   * fix option escaping for fusermount [bnc#641480]

* Wed Apr 28 2010 mszeredi at suse.cz
- keep examples and internal docs in devel package (from jnweiger)

* Mon Apr 26 2010 mszeredi at suse.cz
- update to 2.8.4
   * fix checking for symlinks in umount from /tmp
   * fix umounting if /tmp is a symlink


kind regards
markus froehlich

Am 06.06.2011 21:19, schrieb Anthony J. Biacco:
> Could be fuse, check 'dmesg' for kernel module timeouts.
>
> In a similar vein, has anyone seen signifigant performance/reliability with diff fuse versions? say, latest source vs. Rhel distro rpms vers.
>
> -Tony
>
>
>
> -----Original Message-----
> From: Mohit Anchlia<mohitanchlia at gmail.com>
> Sent: June 06, 2011 1:14 PM
> To: Markus Fröhlich<markus.froehlich at xidras.com>
> Cc: gluster-users at gluster.org<gluster-users at gluster.org>
> Subject: Re: [Gluster-users] uninterruptible processes writing to glusterfsshare
>
> Is there anything in the server logs? Does it follow any particular
> pattern before going in this mode?
>
> Did you upgrade Gluster or is this new install?
>
> 2011/6/6 Markus Fröhlich<markus.froehlich at xidras.com>:
>> hi!
>>
>> sometimes we've on some client-servers hanging uninterruptible processes
>> ("ps aux" stat is on "D" ) and on one the CPU wait I/O grows within some
>> minutes to 100%.
>> you are not able to kill such processes - also "kill -9" doesnt work - when
>> you connect via "strace" to such an process, you wont see anything and you
>> cannot detach it again.
>>
>> there are only two possibilities:
>> killing the glusterfs process (umount GFS share) or rebooting the server.
>>
>> the only log entry I found, was on one client - just a single line:
>> [2011-06-06 10:44:18.593211] I [afr-common.c:581:afr_lookup_collect_xattr]
>> 0-office-data-replicate-0: data self-heal is pending for
>> /pc-partnerbet-public/Promotionaktionen/Mailakquise_2009/Webmaster_2010/HTML/bilder/Thumbs.db.
>>
>> one of the client-servers is a samba-server, the other one a backup-server
>> based on rsync with millions of small files.
>>
>> gfs-servers + gfs-clients: SLES11 x86_64, glusterfs V 3.2.0
>>
>> and here are the configs from server and client:
>> server config
>> "/etc/glusterd/vols/office-data/office-data.gfs-01-01.GFS-office-data02.vol":
>> volume office-data-posix
>>     type storage/posix
>>     option directory /GFS/office-data02
>> end-volume
>>
>> volume office-data-access-control
>>     type features/access-control
>>     subvolumes office-data-posix
>> end-volume
>>
>> volume office-data-locks
>>     type features/locks
>>     subvolumes office-data-access-control
>> end-volume
>>
>> volume office-data-io-threads
>>     type performance/io-threads
>>     subvolumes office-data-locks
>> end-volume
>>
>> volume office-data-marker
>>     type features/marker
>>     option volume-uuid 3c6e633d-a0bb-4c52-8f05-a2db9bc9c659
>>     option timestamp-file /etc/glusterd/vols/office-data/marker.tstamp
>>     option xtime off
>>     option quota off
>>     subvolumes office-data-io-threads
>> end-volume
>>
>> volume /GFS/office-data02
>>     type debug/io-stats
>>     option latency-measurement off
>>     option count-fop-hits off
>>     subvolumes office-data-marker
>> end-volume
>>
>> volume office-data-server
>>     type protocol/server
>>     option transport-type tcp
>>     option auth.addr./GFS/office-data02.allow *
>>     subvolumes /GFS/office-data02
>> end-volume
>>
>>
>> --------------
>> client config "/etc/glusterd/vols/office-data/office-data-fuse.vol":
>> volume office-data-client-0
>>     type protocol/client
>>     option remote-host gfs-01-01
>>     option remote-subvolume /GFS/office-data02
>>     option transport-type tcp
>> end-volume
>>
>> volume office-data-replicate-0
>>     type cluster/replicate
>>     subvolumes office-data-client-0
>> end-volume
>>
>> volume office-data-write-behind
>>     type performance/write-behind
>>     subvolumes office-data-replicate-0
>> end-volume
>>
>> volume office-data-read-ahead
>>     type performance/read-ahead
>>     subvolumes office-data-write-behind
>> end-volume
>>
>> volume office-data-io-cache
>>     type performance/io-cache
>>     subvolumes office-data-read-ahead
>> end-volume
>>
>> volume office-data-quick-read
>>     type performance/quick-read
>>     subvolumes office-data-io-cache
>> end-volume
>>
>> volume office-data-stat-prefetch
>>     type performance/stat-prefetch
>>     subvolumes office-data-quick-read
>> end-volume
>>
>> volume office-data
>>     type debug/io-stats
>>     option latency-measurement off
>>     option count-fop-hits off
>>     subvolumes office-data-stat-prefetch
>> end-volume
>>
>>
>>   -- Mit freundlichen Grüssen
>>
>> Markus Fröhlich
>> Techniker
>>
>> _______________________________________________
>> Gluster-users mailing list
>> Gluster-users at gluster.org
>> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>>
> _______________________________________________
> Gluster-users mailing list
> Gluster-users at gluster.org
> http://gluster.org/cgi-bin/mailman/listinfo/gluster-users
>





More information about the Gluster-users mailing list