gluster-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Gluster-devel] GlusterFS hangs/fails: Transport endpoint is not con


From: Fred Hucht
Subject: Re: [Gluster-devel] GlusterFS hangs/fails: Transport endpoint is not connected
Date: Tue, 25 Nov 2008 17:40:15 +0100


Hi,

I installed and ran mcelog and only found DIMM problems on one node which is not related to our problems.

I don't think that the problems are related to only a few nodes. The problems occur on the nodes where the test job runs, and the queue scheduler always selects different nodes.

Do you think kernel 2.5.25.16 is unstable with respect to GlusterFS?

I am quite sure that it is no network issue.

We use local XFS FS on all nodes and a unify with NUMA as described in my first mail. The client config for one node is

volume scns
  type protocol/client
  option transport-type tcp/client
  option remote-host 127.0.0.1
  option remote-subvolume scns
end-volume

volume sc0
  type protocol/client
  option transport-type tcp/client
  option remote-host 127.0.0.1
  option remote-subvolume sc0
end-volume

[same for sc1 - sc87 ]

volume scratch
  type cluster/unify
subvolumes sc0 sc1 sc2 sc3 sc4 sc5 sc6 sc7 sc8 sc9 sc10 sc11 sc12 sc13 sc14 sc15 sc16 sc17 sc18 sc19 sc20 sc21 sc22 sc23 sc24 sc25 sc26 sc27 sc28 sc29 sc30 sc31 sc32 sc33 sc34 sc35 sc36 sc37 sc38 sc39 sc40 sc41 sc42 sc43 sc44 sc45 sc46 sc47 sc48 sc49 sc50 sc51 sc52 sc53 sc54 sc55 sc56 sc57 sc58 sc59 sc60 sc61 sc62 sc63 sc64 sc65 sc67 sc68 sc69 sc70 sc71 sc72 sc73 sc74 sc75 sc76 sc77 sc78 sc79 sc80 sc81 sc82 sc83 sc84 sc85 sc86 sc87
  option namespace scns
  option scheduler nufa
  option nufa.limits.min-free-disk 15
  option nufa.refresh-interval 10
  option nufa.local-volume-name sc0
end-volume

volume scratch-io-threads
  type performance/io-threads
  option thread-count 4
  subvolumes scratch
end-volume

volume scratch-write-behind
  type performance/write-behind
  option aggregate-size 128kB
  option flush-behind off
  subvolumes scratch-io-threads
end-volume

volume scratch-read-ahead
  type performance/read-ahead
  option page-size 128kB # unit in bytes
  option page-count 2    # cache per file  = (page-count x page-size)
  subvolumes scratch-write-behind
end-volume

volume scratch-io-cache
  type performance/io-cache
  option cache-size 64MB
  option page-size 512kB
  subvolumes scratch-read-ahead
end-volume

Regards,

     Fred

On 25.11.2008, at 15:27, Joe Landman wrote:
Fred Hucht wrote:
Hi,
crawling through all /var/log/messages, I found on one of the failing nodes (node68)

Does your setup use local disk? Is it possible that the backing store is failing?

If you run

        mcelog > /tmp/mce.log 2>&1

on the failing node, do you get any output in /tmp/mce.log ?

My current thoughts in no particular order are

hardware based: failures always concentrated on a few specific nodes (always repeatable only on those nodes)

a) failing local hard drive: backing store failing *could* impact the file system, and you would see this as NFS working on a remote FS while failing on an FS in part storing locally.

b) network issue: possibly a bad driver/flaky port/overloaded switch backplane. This is IMO less likely, as NFS works. Could you post output of "ifconfig" so we can look for error indicators in the port state?

Software based:

c) fuse bugs: I have run into a few in the past, and they have caused errors like this. But umount/mount rarely fixes a hung fuse process, so this is, again, IMO, less likely.

d) GlusterFS bugs: I think the devels would recognize it if it were one. I doubt this at this moment.

e) kernel bug: We are using 2.6.27.5 right now, about to update to . 7 due to some Cert advisories. We have had (stability) issues with kernels from 2.6.24 to 2.6.26.x (x low numbers) under intense loads. It wouldn't surprise me if what you are observing is actually just a symptom of a real problem somewhere else in the kernel. That the state gets resolved when you umount/mount suggests that this could be the case.

Joe




--
Joseph Landman, Ph.D
Founder and CEO
Scalable Informatics LLC,
email: address@hidden
web  : http://www.scalableinformatics.com
      http://jackrabbit.scalableinformatics.com
phone: +1 734 786 8423 x121
fax  : +1 866 888 3112
cell : +1 734 612 4615

Dr. Fred Hucht <address@hidden>
Institute for Theoretical Physics
University of Duisburg-Essen, 47048 Duisburg, Germany





reply via email to

[Prev in Thread] Current Thread [Next in Thread]