gluster-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: {Disarmed} Re: [Gluster-devel] GlusterFS drops mount point


From: Marcus Herou
Subject: Re: {Disarmed} Re: [Gluster-devel] GlusterFS drops mount point
Date: Sat, 13 Sep 2008 19:30:31 +0200

Thanks guys for answering.

GlusterFS-1.3.10 is used on both client and server. Could not get the
patched fuse to compile in Hardy (error below) so until that works I just
took a binary .deb package even though 1.3.12 was available.

FUSE error:
/root/dload/fuse-2.7.2glfs9/kernel/dir.c:1027: error: 'struct iattr' has no
member named 'ia_file'
Linux mapreduce8 2.6.24-19-server #1 SMP Fri Jul 11 21:50:43 UTC 2008 x86_64
GNU/Linux
libfuse2        2.7.2-1ubuntu2

On the client which drops the mount point:
Linux tailsweep2 2.6.17-12-server #2 SMP Thu Jan 31 22:15:27 UTC 2008 i686
GNU/Linux
libfuse2        2.7.0-1ubuntu5
This one compiles the patched fuse 2.7.0...


What we are trying to achieve with the config is something similar to two
replicas of each file spread with afr on three nodes and aggregated with
unify.

Config files:

Server:
volume home
  type storage/posix
  option directory /srv/export/home
end-volume

volume server
  type protocol/server
  subvolumes home
  option transport-type tcp/server # For TCP/IP transport
  option auth.ip.home.allow *
end-volume

Client:
volume v1
  type protocol/client
  option transport-type tcp/client
  option remote-host 192.168.10.30
  option remote-subvolume home
end-volume

volume v2
  type protocol/client
  option transport-type tcp/client
  option remote-host 192.168.10.31
  option remote-subvolume home
end-volume

volume v3
  type protocol/client
  option transport-type tcp/client
  option remote-host 192.168.10.32
  option remote-subvolume home
end-volume

volume afr-1
  type cluster/afr
  subvolumes v1 v2
end-volume

volume afr-2
  type cluster/afr
  subvolumes v2 v3
end-volume

volume afr-3
  type cluster/afr
  subvolumes v3 v1
end-volume

volume ns1
  type protocol/client
  option transport-type tcp/client
  option remote-host 192.168.10.30
  option remote-subvolume home-namespace
end-volume

volume ns2
  type protocol/client
  option transport-type tcp/client
  option remote-host 192.168.10.31
  option remote-subvolume home-namespace
end-volume

volume ns3
  type protocol/client
  option transport-type tcp/client
  option remote-host 192.168.10.32
  option remote-subvolume home-namespace
end-volume

volume namespace
  type cluster/afr
  subvolumes ns1 ns2 ns3
end-volume

volume v
  type cluster/unify
  option scheduler rr
  option namespace namespace
  subvolumes afr-1 afr-2 afr-3
end-volume

I really hope we have misconfigured something since that is the easiest fix
:)

Kindly

//Marcus



On Sat, Sep 13, 2008 at 12:50 AM, Amar S. Tumballi <address@hidden>wrote:

> Also which version of GlusterFS?
>
> address@hidden>
>
>> may be configuration issue...  lets start with config, what does you
>> config look like on client and server?
>>
>> Marcus Herou wrote:
>>
>>> Lots of these on server
>>> 2008-09-12 20:48:14 E [protocol.c:271:gf_block_unserialize_transport]
>>> server: EOF from peer (*MailScanner has detected a possible fraud attempt
>>> from "192.168.10.4:1007" claiming to be* *MailScanner warning: numerical
>>> links are often malicious:* 192.168.10.4:1007 <http://192.168.10.4:1007
>>> >)
>>> ...
>>> 2008-09-12 20:50:12 E [server-protocol.c:4153:server_closedir] server:
>>> not getting enough data, returning EINVAL
>>> ...
>>> 2008-09-12 20:50:12 E [server-protocol.c:4148:server_closedir] server:
>>> unresolved fd 6
>>> ...
>>> 2008-09-12 20:51:47 E [protocol.c:271:gf_block_unserialize_transport]
>>> server: EOF from peer (*MailScanner has detected a possible fraud attempt
>>> from "192.168.10.10:1015" claiming to be* *MailScanner warning:
>>> numerical links are often malicious:* 192.168.10.10:1015 <
>>> http://192.168.10.10:1015>)
>>>
>>> ...
>>>
>>> And lots of these on client
>>>
>>> 2008-09-12 19:54:45 E [afr.c:2201:afr_open] home-namespace: self heal
>>> failed, returning EIO
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3954: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3956: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3958: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3987: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3989: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3991: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:45 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3993: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 C [client-protocol.c:212:call_bail] home3: bailing
>>> transport
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3970: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3971: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3972: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 3974: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 4001: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 4002: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:54:54 E [client-protocol.c:4827:client_protocol_cleanup]
>>> home3: forced unwinding frame type(2) op(5) address@hidden
>>> 2008-09-12 19:54:54 E [client-protocol.c:4239:client_lock_cbk] home3: no
>>> proper reply from server, returning ENOTCONN
>>> 2008-09-12 19:54:54 E [afr.c:1933:afr_selfheal_lock_cbk] home-afr-3:
>>> (path=/rsyncer/.ssh/authorized_keys2 child=home3) op_ret=-1 op_errno=107
>>> 2008-09-12 19:54:54 E [afr.c:2201:afr_open] home-afr-3: self heal failed,
>>> returning EIO
>>> 2008-09-12 19:54:54 E [fuse-bridge.c:715:fuse_fd_cbk] glusterfs-fuse:
>>> 4004: (12) /rsyncer/.ssh/authorized_keys2 => -1 (5)
>>> 2008-09-12 19:55:01 E [unify.c:335:unify_lookup] home: returning ESTALE
>>> for /rsyncer/.ssh/authorized_keys2: file count is 4
>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home:
>>> /rsyncer/.ssh/authorized_keys2: found on home-namespace
>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home:
>>> /rsyncer/.ssh/authorized_keys2: found on home-afr-2
>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home:
>>> /rsyncer/.ssh/authorized_keys2: found on home-afr-1
>>> 2008-09-12 19:55:01 E [unify.c:339:unify_lookup] home:
>>> /rsyncer/.ssh/authorized_keys2: found on home-afr-3
>>>
>>>
>>> Both server and client are spitting out tons of these. Thought "E" was
>>> Error level, seems like DEBUG ?
>>>
>>> Kindly
>>>
>>> //Marcus
>>>
>>>
>>>
>>>
>>> On Fri, Sep 12, 2008 at 8:01 PM, Brian Taber <address@hidden<mailto:
>>> address@hidden>> wrote:
>>>
>>>    What do you see in your server and client logs for gluster?
>>>
>>>    -------------------------
>>>    Brian Taber
>>>    Owner/IT Specialist
>>>    Diverse Computer Group
>>>    Office: 774-206-5592
>>>    Cell: 508-496-9221
>>>    address@hidden <mailto:address@hidden>
>>>
>>>
>>>
>>>
>>>    Marcus Herou wrote:
>>>    > Hi.
>>>    >
>>>    > We have just recently installed a 3 node cluster with 16 SATA
>>>    disks each.
>>>    >
>>>    > We are using Hardy and the glusterfs-3.10 Ubuntu package on both
>>>    client(s)
>>>    > and server.
>>>    >
>>>    > We have only created one export (/home) yet since we want to
>>>    test it a while
>>>    > before putting it into a live high performance environment.
>>>    >
>>>    > The problem is currently that the client looses /home once a day
>>>    or so. This
>>>    > is really bad since this is a machine which all other connect to
>>>    with ssh
>>>    > keys thus making them unable to log in.
>>>    >
>>>    > Anyone seen something similar ?
>>>    >
>>>    > Kindly
>>>    >
>>>    > //Marcus
>>>    > _______________________________________________
>>>    > Gluster-devel mailing list
>>>    > address@hidden <mailto:address@hidden>
>>>    > http://lists.nongnu.org/mailman/listinfo/gluster-devel
>>>    >
>>>
>>>
>>>
>>>
>>> --
>>> Marcus Herou CTO and co-founder Tailsweep AB
>>> +46702561312
>>> address@hidden <mailto:address@hidden>
>>> http://www.tailsweep.com/
>>> http://blogg.tailsweep.com/
>>>
>> _______________________________________________
>> Gluster-devel mailing list
>> address@hidden
>> http://lists.nongnu.org/mailman/listinfo/gluster-devel
>>
>
>
>
> --
> Amar Tumballi
> Gluster/GlusterFS Hacker
> [bulde on #gluster/irc.gnu.org]
> http://www.zresearch.com - Commoditizing Super Storage!
>



-- 
Marcus Herou CTO and co-founder Tailsweep AB
+46702561312
address@hidden
http://www.tailsweep.com/
http://blogg.tailsweep.com/


reply via email to

[Prev in Thread] Current Thread [Next in Thread]