Visit Open-E website
Results 1 to 3 of 3

Thread: use the nfs mount mix

  1. #1

    Angry use the nfs mount mix

    *1. NFS mount issue*

    Hostname : *dss03*
    Model : *SR2612UR*
    Mount clients : *Solaris, Linux*
    Share Volume : *lv0001/vol1, lv0002/vol2, lv0003/vol3*
    Working Job : *upd_0882-DSS-V6.upd patch applied*

    *Working history*
    1-a. NFS clients not unmount
    1-b. Service network cable unplug
    1-c. Admin network 100.100.101.223 using
    1-d. Patch applied
    1-e. dss03 NFS service check (Online)
    1-f. Service network cable connect

    *NFS client problem*
    -------------------------------
    for Solaris:

    *Before patch applied (Normal)*
    vector:/vol2 4194172928 10705788 4183467140 1% /sleds2/vol21
    vector:/vol1 6291324928 677188620 5614136308 11% /vector
    vector:/vol3 3145596928 997564684 2148032244 32% /edsdata

    *After patch applied (Problem)*
    vector:/vol2 6291324928 677188620 5614136308 11%
    /sleds2/vol21
    vector:/vol1 104806400 30124 104776276 1% /vector
    vector:/vol3 4194172928 10705788 4183467140 1% /edsdata

    *mount after umount (Normal)*
    vector:/vol2 4194172928 10705788 4183467140 1% /sleds2/vol21
    vector:/vol1 6291324928 677188620 5614136308 11% /vector
    vector:/vol3 3145596928 997564684 2148032244 32% /edsdata

    * mount option (# mount -F nfs -o rw,soft vector:/vol*[x]* /*[mountpoint]*
    )
    -------------------------------
    for Linux:

    NFS mount is gone after patch applied
    Linux system rebooting or mount command use

    * mount option (# mount -t nfs -o rw,soft vector:/vol*[x]* /*[mountpoint]*
    )
    -------------------------------


    *For Solaris, the size of share has changed.*
    *For Linux, the mount point is gone.*
    *Why is this phenomenon?*


    *2. Data replication issue*

    Hostname : *dss01, dss02*
    Model : *SR2612UR, SR1600UR*
    Problem Logical Volume : *lv0001 (3500 GB)*
    Problem Share Name : *engr*
    Problem Data Replication : *lv01_engr*

    *Problem Description*
    a. dss01 and dss02 is service and backup system
    b. File view: task_reverse.log
    ------------------------------------------------------
    [1313200851.564387095-7d21f2caf]
    summary=;;;;
    status[0]=12
    action=stop
    time=2011-08-13 04:00:51
    type=DREP
    name=lv01_engr
    ------------------------------------------------------
    c. error log
    ------------------------------------------------------
    Error 12 : Error in rsync protocol data stream
    ------------------------------------------------------
    d. Check system (lv0001(dss01) 95% used, lv0001(dss02) data full)
    e. "lv01_engr" data replication running fail

    *The original data is small. Why should not data replication?*

    * Current lv0001(dss02) is data replication success after 650GB resizing


    *3. dss01 system error messages*

    2011/08/17 08:27:25|Pid: 29065, comm: nfsd Not tainted
    2.6.27.39-oe64-00000-gc758227 #39
    2011/08/17 08:27:25|
    2011/08/17 08:27:25|Call Trace:
    2011/08/17 08:27:25|[<ffffffff80235004>] warn_on_slowpath+0x64/0xa0
    2011/08/17 08:27:25|[<ffffffff8024df06>] up+0x16/0x50
    2011/08/17 08:27:25|[<ffffffff80235a35>] release_console_sem+0x1b5/0x1f0
    2011/08/17 08:27:25|[<ffffffff805d9908>] lock_sock_nested+0xc8/0xe0
    2011/08/17 08:27:25|[<ffffffff805d9908>] lock_sock_nested+0xc8/0xe0
    2011/08/17 08:27:25|[<ffffffff806955c5>] schedule_timeout+0x95/0xd0
    2011/08/17 08:27:25|[<ffffffff8062e1c2>] inet_sock_destruct+0x172/0x1c0
    2011/08/17 08:27:25|[<ffffffff805d9f83>] sk_free+0x23/0x110
    2011/08/17 08:27:25|[<ffffffff8062d9be>] inet_release+0x3e/0x70
    2011/08/17 08:27:25|[<ffffffff805d7131>] sock_release+0x21/0x90
    2011/08/17 08:27:25|[<ffffffff805d771e>] sock_close+0x1e/0x40
    2011/08/17 08:27:25|[<ffffffff80296eb1>] __fput+0xb1/0x1d0
    2011/08/17 08:27:25|[<ffffffff80664ae2>] svc_sock_free+0x22/0x60
    2011/08/17 08:27:25|[<ffffffff8066f228>] svc_xprt_free+0x38/0x50
    2011/08/17 08:27:25|[<ffffffff8066f1f0>] svc_xprt_free+0x0/0x50
    2011/08/17 08:27:25|[<ffffffff8041f443>] kref_put+0x33/0x70
    2011/08/17 08:27:25|[<ffffffff8066fa57>] svc_close_all+0x57/0x70
    2011/08/17 08:27:25|[<ffffffff806638be>] svc_destroy+0x5e/0x150
    2011/08/17 08:27:25|[<ffffffff8034ac0e>] nfsd+0x26e/0x2a0
    2011/08/17 08:27:25|[<ffffffff8034a9a0>] nfsd+0x0/0x2a0
    2011/08/17 08:27:25|[<ffffffff802499ab>] kthread+0x4b/0x80
    2011/08/17 08:27:25|[<ffffffff8020d149>] child_rip+0xa/0x11
    2011/08/17 08:27:25|[<ffffffff80249960>] kthread+0x0/0x80
    2011/08/17 08:27:25|[<ffffffff8020d13f>] child_rip+0x0/0x11
    2011/08/17 08:27:25|

    *How fix the problem?*

  2. #2

    Default

    Try setting the Solaris and Linux to use NFS ver. 3 only.

    For Solaris:
    following switches in /etc/vfstab:
    proto=tcp,vers=3

    For Linux:
    try mounting with vers=3

    For the Error 12 see link below:

    http://kb.open-e.com/What-does-Error...ation_982.html
    All the best,

    Todd Maxwell


    Follow the red "E"
    Facebook | Twitter | YouTube

  3. #3
    Join Date
    Oct 2010
    Location
    GA
    Posts
    935

    Default

    2011/08/17 08:27:25|Call Trace:
    2011/08/17 08:27:25|[<ffffffff80235004>] warn_on_slowpath+0x64/0xa0
    2011/08/17 08:27:25|[<ffffffff8024df06>] up+0x16/0x50
    2011/08/17 08:27:25|[<ffffffff80235a35>] release_console_sem+0x1b5/0x1f0
    2011/08/17 08:27:25|[<ffffffff805d9908>] lock_sock_nested+0xc8/0xe0
    2011/08/17 08:27:25|[<ffffffff805d9908>] lock_sock_nested+0xc8/0xe0
    2011/08/17 08:27:25|[<ffffffff806955c5>] schedule_timeout+0x95/0xd0
    2011/08/17 08:27:25|[<ffffffff8062e1c2>] inet_sock_destruct+0x172/0x1c0
    2011/08/17 08:27:25|[<ffffffff805d9f83>] sk_free+0x23/0x110
    2011/08/17 08:27:25|[<ffffffff8062d9be>] inet_release+0x3e/0x70
    2011/08/17 08:27:25|[<ffffffff805d7131>] sock_release+0x21/0x90
    2011/08/17 08:27:25|[<ffffffff805d771e>] sock_close+0x1e/0x40
    2011/08/17 08:27:25|[<ffffffff80296eb1>] __fput+0xb1/0x1d0
    2011/08/17 08:27:25|[<ffffffff80664ae2>] svc_sock_free+0x22/0x60
    2011/08/17 08:27:25|[<ffffffff8066f228>] svc_xprt_free+0x38/0x50
    2011/08/17 08:27:25|[<ffffffff8066f1f0>] svc_xprt_free+0x0/0x50
    2011/08/17 08:27:25|[<ffffffff8041f443>] kref_put+0x33/0x70
    2011/08/17 08:27:25|[<ffffffff8066fa57>] svc_close_all+0x57/0x70
    2011/08/17 08:27:25|[<ffffffff806638be>] svc_destroy+0x5e/0x150
    2011/08/17 08:27:25|[<ffffffff8034ac0e>] nfsd+0x26e/0x2a0
    2011/08/17 08:27:25|[<ffffffff8034a9a0>] nfsd+0x0/0x2a0
    2011/08/17 08:27:25|[<ffffffff802499ab>] kthread+0x4b/0x80
    2011/08/17 08:27:25|[<ffffffff8020d149>] child_rip+0xa/0x11
    2011/08/17 08:27:25|[<ffffffff80249960>] kthread+0x0/0x80
    2011/08/17 08:27:25|[<ffffffff8020d13f>] child_rip+0x0/0x11
    This may indicate a stale NFS mount.
    To correct it, the hosts need to re-mount the NFS share after storage reboots.
    I think this is a known issue with NFS.

Posting Permissions

  • You may not post new threads
  • You may not post replies
  • You may not post attachments
  • You may not edit your posts
  •