HI Patrick,
That worries me. I am taking down the file system for another clean restart. Is there
anything different one should do in this case just to make sure this does not escalate to
worry some repetition?
Thank you for your response.
Amit
-----Original Message-----
From: Patrick Farrell [mailto:paf@cray.com]
Sent: Thursday, July 30, 2015 5:42 PM
To: Kumar, Amit; Mohr Jr, Richard Frank (Rick Mohr)
Cc: hpdd-discuss(a)lists.01.org
Subject: RE: [HPDD-discuss] One of several Clients failing to mount handful
OST's
Amit,
-ENODEV (status=-19) may be from sort of race related to
reconnection/recovery, if I'm reading that LU correctly. If those messages
continue for a long time, it's cause for concern. If they go away within, say,
10-30 minutes of startup (so all clients should be reconnected), I'd it's nothing
to be concerned about, unless they come back.
- Patrick
________________________________________
From: HPDD-discuss [hpdd-discuss-bounces(a)lists.01.org] on behalf of Kumar,
Amit [ahkumar(a)mail.smu.edu]
Sent: Thursday, July 30, 2015 4:39 PM
To: Mohr Jr, Richard Frank (Rick Mohr)
Cc: hpdd-discuss(a)lists.01.org
Subject: Re: [HPDD-discuss] One of several Clients failing to mount handful
OST's
Upon e2fsck -fp /dev/...
These OST's recover from "rc = -30". But I see a trend with them running
into
slow creates, it his part of the recovery cycle and should it subside eventually?
Lustre: scratch-OST001d-osc-MDT0000: slow creates,
last=[0x1001d0000:0xc1941:0x0], next=[0x1001d0000:0xc1941:0x0],
reserved=0, syn_changes=0, syn_rpc_in_progress=0, status=-19
I see similar entries in
https://jira.hpdd.intel.com/browse/LU-2279, any
thoughts on: if this is an issue of concern will be very helpful.
Thank you,
Amit
>-----Original Message-----
>From: HPDD-discuss [mailto:hpdd-discuss-bounces@lists.01.org] On Behalf
>Of Kumar, Amit
>Sent: Thursday, July 30, 2015 1:13 PM
>To: Mohr Jr, Richard Frank (Rick Mohr)
>Cc: hpdd-discuss(a)lists.01.org
>Subject: Re: [HPDD-discuss] One of several Clients failing to mount
>handful OST's
>
>Hi Rick,
>>
>>I have encountered similar errors on some of my OSTs when things did
>>not shutdown cleanly. I just disabled quotas on the OST and then
>>re-enabled them so that they would be recalculated. After that,
>>e2fsck did not report any errors.
>>
>Version of my lustre: 2.4.3
>
>I fail to see if see any difference in the quota settings, after I
>disable them as follows
>
>lctl conf_param scratch.quota.ost=none
>lctl conf_param scratch.quota.mdt=none
>
>Also looking at the troubled OST's parameters "before" and
"after"
>applying the above commands on MGS, I don't think it had any effect.
>Please let me know if you did anything else to disable quota on OST's.
>
>Thank you,
>Amit
>
># lctl get_param osd-*.*.quota_slave.info osd-ldiskfs.scratch-
>OST001a.quota_slave.info=
>target name: scratch-OST001a
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001b.quota_slave.info=
>target name: scratch-OST001b
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001c.quota_slave.info=
>target name: scratch-OST001c
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001d.quota_slave.info=
>target name: scratch-OST001d
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001e.quota_slave.info=
>target name: scratch-OST001e
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001f.quota_slave.info=
>target name: scratch-OST001f
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>[root@mfoss05 scratch-OST001d]# lctl get_param osd-*.*.quota_slave.info
>osd-ldiskfs.scratch-OST001a.quota_slave.info=
>target name: scratch-OST001a
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001b.quota_slave.info=
>target name: scratch-OST001b
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001c.quota_slave.info=
>target name: scratch-OST001c
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001d.quota_slave.info=
>target name: scratch-OST001d
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001e.quota_slave.info=
>target name: scratch-OST001e
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>osd-ldiskfs.scratch-OST001f.quota_slave.info=
>target name: scratch-OST001f
>pool ID: 0
>type: dt
>quota enabled: none
>conn to master: setup
>space acct: ug
>user uptodate: glb[0],slv[0],reint[0]
>group uptodate: glb[0],slv[0],reint[0]
>_______________________________________________
>HPDD-discuss mailing list
>HPDD-discuss(a)lists.01.org
>https://lists.01.org/mailman/listinfo/hpdd-discuss
_______________________________________________
HPDD-discuss mailing list
HPDD-discuss(a)lists.01.org
https://lists.01.org/mailman/listinfo/hpdd-discuss