File Size < on disk size - currently unexplained, size on disk is 2-3 x file size #9855
-
zdb -b filesystem022-OST21.txt Observations:
Investigation: server01]</users/user001>$ du -h --apparent-size /lus/filesystem022/project/file_name/* [server01]</users/user001>$ du -h /lus/filesystem022/project/file_name/*
[server01]</users/user001>$ cp -rp /lus/filesystem022/project/file_name /lus/filesystem022/project/p005j02_2010_SRME_1238A018_JC [server01]</users/user001>$ md5sum /lus/filesystem022/project/file_name/* [server01]</users/user001>$ md5sum /lus/filesystem022/project/p005j02_2010_SRME_1238A018_JC/* [server01]</users/user001>$ du -h /lus/filesystem022/project/p005j02_2010_SRME_1238A018_JC/* [server01]</users/user001>$ du -h --apparent-size /lus/filesystem022/project/p005j02_2010_SRME_1238A018_JC/*
[root@filesystem022-oss6 ~]# zdb -b filesystem022-OST21 loading space map for vdev 0 of 1, metaslab 180 of 181 ... Please would someone be able to advise. Thanks |
Beta Was this translation helpful? Give feedback.
Replies: 4 comments
-
Would you mind not destroying the preformated issue form? |
Beta Was this translation helpful? Give feedback.
-
Apologies. Please see reformatted information below: System information
Describe the problem you're observingFile Size < on disk size - currently unexplained, size on disk is 2-3 x file size Observations:
Investigation:
Describe how to reproduce the problemInclude any warning/errors/backtraces from the system logs |
Beta Was this translation helpful? Give feedback.
-
The main thing to be aware of when using This means the warnings you've seen are probably not indicative of a problem if you ran
As for the original issue is this something you're seeing on every file? Since you're using Lustre I assume the |
Beta Was this translation helpful? Give feedback.
-
Hi Brian,
Thanks for the info. I believe the zdb check was probably run with the zpools imported. Will look at possible opportunity to run offline.
File size < on disk size : We do not see this on every file. The data appears to be ok. The output from the du command is returned from the lustre mount point.
Zfs copies are set to 1
# zfs get copies
NAME PROPERTY VALUE SOURCE
lsi0xx-OST17 copies 1 default
lsi0xx-OST17/lsi0xx-OST0005 copies 1 default
lsi0xx-OST19 copies 1 default
lsi0xx-OST19/lsi0xx-OST0009 copies 1 default
lsi0xx-OST21 copies 1 default
lsi0xx-OST21/lsi0xx-OST000f copies 1 default
lsi0xx-OST23 copies 1 default
lsi0xx-OST23/lsi0xx-OST0013 copies 1 default
We are currently running a zfs scrub on one pool that has exhibited the issue on some files, with 0B repaired so far:
# zpool status lsi0xx-OST21
pool: lsi0xx-OST21
state: ONLINE
scan: scrub in progress since Thu Jan 16 13:42:33 2020
55.8T scanned out of 67.0T at 97.2M/s, 33h37m to go
0B repaired, 83.27% done
config:
NAME STATE READ WRITE CKSUM
lsi0xx-OST21 ONLINE 0 0 0
raidz2-0 ONLINE 0 0 0
JBOD_5_6_SLOT_24 ONLINE 0 0 0
JBOD_5_6_SLOT_25 ONLINE 0 0 0
JBOD_5_6_SLOT_26 ONLINE 0 0 0
JBOD_5_6_SLOT_27 ONLINE 0 0 0
JBOD_5_6_SLOT_28 ONLINE 0 0 0
JBOD_5_6_SLOT_36 ONLINE 0 0 0
JBOD_5_6_SLOT_37 ONLINE 0 0 0
JBOD_5_6_SLOT_38 ONLINE 0 0 0
JBOD_5_6_SLOT_39 ONLINE 0 0 0
JBOD_5_6_SLOT_40 ONLINE 0 0 0
errors: No known data errors
We have several lustre storage systems built with a ZFS filesystem and to a degree they all show some files with the same symptoms
Any further suggestions would be appreciated
Regards
Nick
From: Brian Behlendorf <[email protected]>
Sent: 21 January 2020 22:21
To: zfsonlinux/zfs <[email protected]>
Cc: Nick Skingle <[email protected]>; Author <[email protected]>
Subject: [External] Re: [zfsonlinux/zfs] File Size < on disk size - currently unexplained, size on disk is 2-3 x file size (#9855)
* zdb check for leaks, it walks the entire block tree constructing the space maps in memory and then compares them to the ones stored on disk. If they differ it reports the leak.
The main thing to be aware of when using zdb to check for leaks is the pool must not be online when running zdb. If the pool is imported and active zdb will incorrectly report leaked space. This is because zdb effectively imports the pool read-only in user space, so changes made to the imported pool by the kernel module will not be correctly accounted for by zdb.
This means the warnings you've seen are probably not indicative of a problem if you ran zdb against the imported pool. Unfortunately, to be absolutely certain you would need to stop the OSS then use zdb -e verify the space maps.
File Size < on disk size
As for the original issue is this something you're seeing on every file? Since you're using Lustre I assume the du output posted is what's returned via the Lustre mount point? Is it possible that you've set the zfs copies property to a value other than 1?
—
You are receiving this because you authored the thread.
Reply to this email directly, view it on GitHub<https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_zfsonlinux_zfs_issues_9855-3Femail-5Fsource-3Dnotifications-26email-5Ftoken-3DAKINXQUABTLJURAZ7VARPNLQ65YK5A5CNFSM4KIEWI7KYY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOEJRPWYY-23issuecomment-2D576912227&d=DwMCaQ&c=KV_I7O14pmwRcmAVyJ1eg4Jwb8Y2JAxuL5YgMGHpjcQ&r=2_4LnTcS-T76QITHqbAo2LU1OQ-6bfJjf5CMuaMjSVQ&m=uGzm6HqiQHqtKy-YUC0t0L4aY4O5s_UQAdXzBUsHt9o&s=SGmLKZ8V-g7cKmLUSyKNriN98NXg9pQ6kVnZtHPB0og&e=>, or unsubscribe<https://urldefense.proofpoint.com/v2/url?u=https-3A__github.com_notifications_unsubscribe-2Dauth_AKINXQUX2JS3U6XC5EQF4DTQ65YK5ANCNFSM4KIEWI7A&d=DwMCaQ&c=KV_I7O14pmwRcmAVyJ1eg4Jwb8Y2JAxuL5YgMGHpjcQ&r=2_4LnTcS-T76QITHqbAo2LU1OQ-6bfJjf5CMuaMjSVQ&m=uGzm6HqiQHqtKy-YUC0t0L4aY4O5s_UQAdXzBUsHt9o&s=jc7fYRdkOsqGncpZp8HKT2eMp0a8DgLtwi2HZFb1Nv8&e=>.
|
Beta Was this translation helpful? Give feedback.
The main thing to be aware of when using
zdb
to check for leaks is the pool must not be online when runningzdb
. If the pool is imported and activezdb
will incorrectly report leaked space. This is becausezdb
effectively imports the pool read-only in user space, so changes made to the imported pool by the kernel module will not be correctly accounted for byzdb
.This means the warnings you've seen are probably not indicative of a problem if you ran
zdb
against the imported pool. Unfortunately, to be absolutely…