lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <71009a7a-7875-bf6b-5db4-3eeeb6a787af@web.de>
Date:   Mon, 6 Nov 2017 18:35:22 +0100
From:   mfe555 <mfe555@....de>
To:     "Mattthew L. Martin" <linux@...arskydata.com>,
        linux-ext4@...r.kernel.org
Cc:     mfe555@....de
Subject: Re: Significant difference in 'file size' and 'disk usage' for single
 files

Dear Mathew,

thank you very much for your message and for your offer of helping me.

In my case, the file system has a cluster size of 262144. bigalloc is 
enabled, please see below for details (tune2fs). I have been able to 
confirm that unmounting and re-mounting the file system helps.

Please let me know what else I can do for giving you more clues. For 
example, as our linux system is built for over 100 different settop 
boxes, I might be able to get help from other people, performing tests 
on specific linux kernels.

Kind regards
Lukas

=================================
# tune2fs -l /dev/sdb1
tune2fs 1.43.4 (31-Jan-2017)
Filesystem volume name:   <none>
Last mounted on:          /media/hdd
Filesystem UUID:          1dbc401d-3ff4-4a46-acc7-8ec7b841bdb0
Filesystem magic number:  0xEF53
Filesystem revision #:    1 (dynamic)
Filesystem features:      has_journal ext_attr resize_inode dir_index 
filetype needs_recovery extent flex_bg sparse_super large_file huge_file 
uninit_bg dir_nlink extra_isize bigalloc
Filesystem flags:         signed_directory_hash
Default mount options:    user_xattr acl
Filesystem state:         clean
Errors behavior:          Continue
Filesystem OS type:       Linux
Inode count:              264688
Block count:              488378368
Reserved block count:     0
Free blocks:              146410368
Free inodes:              260432
First block:              0
Block size:               4096
Cluster size:             262144
Reserved GDT blocks:      14
Blocks per group:         2097152
Clusters per group:       32768
Inodes per group:         1136
Inode blocks per group:   71
Flex block group size:    16
Filesystem created:       Sun Mar 13 16:31:29 2016
Last mount time:          Thu Jan  1 01:00:04 1970
Last write time:          Thu Jan  1 01:00:04 1970
Mount count:              884
Maximum mount count:      -1
Last checked:             Sun Mar 13 16:31:29 2016
Check interval:           0 (<none>)
Lifetime writes:          6971 GB
Reserved blocks uid:      0 (user root)
Reserved blocks gid:      0 (group root)
First inode:              11
Inode size:               256
Required extra isize:     28
Desired extra isize:      28
Journal inode:            8
Default directory hash:   half_md4
Directory Hash Seed:      c69a1039-0065-4c1b-8732-ff1b52b57313
Journal backup:           inode blocks




Am 06.11.2017 um 16:35 schrieb Mattthew L. Martin:
> I filed a bug for this a while ago:
>
> https://bugzilla.kernel.org/show_bug.cgi?id=151491
>
> We would be happy to help track this down as it is a pain to manage 
> this on running servers.
>
> Matthew
>
>
> On 11/5/17 06:16, mfe555 wrote:
>> Some follow-up:
>>
>> The issue only occurs with "bigalloc" enabled.
>>
>>     echo 3 > /proc/sys/vm/drop_caches
>>
>> seems to detach the blocked disk space from the files (so that 'du 
>> file' no longer includes the offset), but it does not free the space, 
>> 'df' still shows all file overheads as used disk space.
>>
>>
>>
>> Am 02.11.2017 um 20:17 schrieb mfe555:
>>> Hi, I'm using ext4 on a Linux based Enigma2 set-top box, kernel 4.8.3.
>>>
>>> When creating a fresh file, there is a significant difference in 
>>> file size (ls -la) and disk usage (du). When making two copies of 
>>> the file ..
>>>
>>> gbquad:/hdd/test# cp file file.copy1
>>> gbquad:/hdd/test# cp file file.copy2
>>> gbquad:/hdd/test# ls -la
>>> -rw-------    1 root     root     581821460 Nov  1 18:52 file
>>> -rw-------    1 root     root     581821460 Nov  1 18:56 file.copy1
>>> -rw-------    1 root     root     581821460 Nov  1 18:57 file.copy2
>>> gbquad:/hdd/test# du *
>>> 607232  file
>>> 658176  file.copy1
>>> 644864  file.copy2
>>>
>>> ... all three files show an overhead in the ~10% range, and the 
>>> overhead is different for these files although their md5sums are equal.
>>>
>>> When deleting a file (rm), the overhead remains occupied on the 
>>> disk. For example, after deleting "file", "df" reports approx. 
>>> 581821460 more bytes free, not 607232 kbytes more free space. The 
>>> overhead (607232 kB - 581821460 B =pprox. 39 MB) remains blocked.
>>>
>>> When re-booting, the blocked space becomes free again, and in 
>>> addition the overhead of those files that were not deleted also 
>>> disappears, so that after a reboot the'file size' and 'disk usage' 
>>> match for all files (except for rounding up to some block size).
>>>
>>> A colleague and I have observed this on two different "kernel 4.8.3" 
>>> boxes and three ext4 disks, but not on a "kernel 3.14" box also 
>>> using ext4.
>>>
>>> Can anyone help me with this ?
>>>
>>> Thanks a lot
>>> Lukas
>>>
>>
>

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ