[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <87bo8v42wx.fsf@devron.myhome.or.jp>
Date: Wed, 01 May 2013 15:42:06 +0900
From: OGAWA Hirofumi <hirofumi@...l.parknet.co.jp>
To: Namjae Jeon <linkinjeon@...il.com>
Cc: akpm@...ux-foundation.org, linux-fsdevel@...r.kernel.org,
linux-kernel@...r.kernel.org,
Namjae Jeon <namjae.jeon@...sung.com>,
Amit Sahrawat <a.sahrawat@...sung.com>
Subject: Re: [PATCH RESEND v5] fat: editions to support fat_fallocate
Namjae Jeon <linkinjeon@...il.com> writes:
> Hi OGAWA.
Hi,
>> I couldn't review fully though.
>>
>>> + if (mmu_private_ideal < MSDOS_I(inode)->mmu_private &&
>>> + filp->f_dentry->d_count == 1)
>>> + fat_truncate_blocks(inode, inode->i_size);
>>
>> Hm, why d_count == 1 check is needed? Feel strange and racy.
> Since, fat_file_release() is called on every close for the file.
What is wrong? IIRC, it is what you choose (i.e. for each last close for
the file descriptor).
> But we want to free up the reserved blocks only at the last reference
> for the file exits.
> So, we have used “d_count ==1” i.e., when there is only one reference
> left for the file and it is being closed.
> Then call the truncate blocks to free up the space.
It probably doesn't work. E.g. if unlink(2) is grabbing refcount, then
close(2) may not be last referencer, right?
So, then, nobody truncates anymore.
>>> + /* Start the allocation.We are not zeroing out the clusters */
>>> + while (nr_cluster-- > 0) {
>>> + err = fat_alloc_clusters(inode, &cluster, 1);
>>
>> Why doesn't allocate clusters at once by fat_alloc_clusters()?
> It is because of default design, where we cannot allocate all the
> clusters at once. For reference if we try to allocate all clusters at
> once, it will trigger a BUG_ON in
> fat_alloc_clusters()->
> BUG_ON(nr_cluster > (MAX_BUF_PER_PAGE / 2)); /* fixed limit */
> And we needed to update the fat chain after each allocation and take
> care of the failure cases as well. So, we have done that sequential.
> That optimization of allocating all clusters at once can be considered
> as a separate changeline.
OK.
>>> + size = i_size_read(inode);
>>> + mmu_private_actual = MSDOS_I(inode)->mmu_private;
>>> + mmu_private_ideal = round_up(size, sb->s_blocksize);
>>> + if ((mmu_private_actual > mmu_private_ideal) && (pos > size)) {
>>> + err = fat_zero_falloc_area(file, mapping, pos);
>>> + if (err) {
>>> + fat_msg(sb, KERN_ERR,
>>> + "Error (%d) zeroing fallocated area", err);
>>> + return err;
>>> + }
>>> + }
>>
>> This way probably inefficient. This would write data twice times (one is
>> zeroed, one is actual data). So, cpu time would be twice higher if
>> user uses fallocated, right?
> We introduced the “zeroing out” after there was a comment regarding
> the security loophole of accessing invalid data.
> So, while doing fallocate we reserved the space. But, if there is a
> request to access the pre-allocated space we zeroout the complete area
> to avoid any security issue.
I know. Question is, why do we need to initialize twice.
1) zeroed for uninitialized area, 2) then copy user data area. We need
only either, right? This seems to be doing both for all fallocated area.
Thanks.
--
OGAWA Hirofumi <hirofumi@...l.parknet.co.jp>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/
Powered by blists - more mailing lists