lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <alpine.DEB.2.00.1308152201170.1479@cobra.newdream.net>
Date:	Thu, 15 Aug 2013 22:02:01 -0700 (PDT)
From:	Sage Weil <sage@...tank.com>
To:	Li Wang <liwang@...ntukylin.com>
cc:	ceph-devel@...r.kernel.org, linux-kernel@...r.kernel.org,
	Yunchuan Wen <yunchuanwen@...ntukylin.com>
Subject: Re: [PATCH v5] Ceph: Punch hole support for kernel client

I've applied this to the testing branch and moved to the better fsx in the 
qa suite.

The ceph-fuse patches are still in wip-fallocate until I can run the fs 
test suite against them.

Thanks!
sage


On Thu, 15 Aug 2013, Li Wang wrote:

> This patch implements fallocate and punch hole support for Ceph kernel client.
> 
> Signed-off-by: Li Wang <liwang@...ntukylin.com>
> Signed-off-by: Yunchuan Wen <yunchuanwen@...ntukylin.com>
> ---
> Against v3:
> 
> Passed the fsx test from xfstests.
> 
> Truncate rather than delete the first object. Thanks go to Sage and Zheng for the explanation.
> 
> Silence the OSD ENOENT complaints.
> ---
>  fs/ceph/file.c        |  196 +++++++++++++++++++++++++++++++++++++++++++++++++
>  net/ceph/osd_client.c |   11 ++-
>  2 files changed, 205 insertions(+), 2 deletions(-)
> 
> diff --git a/fs/ceph/file.c b/fs/ceph/file.c
> index 2ddf061..e2bcd5c 100644
> --- a/fs/ceph/file.c
> +++ b/fs/ceph/file.c
> @@ -8,6 +8,7 @@
>  #include <linux/namei.h>
>  #include <linux/writeback.h>
>  #include <linux/aio.h>
> +#include <linux/falloc.h>
>  
>  #include "super.h"
>  #include "mds_client.h"
> @@ -871,6 +872,200 @@ out:
>  	return offset;
>  }
>  
> +static inline void ceph_zero_partial_page(
> +	struct inode *inode, loff_t offset, unsigned size)
> +{
> +	struct page *page;
> +	pgoff_t index = offset >> PAGE_CACHE_SHIFT;
> +
> +	page = find_lock_page(inode->i_mapping, index);
> +	if (page) {
> +		wait_on_page_writeback(page);
> +		zero_user(page, offset & (PAGE_CACHE_SIZE - 1), size);
> +		unlock_page(page);
> +		page_cache_release(page);
> +	}
> +}
> +
> +static void ceph_zero_pagecache_range(struct inode *inode, loff_t offset,
> +				      loff_t length)
> +{
> +	loff_t nearly = round_up(offset, PAGE_CACHE_SIZE);
> +	if (offset < nearly) {
> +		loff_t size = nearly - offset;
> +		if (length < size)
> +			size = length;
> +		ceph_zero_partial_page(inode, offset, size);
> +		offset += size;
> +		length -= size;
> +	}
> +	if (length >= PAGE_CACHE_SIZE) {
> +		loff_t size = round_down(length, PAGE_CACHE_SIZE);
> +		truncate_pagecache_range(inode, offset, offset + size - 1);
> +		offset += size;
> +		length -= size;
> +	}
> +	if (length)
> +		ceph_zero_partial_page(inode, offset, length);
> +}
> +
> +static int ceph_zero_partial_object(struct inode *inode,
> +				    loff_t offset, loff_t *length)
> +{
> +	struct ceph_inode_info *ci = ceph_inode(inode);
> +	struct ceph_fs_client *fsc = ceph_inode_to_client(inode);
> +	struct ceph_osd_request *req;
> +	int ret = 0;
> +	loff_t zero = 0;
> +	int op;
> +
> +	if (!length) {
> +		op = offset ? CEPH_OSD_OP_DELETE : CEPH_OSD_OP_TRUNCATE;
> +		length = &zero;
> +	} else {
> +		op = CEPH_OSD_OP_ZERO;
> +	}
> +
> +	req = ceph_osdc_new_request(&fsc->client->osdc, &ci->i_layout,
> +					ceph_vino(inode),
> +					offset, length,
> +					1, op,
> +					CEPH_OSD_FLAG_WRITE |
> +					CEPH_OSD_FLAG_ONDISK,
> +					NULL, 0, 0, false);
> +	if (IS_ERR(req)) {
> +		ret = PTR_ERR(req);
> +		goto out;
> +	}
> +
> +	ceph_osdc_build_request(req, offset, NULL, ceph_vino(inode).snap,
> +				&inode->i_mtime);
> +
> +	ret = ceph_osdc_start_request(&fsc->client->osdc, req, false);
> +	if (!ret) {
> +		ret = ceph_osdc_wait_request(&fsc->client->osdc, req);
> +		if (ret == -ENOENT)
> +			ret = 0;
> +	}
> +	ceph_osdc_put_request(req);
> +
> +out:
> +	return ret;
> +}
> +
> +static int ceph_zero_objects(struct inode *inode, loff_t offset, loff_t length)
> +{
> +	int ret = 0;
> +	struct ceph_inode_info *ci = ceph_inode(inode);
> +	__s32 stripe_unit = ceph_file_layout_su(ci->i_layout);
> +	__s32 stripe_count = ceph_file_layout_stripe_count(ci->i_layout);
> +	__s32 object_size = ceph_file_layout_object_size(ci->i_layout);
> +	loff_t object_set_size = (loff_t)object_size * stripe_count;
> +
> +	loff_t nearly = (offset + object_set_size - 1)
> +			/ object_set_size * object_set_size;
> +	while (length && offset < nearly) {
> +		loff_t size = length;
> +		ret = ceph_zero_partial_object(inode, offset, &size);
> +		if (ret < 0)
> +			return ret;
> +		offset += size;
> +		length -= size;
> +	}
> +	while (length >= object_set_size) {
> +		int i;
> +		loff_t pos = offset;
> +		for (i = 0; i < stripe_count; ++i) {
> +			ret = ceph_zero_partial_object(inode, pos, NULL);
> +			if (ret < 0)
> +				return ret;
> +			pos += stripe_unit;
> +		}
> +		offset += object_set_size;
> +		length -= object_set_size;
> +	}
> +	while (length) {
> +		loff_t size = length;
> +		ret = ceph_zero_partial_object(inode, offset, &size);
> +		if (ret < 0)
> +			return ret;
> +		offset += size;
> +		length -= size;
> +	}
> +	return ret;
> +}
> +
> +static long ceph_fallocate(struct file *file, int mode,
> +				loff_t offset, loff_t length)
> +{
> +	struct ceph_file_info *fi = file->private_data;
> +	struct inode *inode = file->f_dentry->d_inode;
> +	struct ceph_inode_info *ci = ceph_inode(inode);
> +	struct ceph_osd_client *osdc =
> +		&ceph_inode_to_client(inode)->client->osdc;
> +	int want, got = 0;
> +	int dirty;
> +	int ret = 0;
> +	loff_t endoff = 0;
> +	loff_t size;
> +
> +	if (!S_ISREG(inode->i_mode))
> +		return -EOPNOTSUPP;
> +
> +	if (IS_SWAPFILE(inode))
> +		return -ETXTBSY;
> +
> +	mutex_lock(&inode->i_mutex);
> +
> +	if (ceph_snap(inode) != CEPH_NOSNAP) {
> +		ret = -EROFS;
> +		goto unlock;
> +	}
> +
> +	if (ceph_osdmap_flag(osdc->osdmap, CEPH_OSDMAP_FULL) &&
> +		!(mode & FALLOC_FL_PUNCH_HOLE)) {
> +		ret = -ENOSPC;
> +		goto unlock;
> +	}
> +
> +	size = i_size_read(inode);
> +	if (!(mode & FALLOC_FL_KEEP_SIZE))
> +		endoff = offset + length;
> +
> +	if (fi->fmode & CEPH_FILE_MODE_LAZY)
> +		want = CEPH_CAP_FILE_BUFFER | CEPH_CAP_FILE_LAZYIO;
> +	else
> +		want = CEPH_CAP_FILE_BUFFER;
> +
> +	ret = ceph_get_caps(ci, CEPH_CAP_FILE_WR, want, &got, endoff);
> +	if (ret < 0)
> +		goto unlock;
> +
> +	if (mode & FALLOC_FL_PUNCH_HOLE) {
> +		if (offset < size)
> +			ceph_zero_pagecache_range(inode, offset, length);
> +		ret = ceph_zero_objects(inode, offset, length);
> +	} else if (endoff > size) {
> +		truncate_pagecache_range(inode, size, -1);
> +		if (ceph_inode_set_size(inode, endoff))
> +			ceph_check_caps(ceph_inode(inode),
> +				CHECK_CAPS_AUTHONLY, NULL);
> +	}
> +
> +	if (!ret) {
> +		spin_lock(&ci->i_ceph_lock);
> +		dirty = __ceph_mark_dirty_caps(ci, CEPH_CAP_FILE_WR);
> +		spin_unlock(&ci->i_ceph_lock);
> +		if (dirty)
> +			__mark_inode_dirty(inode, dirty);
> +	}
> +
> +	ceph_put_cap_refs(ci, got);
> +unlock:
> +	mutex_unlock(&inode->i_mutex);
> +	return ret;
> +}
> +
>  const struct file_operations ceph_file_fops = {
>  	.open = ceph_open,
>  	.release = ceph_release,
> @@ -887,5 +1082,6 @@ const struct file_operations ceph_file_fops = {
>  	.splice_write = generic_file_splice_write,
>  	.unlocked_ioctl = ceph_ioctl,
>  	.compat_ioctl	= ceph_ioctl,
> +	.fallocate	= ceph_fallocate,
>  };
>  
> diff --git a/net/ceph/osd_client.c b/net/ceph/osd_client.c
> index dd47889..c1d15ab 100644
> --- a/net/ceph/osd_client.c
> +++ b/net/ceph/osd_client.c
> @@ -503,7 +503,9 @@ void osd_req_op_extent_init(struct ceph_osd_request *osd_req,
>  	struct ceph_osd_req_op *op = _osd_req_op_init(osd_req, which, opcode);
>  	size_t payload_len = 0;
>  
> -	BUG_ON(opcode != CEPH_OSD_OP_READ && opcode != CEPH_OSD_OP_WRITE);
> +	BUG_ON(opcode != CEPH_OSD_OP_READ && opcode != CEPH_OSD_OP_WRITE &&
> +	       opcode != CEPH_OSD_OP_DELETE && opcode != CEPH_OSD_OP_ZERO &&
> +	       opcode != CEPH_OSD_OP_TRUNCATE);
>  
>  	op->extent.offset = offset;
>  	op->extent.length = length;
> @@ -631,6 +633,9 @@ static u64 osd_req_encode_op(struct ceph_osd_request *req,
>  		break;
>  	case CEPH_OSD_OP_READ:
>  	case CEPH_OSD_OP_WRITE:
> +	case CEPH_OSD_OP_ZERO:
> +	case CEPH_OSD_OP_DELETE:
> +	case CEPH_OSD_OP_TRUNCATE:
>  		if (src->op == CEPH_OSD_OP_WRITE)
>  			request_data_len = src->extent.length;
>  		dst->extent.offset = cpu_to_le64(src->extent.offset);
> @@ -715,7 +720,9 @@ struct ceph_osd_request *ceph_osdc_new_request(struct ceph_osd_client *osdc,
>  	u64 object_base;
>  	int r;
>  
> -	BUG_ON(opcode != CEPH_OSD_OP_READ && opcode != CEPH_OSD_OP_WRITE);
> +	BUG_ON(opcode != CEPH_OSD_OP_READ && opcode != CEPH_OSD_OP_WRITE &&
> +	       opcode != CEPH_OSD_OP_DELETE && opcode != CEPH_OSD_OP_ZERO &&
> +	       opcode != CEPH_OSD_OP_TRUNCATE);
>  
>  	req = ceph_osdc_alloc_request(osdc, snapc, num_ops, use_mempool,
>  					GFP_NOFS);
> -- 
> 1.7.9.5
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@...r.kernel.org
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ