[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <5116690d-0214-d55c-d3de-4cb6850d7a4d@kernel.org>
Date: Sun, 24 Jul 2022 17:27:17 +0800
From: Chao Yu <chao@...nel.org>
To: Daeho Jeong <daehojeong@...gle.com>
Cc: Jaegeuk Kim <jaegeuk@...nel.org>, Daeho Jeong <daeho43@...il.com>,
kernel-team@...roid.com, linux-kernel@...r.kernel.org,
linux-f2fs-devel@...ts.sourceforge.net
Subject: Re: [f2fs-dev] [PATCH 2/2] f2fs: handle decompress only post
processing in softirq
On 2022/7/18 23:47, Daeho Jeong wrote:
> On Sat, Jul 16, 2022 at 8:22 PM Chao Yu <chao@...nel.org> wrote:
>>
>> On 2022/6/22 0:07, Jaegeuk Kim wrote:
>>> Can we do like this which has no functional change but refactored
>>> some functions?
>>>
>>> ---
>>> fs/f2fs/compress.c | 208 ++++++++++++++++++++++++++++-----------------
>>> fs/f2fs/data.c | 52 ++++++++----
>>> fs/f2fs/f2fs.h | 17 ++--
>>> 3 files changed, 172 insertions(+), 105 deletions(-)
>>>
>>> diff --git a/fs/f2fs/compress.c b/fs/f2fs/compress.c
>>> index fa237e5c7173..494ce3634b62 100644
>>> --- a/fs/f2fs/compress.c
>>> +++ b/fs/f2fs/compress.c
>>> @@ -729,28 +729,18 @@ static int f2fs_compress_pages(struct compress_ctx *cc)
>>> return ret;
>>> }
>>>
>>> -void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
>>> +static int f2fs_prepare_decomp_mem(struct decompress_io_ctx *dic, bool end_io)
>>> {
>>> - struct f2fs_sb_info *sbi = F2FS_I_SB(dic->inode);
>>> - struct f2fs_inode_info *fi = F2FS_I(dic->inode);
>>> const struct f2fs_compress_ops *cops =
>>> - f2fs_cops[fi->i_compress_algorithm];
>>> - int ret;
>>> + f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
>>> int i;
>>>
>>> - trace_f2fs_decompress_pages_start(dic->inode, dic->cluster_idx,
>>> - dic->cluster_size, fi->i_compress_algorithm);
>>> -
>>> - if (dic->failed) {
>>> - ret = -EIO;
>>> - goto out_end_io;
>>> - }
>>> + if (!(end_io ^ f2fs_low_mem_mode(F2FS_I_SB(dic->inode))))
>>
>> How about using allow_decompress_in_softirq() to wrap !f2fs_low_mem_mode()
>> to improve readability?
>>
>>> + return 0;
>>>
>>> dic->tpages = page_array_alloc(dic->inode, dic->cluster_size);
>>> - if (!dic->tpages) {
>>> - ret = -ENOMEM;
>>> - goto out_end_io;
>>> - }
>>> + if (!dic->tpages)
>>> + return 1;
>>
>> return -ENOMEM instead of magic number.
>
> The callers already change the error number to ENOMEM when it gets 1.
> This is the same flow in the previous code. Do you want to change
> this?
Yes, please. :)
>
>>
>>>
>>> for (i = 0; i < dic->cluster_size; i++) {
>>> if (dic->rpages[i]) {
>>> @@ -759,28 +749,100 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
>>> }
>>>
>>> dic->tpages[i] = f2fs_compress_alloc_page();
>>> - if (!dic->tpages[i]) {
>>> - ret = -ENOMEM;
>>> - goto out_end_io;
>>> - }
>>> + if (!dic->tpages[i])
>>> + return 1;
>>
>> Ditto,
>>
>>> }
>>>
>>> + dic->rbuf = f2fs_vmap(dic->tpages, dic->cluster_size);
>>> + if (!dic->rbuf)
>>> + return 1;
>>
>> Ditto,
>>
>>> +
>>> + dic->cbuf = f2fs_vmap(dic->cpages, dic->nr_cpages);
>>> + if (!dic->cbuf)
>>> + return 1;
>>
>> Ditto,
>>
>>> +
>>> + cops = f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
>>
>> No need to assign cops again?
>
> Ack
>
>>
>>> if (cops->init_decompress_ctx) {
>>> - ret = cops->init_decompress_ctx(dic);
>>> + int ret = cops->init_decompress_ctx(dic);
>>> +
>>> if (ret)
>>> - goto out_end_io;
>>> + return 1;
>>
>> How about returning ret here instead of magic number?
>>
>>> }
>>>
>>> - dic->rbuf = f2fs_vmap(dic->tpages, dic->cluster_size);
>>> - if (!dic->rbuf) {
>>> - ret = -ENOMEM;
>>> - goto out_destroy_decompress_ctx;
>>> + return 0;
>>> +}
>>> +
>>> +static void f2fs_release_decomp_mem(struct decompress_io_ctx *dic,
>>> + bool bypass_destroy_callback, bool end_io)
>>> +{
>>> + const struct f2fs_compress_ops *cops =
>>> + f2fs_cops[F2FS_I(dic->inode)->i_compress_algorithm];
>>> +
>>> + if (end_io ^ f2fs_low_mem_mode(F2FS_I_SB(dic->inode)))
>>> + return;
>>> +
>>> + if (!bypass_destroy_callback && cops->destroy_decompress_ctx)
>>> + cops->destroy_decompress_ctx(dic);
>>> +
>>> + if (dic->cbuf)
>>> + vm_unmap_ram(dic->cbuf, dic->nr_cpages);
>>> +
>>> + if (dic->rbuf)
>>> + vm_unmap_ram(dic->rbuf, dic->cluster_size);
>>> +}
>>> +
>>> +static void f2fs_free_dic(struct decompress_io_ctx *dic,
>>> + bool bypass_destroy_callback)
>>> +{
>>> + int i;
>>> +
>>> + f2fs_release_decomp_mem(dic, bypass_destroy_callback, false);
>>> +
>>> + if (dic->tpages) {
>>> + for (i = 0; i < dic->cluster_size; i++) {
>>> + if (dic->rpages[i])
>>> + continue;
>>> + if (!dic->tpages[i])
>>> + continue;
>>> + f2fs_compress_free_page(dic->tpages[i]);
>>> + }
>>> + page_array_free(dic->inode, dic->tpages, dic->cluster_size);
>>> }
>>>
>>> - dic->cbuf = f2fs_vmap(dic->cpages, dic->nr_cpages);
>>> - if (!dic->cbuf) {
>>> + if (dic->cpages) {
>>> + for (i = 0; i < dic->nr_cpages; i++) {
>>> + if (!dic->cpages[i])
>>> + continue;
>>> + f2fs_compress_free_page(dic->cpages[i]);
>>> + }
>>> + page_array_free(dic->inode, dic->cpages, dic->nr_cpages);
>>> + }
>>> +
>>> + page_array_free(dic->inode, dic->rpages, dic->nr_rpages);
>>> + kmem_cache_free(dic_entry_slab, dic);
>>> +}
>>> +
>>> +void f2fs_decompress_cluster(struct decompress_io_ctx *dic, bool in_task)
>>> +{
>>> + struct f2fs_sb_info *sbi = F2FS_I_SB(dic->inode);
>>> + struct f2fs_inode_info *fi = F2FS_I(dic->inode);
>>> + const struct f2fs_compress_ops *cops =
>>> + f2fs_cops[fi->i_compress_algorithm];
>>> + bool bypass_callback = false;
>>> + int ret;
>>> +
>>> + trace_f2fs_decompress_pages_start(dic->inode, dic->cluster_idx,
>>> + dic->cluster_size, fi->i_compress_algorithm);
>>> +
>>> + if (dic->failed) {
>>> + ret = -EIO;
>>> + goto out_end_io;
>>> + }
>>> +
>>> + if (f2fs_prepare_decomp_mem(dic, true)) {
>>> + bypass_callback = true;
>>> ret = -ENOMEM;
>>> - goto out_vunmap_rbuf;
>>> + goto out_release;
>>> }
>>>
>>> dic->clen = le32_to_cpu(dic->cbuf->clen);
>>> @@ -788,7 +850,7 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
>>>
>>> if (dic->clen > PAGE_SIZE * dic->nr_cpages - COMPRESS_HEADER_SIZE) {
>>> ret = -EFSCORRUPTED;
>>> - goto out_vunmap_cbuf;
>>> + goto out_release;
>>> }
>>>
>>> ret = cops->decompress_pages(dic);
>>> @@ -809,17 +871,13 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
>>> }
>>> }
>>>
>>> -out_vunmap_cbuf:
>>> - vm_unmap_ram(dic->cbuf, dic->nr_cpages);
>>> -out_vunmap_rbuf:
>>> - vm_unmap_ram(dic->rbuf, dic->cluster_size);
>>> -out_destroy_decompress_ctx:
>>> - if (cops->destroy_decompress_ctx)
>>> - cops->destroy_decompress_ctx(dic);
>>> +out_release:
>>> + f2fs_release_decomp_mem(dic, bypass_callback, true);
>>> +
>>> out_end_io:
>>> trace_f2fs_decompress_pages_end(dic->inode, dic->cluster_idx,
>>> dic->clen, ret);
>>> - f2fs_decompress_end_io(dic, ret);
>>> + f2fs_decompress_end_io(dic, ret, in_task);
>>> }
>>>
>>> /*
>>> @@ -829,7 +887,7 @@ void f2fs_decompress_cluster(struct decompress_io_ctx *dic)
>>> * (or in the case of a failure, cleans up without actually decompressing).
>>> */
>>> void f2fs_end_read_compressed_page(struct page *page, bool failed,
>>> - block_t blkaddr)
>>> + block_t blkaddr, bool in_task)
>>> {
>>> struct decompress_io_ctx *dic =
>>> (struct decompress_io_ctx *)page_private(page);
>>> @@ -839,12 +897,12 @@ void f2fs_end_read_compressed_page(struct page *page, bool failed,
>>>
>>> if (failed)
>>> WRITE_ONCE(dic->failed, true);
>>> - else if (blkaddr)
>>> + else if (blkaddr && in_task)
>>> f2fs_cache_compressed_page(sbi, page,
>>> dic->inode->i_ino, blkaddr);
>>>
>>> if (atomic_dec_and_test(&dic->remaining_pages))
>>> - f2fs_decompress_cluster(dic);
>>> + f2fs_decompress_cluster(dic, in_task);
>>> }
>>>
>>> static bool is_page_in_cluster(struct compress_ctx *cc, pgoff_t index)
>>> @@ -1552,16 +1610,14 @@ int f2fs_write_multi_pages(struct compress_ctx *cc,
>>> return err;
>>> }
>>>
>>> -static void f2fs_free_dic(struct decompress_io_ctx *dic);
>>> -
>>> struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc)
>>> {
>>> struct decompress_io_ctx *dic;
>>> pgoff_t start_idx = start_idx_of_cluster(cc);
>>> + struct f2fs_sb_info *sbi = F2FS_I_SB(cc->inode);
>>> int i;
>>>
>>> - dic = f2fs_kmem_cache_alloc(dic_entry_slab, GFP_F2FS_ZERO,
>>> - false, F2FS_I_SB(cc->inode));
>>> + dic = f2fs_kmem_cache_alloc(dic_entry_slab, GFP_F2FS_ZERO, false, sbi);
>>> if (!dic)
>>> return ERR_PTR(-ENOMEM);
>>>
>>> @@ -1602,52 +1658,43 @@ struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc)
>>> dic->cpages[i] = page;
>>> }
>>>
>>> + if (f2fs_prepare_decomp_mem(dic, false))
>>> + goto out_free;
>>> +
>>> return dic;
>>>
>>> out_free:
>>> - f2fs_free_dic(dic);
>>> + f2fs_free_dic(dic, true);
>>> return ERR_PTR(-ENOMEM);
>>> }
>>>
>>> -static void f2fs_free_dic(struct decompress_io_ctx *dic)
>>> +static void f2fs_late_free_dic(struct work_struct *work)
>>> {
>>> - int i;
>>> -
>>> - if (dic->tpages) {
>>> - for (i = 0; i < dic->cluster_size; i++) {
>>> - if (dic->rpages[i])
>>> - continue;
>>> - if (!dic->tpages[i])
>>> - continue;
>>> - f2fs_compress_free_page(dic->tpages[i]);
>>> - }
>>> - page_array_free(dic->inode, dic->tpages, dic->cluster_size);
>>> - }
>>> -
>>> - if (dic->cpages) {
>>> - for (i = 0; i < dic->nr_cpages; i++) {
>>> - if (!dic->cpages[i])
>>> - continue;
>>> - f2fs_compress_free_page(dic->cpages[i]);
>>> - }
>>> - page_array_free(dic->inode, dic->cpages, dic->nr_cpages);
>>> - }
>>> + struct decompress_io_ctx *dic =
>>> + container_of(work, struct decompress_io_ctx, free_work);
>>>
>>> - page_array_free(dic->inode, dic->rpages, dic->nr_rpages);
>>> - kmem_cache_free(dic_entry_slab, dic);
>>> + f2fs_free_dic(dic, false);
>>> }
>>>
>>> -static void f2fs_put_dic(struct decompress_io_ctx *dic)
>>> +static void f2fs_put_dic(struct decompress_io_ctx *dic, bool in_task)
>>> {
>>> - if (refcount_dec_and_test(&dic->refcnt))
>>> - f2fs_free_dic(dic);
>>> + if (refcount_dec_and_test(&dic->refcnt)) {
>>> + if (in_task) {
>>> + f2fs_free_dic(dic, false);
>>> + } else {
>>> + INIT_WORK(&dic->free_work, f2fs_late_free_dic);
>>> + queue_work(F2FS_I_SB(dic->inode)->post_read_wq,
>>> + &dic->free_work);
>>> + }
>>> + }
>>> }
>>>
>>> /*
>>> * Update and unlock the cluster's pagecache pages, and release the reference to
>>> * the decompress_io_ctx that was being held for I/O completion.
>>> */
>>> -static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
>>> +static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
>>> + bool in_task)
>>> {
>>> int i;
>>>
>>> @@ -1668,7 +1715,7 @@ static void __f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
>>> unlock_page(rpage);
>>> }
>>>
>>> - f2fs_put_dic(dic);
>>> + f2fs_put_dic(dic, in_task);
>>> }
>>>
>>> static void f2fs_verify_cluster(struct work_struct *work)
>>> @@ -1685,14 +1732,15 @@ static void f2fs_verify_cluster(struct work_struct *work)
>>> SetPageError(rpage);
>>> }
>>>
>>> - __f2fs_decompress_end_io(dic, false);
>>> + __f2fs_decompress_end_io(dic, false, true);
>>> }
>>>
>>> /*
>>> * This is called when a compressed cluster has been decompressed
>>> * (or failed to be read and/or decompressed).
>>> */
>>> -void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
>>> +void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
>>> + bool in_task)
>>> {
>>> if (!failed && dic->need_verity) {
>>> /*
>>> @@ -1704,7 +1752,7 @@ void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
>>> INIT_WORK(&dic->verity_work, f2fs_verify_cluster);
>>> fsverity_enqueue_verify_work(&dic->verity_work);
>>> } else {
>>> - __f2fs_decompress_end_io(dic, failed);
>>> + __f2fs_decompress_end_io(dic, failed, in_task);
>>> }
>>> }
>>>
>>> @@ -1713,12 +1761,12 @@ void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed)
>>> *
>>> * This is called when the page is no longer needed and can be freed.
>>> */
>>> -void f2fs_put_page_dic(struct page *page)
>>> +void f2fs_put_page_dic(struct page *page, bool in_task)
>>> {
>>> struct decompress_io_ctx *dic =
>>> (struct decompress_io_ctx *)page_private(page);
>>>
>>> - f2fs_put_dic(dic);
>>> + f2fs_put_dic(dic, in_task);
>>> }
>>>
>>> /*
>>> diff --git a/fs/f2fs/data.c b/fs/f2fs/data.c
>>> index 7fcbcf979737..c448c3ee7ac3 100644
>>> --- a/fs/f2fs/data.c
>>> +++ b/fs/f2fs/data.c
>>> @@ -119,7 +119,7 @@ struct bio_post_read_ctx {
>>> block_t fs_blkaddr;
>>> };
>>>
>>> -static void f2fs_finish_read_bio(struct bio *bio)
>>> +static void f2fs_finish_read_bio(struct bio *bio, bool in_task)
>>> {
>>> struct bio_vec *bv;
>>> struct bvec_iter_all iter_all;
>>> @@ -133,8 +133,9 @@ static void f2fs_finish_read_bio(struct bio *bio)
>>>
>>> if (f2fs_is_compressed_page(page)) {
>>> if (bio->bi_status)
>>> - f2fs_end_read_compressed_page(page, true, 0);
>>> - f2fs_put_page_dic(page);
>>> + f2fs_end_read_compressed_page(page, true, 0,
>>> + in_task);
>>> + f2fs_put_page_dic(page, in_task);
>>> continue;
>>> }
>>>
>>> @@ -191,7 +192,7 @@ static void f2fs_verify_bio(struct work_struct *work)
>>> fsverity_verify_bio(bio);
>>> }
>>>
>>> - f2fs_finish_read_bio(bio);
>>> + f2fs_finish_read_bio(bio, true);
>>> }
>>>
>>> /*
>>> @@ -203,7 +204,7 @@ static void f2fs_verify_bio(struct work_struct *work)
>>> * can involve reading verity metadata pages from the file, and these verity
>>> * metadata pages may be encrypted and/or compressed.
>>> */
>>> -static void f2fs_verify_and_finish_bio(struct bio *bio)
>>> +static void f2fs_verify_and_finish_bio(struct bio *bio, bool in_task)
>>> {
>>> struct bio_post_read_ctx *ctx = bio->bi_private;
>>>
>>> @@ -211,7 +212,7 @@ static void f2fs_verify_and_finish_bio(struct bio *bio)
>>> INIT_WORK(&ctx->work, f2fs_verify_bio);
>>> fsverity_enqueue_verify_work(&ctx->work);
>>> } else {
>>> - f2fs_finish_read_bio(bio);
>>> + f2fs_finish_read_bio(bio, in_task);
>>> }
>>> }
>>>
>>> @@ -224,7 +225,8 @@ static void f2fs_verify_and_finish_bio(struct bio *bio)
>>> * that the bio includes at least one compressed page. The actual decompression
>>> * is done on a per-cluster basis, not a per-bio basis.
>>> */
>>> -static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx)
>>> +static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx,
>>> + bool in_task)
>>> {
>>> struct bio_vec *bv;
>>> struct bvec_iter_all iter_all;
>>> @@ -237,7 +239,7 @@ static void f2fs_handle_step_decompress(struct bio_post_read_ctx *ctx)
>>> /* PG_error was set if decryption failed. */
>>> if (f2fs_is_compressed_page(page))
>>> f2fs_end_read_compressed_page(page, PageError(page),
>>> - blkaddr);
>>> + blkaddr, in_task);
>>> else
>>> all_compressed = false;
>>>
>>> @@ -262,15 +264,16 @@ static void f2fs_post_read_work(struct work_struct *work)
>>> fscrypt_decrypt_bio(ctx->bio);
>>>
>>> if (ctx->enabled_steps & STEP_DECOMPRESS)
>>> - f2fs_handle_step_decompress(ctx);
>>> + f2fs_handle_step_decompress(ctx, true);
>>>
>>> - f2fs_verify_and_finish_bio(ctx->bio);
>>> + f2fs_verify_and_finish_bio(ctx->bio, true);
>>> }
>>>
>>> static void f2fs_read_end_io(struct bio *bio)
>>> {
>>> struct f2fs_sb_info *sbi = F2FS_P_SB(bio_first_page_all(bio));
>>> struct bio_post_read_ctx *ctx;
>>> + bool intask = in_task();
>>
>> Is there any condition that in_task() is true here? Maybe I'm missing
>> something here....
>
> I think that error handling cases in block layer before submitting bio
> falls in this.
Yup, I guess you're right. :)
Thanks,
>
>>
>> Thanks,
>>
>>>
>>> iostat_update_and_unbind_ctx(bio, 0);
>>> ctx = bio->bi_private;
>>> @@ -281,16 +284,29 @@ static void f2fs_read_end_io(struct bio *bio)
>>> }
>>>
>>> if (bio->bi_status) {
>>> - f2fs_finish_read_bio(bio);
>>> + f2fs_finish_read_bio(bio, intask);
>>> return;
>>> }
>>>
>>> - if (ctx && (ctx->enabled_steps & (STEP_DECRYPT | STEP_DECOMPRESS))) {
>>> - INIT_WORK(&ctx->work, f2fs_post_read_work);
>>> - queue_work(ctx->sbi->post_read_wq, &ctx->work);
>>> - } else {
>>> - f2fs_verify_and_finish_bio(bio);
>>> + if (ctx) {
>>> + unsigned int enabled_steps = ctx->enabled_steps &
>>> + (STEP_DECRYPT | STEP_DECOMPRESS);
>>> +
>>> + /*
>>> + * If we have only decompression step between decompression and
>>> + * decrypt, we don't need post processing for this.
>>> + */
>>> + if (enabled_steps == STEP_DECOMPRESS &&
>>> + !f2fs_low_mem_mode(sbi)) {
>>> + f2fs_handle_step_decompress(ctx, intask);
>>> + } else if (enabled_steps) {
>>> + INIT_WORK(&ctx->work, f2fs_post_read_work);
>>> + queue_work(ctx->sbi->post_read_wq, &ctx->work);
>>> + return;
>>> + }
>>> }
>>> +
>>> + f2fs_verify_and_finish_bio(bio, intask);
>>> }
>>>
>>> static void f2fs_write_end_io(struct bio *bio)
>>> @@ -2222,7 +2238,7 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret,
>>>
>>> if (f2fs_load_compressed_page(sbi, page, blkaddr)) {
>>> if (atomic_dec_and_test(&dic->remaining_pages))
>>> - f2fs_decompress_cluster(dic);
>>> + f2fs_decompress_cluster(dic, true);
>>> continue;
>>> }
>>>
>>> @@ -2240,7 +2256,7 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret,
>>> page->index, for_write);
>>> if (IS_ERR(bio)) {
>>> ret = PTR_ERR(bio);
>>> - f2fs_decompress_end_io(dic, ret);
>>> + f2fs_decompress_end_io(dic, ret, true);
>>> f2fs_put_dnode(&dn);
>>> *bio_ret = NULL;
>>> return ret;
>>> diff --git a/fs/f2fs/f2fs.h b/fs/f2fs/f2fs.h
>>> index fea97093d927..c9a31934b948 100644
>>> --- a/fs/f2fs/f2fs.h
>>> +++ b/fs/f2fs/f2fs.h
>>> @@ -1588,6 +1588,7 @@ struct decompress_io_ctx {
>>> void *private; /* payload buffer for specified decompression algorithm */
>>> void *private2; /* extra payload buffer */
>>> struct work_struct verity_work; /* work to verify the decompressed pages */
>>> + struct work_struct free_work; /* work for late free this structure itself */
>>> };
>>>
>>> #define NULL_CLUSTER ((unsigned int)(~0))
>>> @@ -4166,9 +4167,9 @@ void f2fs_compress_write_end_io(struct bio *bio, struct page *page);
>>> bool f2fs_is_compress_backend_ready(struct inode *inode);
>>> int f2fs_init_compress_mempool(void);
>>> void f2fs_destroy_compress_mempool(void);
>>> -void f2fs_decompress_cluster(struct decompress_io_ctx *dic);
>>> +void f2fs_decompress_cluster(struct decompress_io_ctx *dic, bool in_task);
>>> void f2fs_end_read_compressed_page(struct page *page, bool failed,
>>> - block_t blkaddr);
>>> + block_t blkaddr, bool in_task);
>>> bool f2fs_cluster_is_empty(struct compress_ctx *cc);
>>> bool f2fs_cluster_can_merge_page(struct compress_ctx *cc, pgoff_t index);
>>> bool f2fs_all_cluster_page_loaded(struct compress_ctx *cc, struct pagevec *pvec,
>>> @@ -4187,8 +4188,9 @@ int f2fs_read_multi_pages(struct compress_ctx *cc, struct bio **bio_ret,
>>> unsigned nr_pages, sector_t *last_block_in_bio,
>>> bool is_readahead, bool for_write);
>>> struct decompress_io_ctx *f2fs_alloc_dic(struct compress_ctx *cc);
>>> -void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed);
>>> -void f2fs_put_page_dic(struct page *page);
>>> +void f2fs_decompress_end_io(struct decompress_io_ctx *dic, bool failed,
>>> + bool in_task);
>>> +void f2fs_put_page_dic(struct page *page, bool in_task);
>>> unsigned int f2fs_cluster_blocks_are_contiguous(struct dnode_of_data *dn);
>>> int f2fs_init_compress_ctx(struct compress_ctx *cc);
>>> void f2fs_destroy_compress_ctx(struct compress_ctx *cc, bool reuse);
>>> @@ -4234,13 +4236,14 @@ static inline struct page *f2fs_compress_control_page(struct page *page)
>>> }
>>> static inline int f2fs_init_compress_mempool(void) { return 0; }
>>> static inline void f2fs_destroy_compress_mempool(void) { }
>>> -static inline void f2fs_decompress_cluster(struct decompress_io_ctx *dic) { }
>>> +static inline void f2fs_decompress_cluster(struct decompress_io_ctx *dic,
>>> + bool in_task) { }
>>> static inline void f2fs_end_read_compressed_page(struct page *page,
>>> - bool failed, block_t blkaddr)
>>> + bool failed, block_t blkaddr, bool in_task)
>>> {
>>> WARN_ON_ONCE(1);
>>> }
>>> -static inline void f2fs_put_page_dic(struct page *page)
>>> +static inline void f2fs_put_page_dic(struct page *page, bool in_task)
>>> {
>>> WARN_ON_ONCE(1);
>>> }
Powered by blists - more mailing lists