Re: [f2fs-dev] [PATCH 2/2] f2fs: support data compression

From: Chao Yu
Date: Tue Oct 29 2019 - 04:33:44 EST


On 2019/10/28 6:50, Eric Biggers wrote:
>> +bool f2fs_is_compressed_page(struct page *page)
>> +{
>> + if (!page_private(page))
>> + return false;
>> + if (IS_ATOMIC_WRITTEN_PAGE(page) || IS_DUMMY_WRITTEN_PAGE(page))
>> + return false;
>> + return *((u32 *)page_private(page)) == F2FS_COMPRESSED_PAGE_MAGIC;
>> +}
>
> This code implies that there can be multiple page private structures each of
> which has a different magic number. But I only see F2FS_COMPRESSED_PAGE_MAGIC.
> Where in the code is the other one(s)?

I'm not sure I understood you correctly, did you mean it needs to introduce
f2fs_is_atomic_written_page() and f2fs_is_dummy_written_page() like
f2fs_is_compressed_page()?

>
>> +
>> +static void f2fs_set_compressed_page(struct page *page,
>> + struct inode *inode, pgoff_t index, void *data, refcount_t *r)
>> +{
>> + SetPagePrivate(page);
>> + set_page_private(page, (unsigned long)data);
>> +
>> + /* i_crypto_info and iv index */
>> + page->index = index;
>> + page->mapping = inode->i_mapping;
>> + if (r)
>> + refcount_inc(r);
>> +}
>
> It isn't really appropriate to create fake pagecache pages like this. Did you
> consider changing f2fs to use fscrypt_decrypt_block_inplace() instead?

We need to store i_crypto_info and iv index somewhere, in order to pass them to
fscrypt_decrypt_block_inplace(), where did you suggest to store them?

>> +
>> +void f2fs_destroy_compress_ctx(struct compress_ctx *cc)
>> +{
>> + kvfree(cc->rpages);
>> +}
>
> The memory is allocated with kzalloc(), so why is it freed with kvfree() and not
> just kfree()?

It was allocated by f2fs_*alloc() which will fallback to kvmalloc() once
kmalloc() failed.

>> +static int lzo_compress_pages(struct compress_ctx *cc)
>> +{
>> + int ret;
>> +
>> + ret = lzo1x_1_compress(cc->rbuf, cc->rlen, cc->cbuf->cdata,
>> + &cc->clen, cc->private);
>> + if (ret != LZO_E_OK) {
>> + printk_ratelimited("%sF2FS-fs: lzo compress failed, ret:%d\n",
>> + KERN_ERR, ret);
>> + return -EIO;
>> + }
>> + return 0;
>> +}
>
> Why not using f2fs_err()? Same in lots of other places.

We use printk_ratelimited at some points where we can afford to lose logs,
otherwise we use f2fs_{err,warn...} to record info as much as possible for
troubleshoot.

>> +
>> + ret = cops->compress_pages(cc);
>> + if (ret)
>> + goto out_vunmap_cbuf;
>> +
>> + max_len = PAGE_SIZE * (cc->cluster_size - 1) - COMPRESS_HEADER_SIZE;
>> +
>> + if (cc->clen > max_len) {
>> + ret = -EAGAIN;
>> + goto out_vunmap_cbuf;
>> + }
>
> Since we already know the max length we're willing to compress to (the max
> length for any space to be saved), why is more space than that being allocated?
> LZ4_compress_default() will return an error if there isn't enough space, so that
> error could just be used as the indication to store the data uncompressed.

AFAIK, there is no such common error code returned from all compression
algorithms indicating there is no room for limited target size, however we need
that information to fallback to write raw pages. Any better idea?

>
>> +
>> + cc->cbuf->clen = cpu_to_le32(cc->clen);
>> + cc->cbuf->chksum = 0;
>
> What is the point of the chksum field? It's always set to 0 and never checked.

When I written initial codes, I doubt that I may lose to check some SPO corner
cases, in where we missed to write whole cluster, so I added that to help to
recall that case, however I didn't have time to cover those cases, resulting
leaving unfinished code there... :(, I'm okay to delete it in a formal version.

BTW, for data checksum feature, I guess we need to reconstruct dnode layout to
cover both compressed/non-compressed data.

>
>> +
>> +static bool __cluster_may_compress(struct compress_ctx *cc)
>> +{
>> + struct f2fs_sb_info *sbi = F2FS_I_SB(cc->inode);
>> + loff_t i_size = i_size_read(cc->inode);
>> + const pgoff_t end_index = ((unsigned long long)i_size)
>> + >> PAGE_SHIFT;
>> + unsigned offset;
>> + int i;
>> +
>> + for (i = 0; i < cc->cluster_size; i++) {
>> + struct page *page = cc->rpages[i];
>> +
>> + f2fs_bug_on(sbi, !page);
>> +
>> + if (unlikely(f2fs_cp_error(sbi)))
>> + return false;
>> + if (unlikely(is_sbi_flag_set(sbi, SBI_POR_DOING)))
>> + return false;
>> + if (f2fs_is_drop_cache(cc->inode))
>> + return false;
>> + if (f2fs_is_volatile_file(cc->inode))
>> + return false;
>> +
>> + offset = i_size & (PAGE_SIZE - 1);
>> + if ((page->index > end_index) ||
>> + (page->index == end_index && !offset))
>> + return false;
>
> No need to have a special case for when i_size is a multiple of the page size.
> Just replace end_index with 'nr_pages = DIV_ROUND_UP(i_size, PAGE_SIZE)' and
> check for page->index >= nr_pages.

That is copied from f2fs_write_data_page(), let's clean up in a separated patch.

>
>> +out_fail:
>> + /* TODO: revoke partially updated block addresses */
>> + for (i += 1; i < cc->cluster_size; i++) {
>> + if (!cc->rpages[i])
>> + continue;
>> + redirty_page_for_writepage(wbc, cc->rpages[i]);
>> + unlock_page(cc->rpages[i]);
>> + }
>> + return err;
>
> Un-addressed TODO.

Will fix a little later.

>> static void bio_post_read_processing(struct bio_post_read_ctx *ctx)
>> {
>> - /*
>> - * We use different work queues for decryption and for verity because
>> - * verity may require reading metadata pages that need decryption, and
>> - * we shouldn't recurse to the same workqueue.
>> - */
>
> Why is it okay (i.e., no deadlocks) to no longer use different work queues for
> decryption and for verity? See the comment above which is being deleted.

Could you explain more about how deadlock happen? or share me a link address if
you have described that case somewhere?

>
>> + /* TODO: cluster can be compressed due to race with .writepage */
>> +
>
> Another un-addressed TODO.

Will fix a little later.

>
>> +int f2fs_init_post_read_wq(struct f2fs_sb_info *sbi)
>> +{
>> + if (!f2fs_sb_has_encrypt(sbi) &&
>> + !f2fs_sb_has_compression(sbi))
>> + return 0;
>> +
>> + sbi->post_read_wq = alloc_workqueue("f2fs_post_read_wq",
>> + WQ_UNBOUND | WQ_HIGHPRI,
>> + num_online_cpus());
>
> post_read_wq is also needed if verity is enabled.

Yes, we missed this as verity was not merged when implementing this....

Thanks,