Lines Matching refs:new_size
235 u64 new_size, struct runs_tree *run, in attr_make_nonresident() argument
360 u64 new_size, struct runs_tree *run, in attr_set_size_res() argument
371 s64 dsize = ALIGN(new_size, 8) - ALIGN(rsize, 8); in attr_set_size_res()
377 return attr_make_nonresident(ni, attr, le, mi, new_size, in attr_set_size_res()
384 if (new_size > rsize) in attr_set_size_res()
386 new_size - rsize); in attr_set_size_res()
390 attr->res.data_size = cpu_to_le32(new_size); in attr_set_size_res()
408 u64 new_size, const u64 *new_valid, bool keep_prealloc, in attr_set_size() argument
436 err = attr_set_size_res(ni, attr_b, le_b, mi_b, new_size, run, in attr_set_size()
458 new_alloc = (new_size + align - 1) & ~(u64)(align - 1); in attr_set_size()
461 if (keep_prealloc && new_size < old_size) { in attr_set_size()
462 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
500 if (new_size > old_size) { in attr_set_size()
505 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
524 sbi, get_pre_allocated(new_size)); in attr_set_size()
592 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
599 new_size = new_alloc_tmp; in attr_set_size()
669 if (new_size != old_size || in attr_set_size()
712 attr_b->nres.data_size = cpu_to_le64(new_size); in attr_set_size()
713 if (new_size < old_valid) in attr_set_size()
767 __le64 valid = cpu_to_le64(min(*new_valid, new_size)); in attr_set_size()
784 if (ni->vfs_inode.i_size != new_size) { in attr_set_size()
785 ni->vfs_inode.i_size = new_size; in attr_set_size()