while (write_bytes > 0) {
size_t count = min_t(size_t,
while (write_bytes > 0) {
size_t count = min_t(size_t,
err = prepare_uptodate_page(inode, pages[i],
pos + write_bytes, false);
if (err) {
err = prepare_uptodate_page(inode, pages[i],
pos + write_bytes, false);
if (err) {
}
btrfs_start_ordered_extent(inode, ordered, 1);
btrfs_put_ordered_extent(ordered);
}
btrfs_start_ordered_extent(inode, ordered, 1);
btrfs_put_ordered_extent(ordered);
- nrptrs = min(DIV_ROUND_UP(iov_iter_count(i), PAGE_CACHE_SIZE),
- PAGE_CACHE_SIZE / (sizeof(struct page *)));
+ nrptrs = min(DIV_ROUND_UP(iov_iter_count(i), PAGE_SIZE),
+ PAGE_SIZE / (sizeof(struct page *)));
nrptrs = min(nrptrs, current->nr_dirtied_pause - current->nr_dirtied);
nrptrs = max(nrptrs, 8);
pages = kmalloc_array(nrptrs, sizeof(struct page *), GFP_KERNEL);
nrptrs = min(nrptrs, current->nr_dirtied_pause - current->nr_dirtied);
nrptrs = max(nrptrs, 8);
pages = kmalloc_array(nrptrs, sizeof(struct page *), GFP_KERNEL);
* write_bytes, so scale down.
*/
num_pages = DIV_ROUND_UP(write_bytes + offset,
* write_bytes, so scale down.
*/
num_pages = DIV_ROUND_UP(write_bytes + offset,
} else {
force_page_uptodate = false;
dirty_pages = DIV_ROUND_UP(copied + offset,
} else {
force_page_uptodate = false;
dirty_pages = DIV_ROUND_UP(copied + offset,
- invalidate_mapping_pages(file->f_mapping, pos >> PAGE_CACHE_SHIFT,
- endbyte >> PAGE_CACHE_SHIFT);
+ invalidate_mapping_pages(file->f_mapping, pos >> PAGE_SHIFT,
+ endbyte >> PAGE_SHIFT);
*/
int btrfs_sync_file(struct file *file, loff_t start, loff_t end, int datasync)
{
*/
int btrfs_sync_file(struct file *file, loff_t start, loff_t end, int datasync)
{
struct inode *inode = d_inode(dentry);
struct btrfs_root *root = BTRFS_I(inode)->root;
struct btrfs_trans_handle *trans;
struct inode *inode = d_inode(dentry);
struct btrfs_root *root = BTRFS_I(inode)->root;
struct btrfs_trans_handle *trans;