1 #include <linux/ceph/ceph_debug.h>
3 #include <linux/module.h>
4 #include <linux/sched.h>
5 #include <linux/slab.h>
6 #include <linux/file.h>
7 #include <linux/namei.h>
8 #include <linux/writeback.h>
10 #include <linux/ceph/libceph.h>
13 * build a vector of user pages
15 struct page
**ceph_get_direct_page_vector(const void __user
*data
,
16 int num_pages
, bool write_page
)
22 pages
= kmalloc(sizeof(*pages
) * num_pages
, GFP_NOFS
);
24 return ERR_PTR(-ENOMEM
);
26 while (got
< num_pages
) {
27 rc
= get_user_pages_unlocked(
28 (unsigned long)data
+ ((unsigned long)got
* PAGE_SIZE
),
29 num_pages
- got
, pages
+ got
, write_page
? FOLL_WRITE
: 0);
40 ceph_put_page_vector(pages
, got
, false);
43 EXPORT_SYMBOL(ceph_get_direct_page_vector
);
45 void ceph_put_page_vector(struct page
**pages
, int num_pages
, bool dirty
)
49 for (i
= 0; i
< num_pages
; i
++) {
51 set_page_dirty_lock(pages
[i
]);
56 EXPORT_SYMBOL(ceph_put_page_vector
);
58 void ceph_release_page_vector(struct page
**pages
, int num_pages
)
62 for (i
= 0; i
< num_pages
; i
++)
63 __free_pages(pages
[i
], 0);
66 EXPORT_SYMBOL(ceph_release_page_vector
);
69 * allocate a vector new pages
71 struct page
**ceph_alloc_page_vector(int num_pages
, gfp_t flags
)
76 pages
= kmalloc(sizeof(*pages
) * num_pages
, flags
);
78 return ERR_PTR(-ENOMEM
);
79 for (i
= 0; i
< num_pages
; i
++) {
80 pages
[i
] = __page_cache_alloc(flags
);
81 if (pages
[i
] == NULL
) {
82 ceph_release_page_vector(pages
, i
);
83 return ERR_PTR(-ENOMEM
);
88 EXPORT_SYMBOL(ceph_alloc_page_vector
);
91 * copy user data into a page vector
93 int ceph_copy_user_to_page_vector(struct page
**pages
,
94 const void __user
*data
,
95 loff_t off
, size_t len
)
98 int po
= off
& ~PAGE_MASK
;
103 l
= min_t(int, PAGE_SIZE
-po
, left
);
104 bad
= copy_from_user(page_address(pages
[i
]) + po
, data
, l
);
110 if (po
== PAGE_SIZE
) {
117 EXPORT_SYMBOL(ceph_copy_user_to_page_vector
);
119 void ceph_copy_to_page_vector(struct page
**pages
,
121 loff_t off
, size_t len
)
124 size_t po
= off
& ~PAGE_MASK
;
128 size_t l
= min_t(size_t, PAGE_SIZE
-po
, left
);
130 memcpy(page_address(pages
[i
]) + po
, data
, l
);
134 if (po
== PAGE_SIZE
) {
140 EXPORT_SYMBOL(ceph_copy_to_page_vector
);
142 void ceph_copy_from_page_vector(struct page
**pages
,
144 loff_t off
, size_t len
)
147 size_t po
= off
& ~PAGE_MASK
;
151 size_t l
= min_t(size_t, PAGE_SIZE
-po
, left
);
153 memcpy(data
, page_address(pages
[i
]) + po
, l
);
157 if (po
== PAGE_SIZE
) {
163 EXPORT_SYMBOL(ceph_copy_from_page_vector
);
166 * Zero an extent within a page vector. Offset is relative to the
167 * start of the first page.
169 void ceph_zero_page_vector_range(int off
, int len
, struct page
**pages
)
171 int i
= off
>> PAGE_SHIFT
;
175 dout("zero_page_vector_page %u~%u\n", off
, len
);
177 /* leading partial page? */
179 int end
= min((int)PAGE_SIZE
, off
+ len
);
180 dout("zeroing %d %p head from %d\n", i
, pages
[i
],
182 zero_user_segment(pages
[i
], off
, end
);
186 while (len
>= PAGE_SIZE
) {
187 dout("zeroing %d %p len=%d\n", i
, pages
[i
], len
);
188 zero_user_segment(pages
[i
], 0, PAGE_SIZE
);
192 /* trailing partial page? */
194 dout("zeroing %d %p tail to %d\n", i
, pages
[i
], (int)len
);
195 zero_user_segment(pages
[i
], 0, len
);
198 EXPORT_SYMBOL(ceph_zero_page_vector_range
);