]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/blame - ubuntu/vbox/vboxsf/regops.c
UBUNTU: ubuntu: vbox -- update to 5.2.0-dfsg-2
[mirror_ubuntu-bionic-kernel.git] / ubuntu / vbox / vboxsf / regops.c
CommitLineData
056a1eb7
SF
1/* $Id: regops.c $ */
2/** @file
3 * vboxsf - VBox Linux Shared Folders, Regular file inode and file operations.
4 */
5
6/*
7 * Copyright (C) 2006-2016 Oracle Corporation
8 *
9 * This file is part of VirtualBox Open Source Edition (OSE), as
10 * available from http://www.virtualbox.org. This file is free software;
11 * you can redistribute it and/or modify it under the terms of the GNU
12 * General Public License (GPL) as published by the Free Software
13 * Foundation, in version 2 as it comes in the "COPYING" file of the
14 * VirtualBox OSE distribution. VirtualBox OSE is distributed in the
15 * hope that it will be useful, but WITHOUT ANY WARRANTY of any kind.
16 */
17
18/*
19 * Limitations: only COW memory mapping is supported
20 */
21
22#include "vfsmod.h"
23
24static void *alloc_bounce_buffer(size_t *tmp_sizep, PRTCCPHYS physp, size_t
25 xfer_size, const char *caller)
26{
27 size_t tmp_size;
28 void *tmp;
29
30 /* try for big first. */
31 tmp_size = RT_ALIGN_Z(xfer_size, PAGE_SIZE);
32 if (tmp_size > 16U*_1K)
33 tmp_size = 16U*_1K;
34 tmp = kmalloc(tmp_size, GFP_KERNEL);
35 if (!tmp)
36 {
37 /* fall back on a page sized buffer. */
38 tmp = kmalloc(PAGE_SIZE, GFP_KERNEL);
39 if (!tmp)
40 {
41 LogRel(("%s: could not allocate bounce buffer for xfer_size=%zu %s\n", caller, xfer_size));
42 return NULL;
43 }
44 tmp_size = PAGE_SIZE;
45 }
46
47 *tmp_sizep = tmp_size;
48 *physp = virt_to_phys(tmp);
49 return tmp;
50}
51
52static void free_bounce_buffer(void *tmp)
53{
54 kfree (tmp);
55}
56
57
58/* fops */
59static int sf_reg_read_aux(const char *caller, struct sf_glob_info *sf_g,
60 struct sf_reg_info *sf_r, void *buf,
61 uint32_t *nread, uint64_t pos)
62{
63 /** @todo bird: yes, kmap() and kmalloc() input only. Since the buffer is
64 * contiguous in physical memory (kmalloc or single page), we should
65 * use a physical address here to speed things up. */
66 int rc = VbglR0SfRead(&client_handle, &sf_g->map, sf_r->handle,
67 pos, nread, buf, false /* already locked? */);
68 if (RT_FAILURE(rc))
69 {
70 LogFunc(("VbglR0SfRead failed. caller=%s, rc=%Rrc\n", caller, rc));
71 return -EPROTO;
72 }
73 return 0;
74}
75
76static int sf_reg_write_aux(const char *caller, struct sf_glob_info *sf_g,
77 struct sf_reg_info *sf_r, void *buf,
78 uint32_t *nwritten, uint64_t pos)
79{
80 /** @todo bird: yes, kmap() and kmalloc() input only. Since the buffer is
81 * contiguous in physical memory (kmalloc or single page), we should
82 * use a physical address here to speed things up. */
83 int rc = VbglR0SfWrite(&client_handle, &sf_g->map, sf_r->handle,
84 pos, nwritten, buf, false /* already locked? */);
85 if (RT_FAILURE(rc))
86 {
87 LogFunc(("VbglR0SfWrite failed. caller=%s, rc=%Rrc\n",
88 caller, rc));
89 return -EPROTO;
90 }
91 return 0;
92}
93
94/**
95 * Read from a regular file.
96 *
97 * @param file the file
98 * @param buf the buffer
99 * @param size length of the buffer
100 * @param off offset within the file
101 * @returns the number of read bytes on success, Linux error code otherwise
102 */
103static ssize_t sf_reg_read(struct file *file, char *buf, size_t size, loff_t *off)
104{
105 int err;
106 void *tmp;
107 RTCCPHYS tmp_phys;
108 size_t tmp_size;
109 size_t left = size;
110 ssize_t total_bytes_read = 0;
111 struct inode *inode = GET_F_DENTRY(file)->d_inode;
112 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
113 struct sf_reg_info *sf_r = file->private_data;
114 loff_t pos = *off;
115
116 TRACE();
117 if (!S_ISREG(inode->i_mode))
118 {
119 LogFunc(("read from non regular file %d\n", inode->i_mode));
120 return -EINVAL;
121 }
122
123 /** XXX Check read permission according to inode->i_mode! */
124
125 if (!size)
126 return 0;
127
128 tmp = alloc_bounce_buffer(&tmp_size, &tmp_phys, size, __PRETTY_FUNCTION__);
129 if (!tmp)
130 return -ENOMEM;
131
132 while (left)
133 {
134 uint32_t to_read, nread;
135
136 to_read = tmp_size;
137 if (to_read > left)
138 to_read = (uint32_t) left;
139
140 nread = to_read;
141
142 err = sf_reg_read_aux(__func__, sf_g, sf_r, tmp, &nread, pos);
143 if (err)
144 goto fail;
145
146 if (copy_to_user(buf, tmp, nread))
147 {
148 err = -EFAULT;
149 goto fail;
150 }
151
152 pos += nread;
153 left -= nread;
154 buf += nread;
155 total_bytes_read += nread;
156 if (nread != to_read)
157 break;
158 }
159
160 *off += total_bytes_read;
161 free_bounce_buffer(tmp);
162 return total_bytes_read;
163
164fail:
165 free_bounce_buffer(tmp);
166 return err;
167}
168
169/**
170 * Write to a regular file.
171 *
172 * @param file the file
173 * @param buf the buffer
174 * @param size length of the buffer
175 * @param off offset within the file
176 * @returns the number of written bytes on success, Linux error code otherwise
177 */
178static ssize_t sf_reg_write(struct file *file, const char *buf, size_t size, loff_t *off)
179{
180 int err;
181 void *tmp;
182 RTCCPHYS tmp_phys;
183 size_t tmp_size;
184 size_t left = size;
185 ssize_t total_bytes_written = 0;
186 struct inode *inode = GET_F_DENTRY(file)->d_inode;
187 struct sf_inode_info *sf_i = GET_INODE_INFO(inode);
188 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
189 struct sf_reg_info *sf_r = file->private_data;
190 loff_t pos;
191
192 TRACE();
193 BUG_ON(!sf_i);
194 BUG_ON(!sf_g);
195 BUG_ON(!sf_r);
196
197 if (!S_ISREG(inode->i_mode))
198 {
199 LogFunc(("write to non regular file %d\n", inode->i_mode));
200 return -EINVAL;
201 }
202
203 pos = *off;
204 if (file->f_flags & O_APPEND)
205 {
206 pos = inode->i_size;
207 *off = pos;
208 }
209
210 /** XXX Check write permission according to inode->i_mode! */
211
212 if (!size)
213 return 0;
214
215 tmp = alloc_bounce_buffer(&tmp_size, &tmp_phys, size, __PRETTY_FUNCTION__);
216 if (!tmp)
217 return -ENOMEM;
218
219 while (left)
220 {
221 uint32_t to_write, nwritten;
222
223 to_write = tmp_size;
224 if (to_write > left)
225 to_write = (uint32_t) left;
226
227 nwritten = to_write;
228
229 if (copy_from_user(tmp, buf, to_write))
230 {
231 err = -EFAULT;
232 goto fail;
233 }
234
6d209b23
SF
235 err = VbglR0SfWritePhysCont(&client_handle, &sf_g->map, sf_r->handle,
236 pos, &nwritten, tmp_phys);
056a1eb7
SF
237 if (err)
238 goto fail;
239
240 pos += nwritten;
241 left -= nwritten;
242 buf += nwritten;
243 total_bytes_written += nwritten;
244 if (nwritten != to_write)
245 break;
246 }
247
248 *off += total_bytes_written;
249 if (*off > inode->i_size)
250 inode->i_size = *off;
251
252 sf_i->force_restat = 1;
253 free_bounce_buffer(tmp);
254 return total_bytes_written;
255
256fail:
257 free_bounce_buffer(tmp);
258 return err;
259}
260
261/**
262 * Open a regular file.
263 *
264 * @param inode the inode
265 * @param file the file
266 * @returns 0 on success, Linux error code otherwise
267 */
268static int sf_reg_open(struct inode *inode, struct file *file)
269{
270 int rc, rc_linux = 0;
271 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
272 struct sf_inode_info *sf_i = GET_INODE_INFO(inode);
273 struct sf_reg_info *sf_r;
274 SHFLCREATEPARMS params;
275
276 TRACE();
277 BUG_ON(!sf_g);
278 BUG_ON(!sf_i);
279
280 LogFunc(("open %s\n", sf_i->path->String.utf8));
281
282 sf_r = kmalloc(sizeof(*sf_r), GFP_KERNEL);
283 if (!sf_r)
284 {
285 LogRelFunc(("could not allocate reg info\n"));
286 return -ENOMEM;
287 }
288
289 /* Already open? */
290 if (sf_i->handle != SHFL_HANDLE_NIL)
291 {
292 /*
293 * This inode was created with sf_create_aux(). Check the CreateFlags:
294 * O_CREAT, O_TRUNC: inherent true (file was just created). Not sure
295 * about the access flags (SHFL_CF_ACCESS_*).
296 */
297 sf_i->force_restat = 1;
298 sf_r->handle = sf_i->handle;
299 sf_i->handle = SHFL_HANDLE_NIL;
300 sf_i->file = file;
301 file->private_data = sf_r;
302 return 0;
303 }
304
305 RT_ZERO(params);
306 params.Handle = SHFL_HANDLE_NIL;
307 /* We check the value of params.Handle afterwards to find out if
308 * the call succeeded or failed, as the API does not seem to cleanly
309 * distinguish error and informational messages.
310 *
311 * Furthermore, we must set params.Handle to SHFL_HANDLE_NIL to
312 * make the shared folders host service use our fMode parameter */
313
314 if (file->f_flags & O_CREAT)
315 {
316 LogFunc(("O_CREAT set\n"));
317 params.CreateFlags |= SHFL_CF_ACT_CREATE_IF_NEW;
318 /* We ignore O_EXCL, as the Linux kernel seems to call create
319 beforehand itself, so O_EXCL should always fail. */
320 if (file->f_flags & O_TRUNC)
321 {
322 LogFunc(("O_TRUNC set\n"));
323 params.CreateFlags |= SHFL_CF_ACT_OVERWRITE_IF_EXISTS;
324 }
325 else
326 params.CreateFlags |= SHFL_CF_ACT_OPEN_IF_EXISTS;
327 }
328 else
329 {
330 params.CreateFlags |= SHFL_CF_ACT_FAIL_IF_NEW;
331 if (file->f_flags & O_TRUNC)
332 {
333 LogFunc(("O_TRUNC set\n"));
334 params.CreateFlags |= SHFL_CF_ACT_OVERWRITE_IF_EXISTS;
335 }
336 }
337
338 switch (file->f_flags & O_ACCMODE)
339 {
340 case O_RDONLY:
341 params.CreateFlags |= SHFL_CF_ACCESS_READ;
342 break;
343
344 case O_WRONLY:
345 params.CreateFlags |= SHFL_CF_ACCESS_WRITE;
346 break;
347
348 case O_RDWR:
349 params.CreateFlags |= SHFL_CF_ACCESS_READWRITE;
350 break;
351
352 default:
353 BUG ();
354 }
355
356 if (file->f_flags & O_APPEND)
357 {
358 LogFunc(("O_APPEND set\n"));
359 params.CreateFlags |= SHFL_CF_ACCESS_APPEND;
360 }
361
362 params.Info.Attr.fMode = inode->i_mode;
363 LogFunc(("sf_reg_open: calling VbglR0SfCreate, file %s, flags=%#x, %#x\n",
364 sf_i->path->String.utf8 , file->f_flags, params.CreateFlags));
365 rc = VbglR0SfCreate(&client_handle, &sf_g->map, sf_i->path, &params);
366 if (RT_FAILURE(rc))
367 {
368 LogFunc(("VbglR0SfCreate failed flags=%d,%#x rc=%Rrc\n",
369 file->f_flags, params.CreateFlags, rc));
370 kfree(sf_r);
371 return -RTErrConvertToErrno(rc);
372 }
373
374 if (SHFL_HANDLE_NIL == params.Handle)
375 {
376 switch (params.Result)
377 {
378 case SHFL_PATH_NOT_FOUND:
379 case SHFL_FILE_NOT_FOUND:
380 rc_linux = -ENOENT;
381 break;
382 case SHFL_FILE_EXISTS:
383 rc_linux = -EEXIST;
384 break;
385 default:
386 break;
387 }
388 }
389
390 sf_i->force_restat = 1;
391 sf_r->handle = params.Handle;
392 sf_i->file = file;
393 file->private_data = sf_r;
394 return rc_linux;
395}
396
397/**
398 * Close a regular file.
399 *
400 * @param inode the inode
401 * @param file the file
402 * @returns 0 on success, Linux error code otherwise
403 */
404static int sf_reg_release(struct inode *inode, struct file *file)
405{
406 int rc;
407 struct sf_reg_info *sf_r;
408 struct sf_glob_info *sf_g;
409 struct sf_inode_info *sf_i = GET_INODE_INFO(inode);
410
411 TRACE();
412 sf_g = GET_GLOB_INFO(inode->i_sb);
413 sf_r = file->private_data;
414
415 BUG_ON(!sf_g);
416 BUG_ON(!sf_r);
417
418#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 4, 25)
419 /* See the smbfs source (file.c). mmap in particular can cause data to be
420 * written to the file after it is closed, which we can't cope with. We
421 * copy and paste the body of filemap_write_and_wait() here as it was not
422 * defined before 2.6.6 and not exported until quite a bit later. */
423 /* filemap_write_and_wait(inode->i_mapping); */
424 if ( inode->i_mapping->nrpages
425 && filemap_fdatawrite(inode->i_mapping) != -EIO)
426 filemap_fdatawait(inode->i_mapping);
427#endif
428 rc = VbglR0SfClose(&client_handle, &sf_g->map, sf_r->handle);
429 if (RT_FAILURE(rc))
430 LogFunc(("VbglR0SfClose failed rc=%Rrc\n", rc));
431
432 kfree(sf_r);
433 sf_i->file = NULL;
434 sf_i->handle = SHFL_HANDLE_NIL;
435 file->private_data = NULL;
436 return 0;
437}
438
439#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0)
440static int sf_reg_fault(struct vm_fault *vmf)
441#elif LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
442static int sf_reg_fault(struct vm_area_struct *vma, struct vm_fault *vmf)
443#elif LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 0)
444static struct page *sf_reg_nopage(struct vm_area_struct *vma, unsigned long vaddr, int *type)
445# define SET_TYPE(t) *type = (t)
446#else /* LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 0) */
447static struct page *sf_reg_nopage(struct vm_area_struct *vma, unsigned long vaddr, int unused)
448# define SET_TYPE(t)
449#endif
450{
451 struct page *page;
452 char *buf;
453 loff_t off;
454 uint32_t nread = PAGE_SIZE;
455 int err;
456#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 11, 0)
457 struct vm_area_struct *vma = vmf->vma;
458#endif
459 struct file *file = vma->vm_file;
460 struct inode *inode = GET_F_DENTRY(file)->d_inode;
461 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
462 struct sf_reg_info *sf_r = file->private_data;
463
464 TRACE();
465#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
466 if (vmf->pgoff > vma->vm_end)
467 return VM_FAULT_SIGBUS;
468#else
469 if (vaddr > vma->vm_end)
470 {
471 SET_TYPE(VM_FAULT_SIGBUS);
472 return NOPAGE_SIGBUS;
473 }
474#endif
475
476 /* Don't use GFP_HIGHUSER as long as sf_reg_read_aux() calls VbglR0SfRead()
477 * which works on virtual addresses. On Linux cannot reliably determine the
478 * physical address for high memory, see rtR0MemObjNativeLockKernel(). */
479 page = alloc_page(GFP_USER);
480 if (!page) {
481 LogRelFunc(("failed to allocate page\n"));
482#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
483 return VM_FAULT_OOM;
484#else
485 SET_TYPE(VM_FAULT_OOM);
486 return NOPAGE_OOM;
487#endif
488 }
489
490 buf = kmap(page);
491#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
492 off = (vmf->pgoff << PAGE_SHIFT);
493#else
494 off = (vaddr - vma->vm_start) + (vma->vm_pgoff << PAGE_SHIFT);
495#endif
496 err = sf_reg_read_aux(__func__, sf_g, sf_r, buf, &nread, off);
497 if (err)
498 {
499 kunmap(page);
500 put_page(page);
501#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
502 return VM_FAULT_SIGBUS;
503#else
504 SET_TYPE(VM_FAULT_SIGBUS);
505 return NOPAGE_SIGBUS;
506#endif
507 }
508
509 BUG_ON (nread > PAGE_SIZE);
510 if (!nread)
511 {
512#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
513 clear_user_page(page_address(page), vmf->pgoff, page);
514#elif LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 0)
515 clear_user_page(page_address(page), vaddr, page);
516#else
517 clear_user_page(page_address(page), vaddr);
518#endif
519 }
520 else
521 memset(buf + nread, 0, PAGE_SIZE - nread);
522
523 flush_dcache_page(page);
524 kunmap(page);
525#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
526 vmf->page = page;
527 return 0;
528#else
529 SET_TYPE(VM_FAULT_MAJOR);
530 return page;
531#endif
532}
533
534static struct vm_operations_struct sf_vma_ops =
535{
536#if LINUX_VERSION_CODE > KERNEL_VERSION(2, 6, 25)
537 .fault = sf_reg_fault
538#else
539 .nopage = sf_reg_nopage
540#endif
541};
542
543static int sf_reg_mmap(struct file *file, struct vm_area_struct *vma)
544{
545 TRACE();
546 if (vma->vm_flags & VM_SHARED)
547 {
548 LogFunc(("shared mmapping not available\n"));
549 return -EINVAL;
550 }
551
552 vma->vm_ops = &sf_vma_ops;
553 return 0;
554}
555
556struct file_operations sf_reg_fops =
557{
558 .read = sf_reg_read,
559 .open = sf_reg_open,
560 .write = sf_reg_write,
561 .release = sf_reg_release,
562 .mmap = sf_reg_mmap,
563#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 0)
564# if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 23)
565 .splice_read = generic_file_splice_read,
566# else
567 .sendfile = generic_file_sendfile,
568# endif
569# if LINUX_VERSION_CODE >= KERNEL_VERSION(3, 16, 0)
570 .read_iter = generic_file_read_iter,
571 .write_iter = generic_file_write_iter,
572# else
573 .aio_read = generic_file_aio_read,
574 .aio_write = generic_file_aio_write,
575# endif
576# if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 35)
577 .fsync = noop_fsync,
578# else
579 .fsync = simple_sync_file,
580# endif
581 .llseek = generic_file_llseek,
582#endif
583};
584
585
586struct inode_operations sf_reg_iops =
587{
588#if LINUX_VERSION_CODE < KERNEL_VERSION(2, 6, 0)
589 .revalidate = sf_inode_revalidate
590#else
591 .getattr = sf_getattr,
592 .setattr = sf_setattr
593#endif
594};
595
596
597#if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 0)
598static int sf_readpage(struct file *file, struct page *page)
599{
600 struct inode *inode = GET_F_DENTRY(file)->d_inode;
601 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
602 struct sf_reg_info *sf_r = file->private_data;
603 uint32_t nread = PAGE_SIZE;
604 char *buf;
605 loff_t off = ((loff_t)page->index) << PAGE_SHIFT;
606 int ret;
607
608 TRACE();
609
610 buf = kmap(page);
611 ret = sf_reg_read_aux(__func__, sf_g, sf_r, buf, &nread, off);
612 if (ret)
613 {
614 kunmap(page);
615 if (PageLocked(page))
616 unlock_page(page);
617 return ret;
618 }
619 BUG_ON(nread > PAGE_SIZE);
620 memset(&buf[nread], 0, PAGE_SIZE - nread);
621 flush_dcache_page(page);
622 kunmap(page);
623 SetPageUptodate(page);
624 unlock_page(page);
625 return 0;
626}
627
628static int
629sf_writepage(struct page *page, struct writeback_control *wbc)
630{
631 struct address_space *mapping = page->mapping;
632 struct inode *inode = mapping->host;
633 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
634 struct sf_inode_info *sf_i = GET_INODE_INFO(inode);
635 struct file *file = sf_i->file;
636 struct sf_reg_info *sf_r = file->private_data;
637 char *buf;
638 uint32_t nwritten = PAGE_SIZE;
639 int end_index = inode->i_size >> PAGE_SHIFT;
640 loff_t off = ((loff_t) page->index) << PAGE_SHIFT;
641 int err;
642
643 TRACE();
644
645 if (page->index >= end_index)
646 nwritten = inode->i_size & (PAGE_SIZE-1);
647
648 buf = kmap(page);
649
650 err = sf_reg_write_aux(__func__, sf_g, sf_r, buf, &nwritten, off);
651 if (err < 0)
652 {
653 ClearPageUptodate(page);
654 goto out;
655 }
656
657 if (off > inode->i_size)
658 inode->i_size = off;
659
660 if (PageError(page))
661 ClearPageError(page);
662 err = 0;
663
664out:
665 kunmap(page);
666
667 unlock_page(page);
668 return err;
669}
670
671# if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 24)
672int sf_write_begin(struct file *file, struct address_space *mapping, loff_t pos,
673 unsigned len, unsigned flags, struct page **pagep, void **fsdata)
674{
675 TRACE();
676
677 return simple_write_begin(file, mapping, pos, len, flags, pagep, fsdata);
678}
679
680int sf_write_end(struct file *file, struct address_space *mapping, loff_t pos,
681 unsigned len, unsigned copied, struct page *page, void *fsdata)
682{
683 struct inode *inode = mapping->host;
684 struct sf_glob_info *sf_g = GET_GLOB_INFO(inode->i_sb);
685 struct sf_reg_info *sf_r = file->private_data;
686 void *buf;
687 unsigned from = pos & (PAGE_SIZE - 1);
688 uint32_t nwritten = len;
689 int err;
690
691 TRACE();
692
693 buf = kmap(page);
694 err = sf_reg_write_aux(__func__, sf_g, sf_r, buf+from, &nwritten, pos);
695 kunmap(page);
696
697 if (!PageUptodate(page) && err == PAGE_SIZE)
698 SetPageUptodate(page);
699
700 if (err >= 0) {
701 pos += nwritten;
702 if (pos > inode->i_size)
703 inode->i_size = pos;
704 }
705
706 unlock_page(page);
707#if LINUX_VERSION_CODE >= KERNEL_VERSION(4, 6, 0)
708 put_page(page);
709#else
710 page_cache_release(page);
711#endif
712
713 return nwritten;
714}
715
716# endif /* KERNEL_VERSION >= 2.6.24 */
717
718struct address_space_operations sf_reg_aops =
719{
720 .readpage = sf_readpage,
721 .writepage = sf_writepage,
722# if LINUX_VERSION_CODE >= KERNEL_VERSION(2, 6, 24)
723 .write_begin = sf_write_begin,
724 .write_end = sf_write_end,
725# else
726 .prepare_write = simple_prepare_write,
727 .commit_write = simple_commit_write,
728# endif
729};
730#endif