]> git.proxmox.com Git - mirror_ubuntu-jammy-kernel.git/blame - fs/xfs/kmem.h
xfs: Replace zero-length array with flexible-array
[mirror_ubuntu-jammy-kernel.git] / fs / xfs / kmem.h
CommitLineData
0b61f8a4 1// SPDX-License-Identifier: GPL-2.0
1da177e4 2/*
7b718769
NS
3 * Copyright (c) 2000-2005 Silicon Graphics, Inc.
4 * All Rights Reserved.
1da177e4
LT
5 */
6#ifndef __XFS_SUPPORT_KMEM_H__
7#define __XFS_SUPPORT_KMEM_H__
8
9#include <linux/slab.h>
10#include <linux/sched.h>
11#include <linux/mm.h>
bdfb0430 12#include <linux/vmalloc.h>
1da177e4 13
8758280f
NS
14/*
15 * General memory allocation interfaces
16 */
17
77ba7877 18typedef unsigned __bitwise xfs_km_flags_t;
77ba7877
AV
19#define KM_NOFS ((__force xfs_km_flags_t)0x0004u)
20#define KM_MAYFAIL ((__force xfs_km_flags_t)0x0008u)
359d992b 21#define KM_ZERO ((__force xfs_km_flags_t)0x0010u)
8758280f
NS
22
23/*
24 * We use a special process flag to avoid recursive callbacks into
25 * the filesystem during transactions. We will also issue our own
26 * warnings, so we explicitly skip any generic ones (silly of us).
27 */
28static inline gfp_t
77ba7877 29kmem_flags_convert(xfs_km_flags_t flags)
1da177e4 30{
8758280f 31 gfp_t lflags;
1da177e4 32
707e0dda 33 BUG_ON(flags & ~(KM_NOFS|KM_MAYFAIL|KM_ZERO));
1da177e4 34
707e0dda
TH
35 lflags = GFP_KERNEL | __GFP_NOWARN;
36 if (flags & KM_NOFS)
37 lflags &= ~__GFP_FS;
359d992b 38
91c63ecd
MH
39 /*
40 * Default page/slab allocator behavior is to retry for ever
41 * for small allocations. We can override this behavior by using
42 * __GFP_RETRY_MAYFAIL which will tell the allocator to retry as long
43 * as it is feasible but rather fail than retry forever for all
44 * request sizes.
45 */
46 if (flags & KM_MAYFAIL)
47 lflags |= __GFP_RETRY_MAYFAIL;
48
359d992b
GZ
49 if (flags & KM_ZERO)
50 lflags |= __GFP_ZERO;
51
8758280f 52 return lflags;
1da177e4
LT
53}
54
77ba7877 55extern void *kmem_alloc(size_t, xfs_km_flags_t);
f8f9ee47 56extern void *kmem_alloc_io(size_t size, int align_mask, xfs_km_flags_t flags);
cb0a8d23 57extern void *kmem_alloc_large(size_t size, xfs_km_flags_t);
664b60f6 58extern void *kmem_realloc(const void *, size_t, xfs_km_flags_t);
f3d21552
WY
59static inline void kmem_free(const void *ptr)
60{
61 kvfree(ptr);
62}
8758280f 63
bdfb0430 64
359d992b
GZ
65static inline void *
66kmem_zalloc(size_t size, xfs_km_flags_t flags)
67{
68 return kmem_alloc(size, flags | KM_ZERO);
cb0a8d23
DC
69}
70
71static inline void *
72kmem_zalloc_large(size_t size, xfs_km_flags_t flags)
73{
74 return kmem_alloc_large(size, flags | KM_ZERO);
359d992b
GZ
75}
76
8758280f
NS
77/*
78 * Zone interfaces
79 */
80
8758280f
NS
81#define kmem_zone kmem_cache
82#define kmem_zone_t struct kmem_cache
83
77ba7877 84extern void *kmem_zone_alloc(kmem_zone_t *, xfs_km_flags_t);
359d992b
GZ
85
86static inline void *
87kmem_zone_zalloc(kmem_zone_t *zone, xfs_km_flags_t flags)
88{
89 return kmem_zone_alloc(zone, flags | KM_ZERO);
90}
1da177e4 91
72945d86
CH
92static inline struct page *
93kmem_to_page(void *addr)
94{
95 if (is_vmalloc_addr(addr))
96 return vmalloc_to_page(addr);
97 return virt_to_page(addr);
98}
99
1da177e4 100#endif /* __XFS_SUPPORT_KMEM_H__ */