]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * JFFS2 -- Journalling Flash File System, Version 2. | |
3 | * | |
4 | * Copyright (C) 2001-2003 Red Hat, Inc. | |
5 | * | |
6 | * Created by David Woodhouse <dwmw2@infradead.org> | |
7 | * | |
8 | * For licensing information, see the file 'LICENCE' in this directory. | |
9 | * | |
10 | * $Id: malloc.c,v 1.28 2004/11/16 20:36:11 dwmw2 Exp $ | |
11 | * | |
12 | */ | |
13 | ||
14 | #include <linux/kernel.h> | |
15 | #include <linux/slab.h> | |
16 | #include <linux/init.h> | |
17 | #include <linux/jffs2.h> | |
18 | #include "nodelist.h" | |
19 | ||
20 | #if 0 | |
21 | #define JFFS2_SLAB_POISON SLAB_POISON | |
22 | #else | |
23 | #define JFFS2_SLAB_POISON 0 | |
24 | #endif | |
25 | ||
26 | // replace this by #define D3 (x) x for cache debugging | |
27 | #define D3(x) | |
28 | ||
29 | /* These are initialised to NULL in the kernel startup code. | |
30 | If you're porting to other operating systems, beware */ | |
31 | static kmem_cache_t *full_dnode_slab; | |
32 | static kmem_cache_t *raw_dirent_slab; | |
33 | static kmem_cache_t *raw_inode_slab; | |
34 | static kmem_cache_t *tmp_dnode_info_slab; | |
35 | static kmem_cache_t *raw_node_ref_slab; | |
36 | static kmem_cache_t *node_frag_slab; | |
37 | static kmem_cache_t *inode_cache_slab; | |
38 | ||
39 | int __init jffs2_create_slab_caches(void) | |
40 | { | |
41 | full_dnode_slab = kmem_cache_create("jffs2_full_dnode", | |
42 | sizeof(struct jffs2_full_dnode), | |
43 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
44 | if (!full_dnode_slab) | |
45 | goto err; | |
46 | ||
47 | raw_dirent_slab = kmem_cache_create("jffs2_raw_dirent", | |
48 | sizeof(struct jffs2_raw_dirent), | |
49 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
50 | if (!raw_dirent_slab) | |
51 | goto err; | |
52 | ||
53 | raw_inode_slab = kmem_cache_create("jffs2_raw_inode", | |
54 | sizeof(struct jffs2_raw_inode), | |
55 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
56 | if (!raw_inode_slab) | |
57 | goto err; | |
58 | ||
59 | tmp_dnode_info_slab = kmem_cache_create("jffs2_tmp_dnode", | |
60 | sizeof(struct jffs2_tmp_dnode_info), | |
61 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
62 | if (!tmp_dnode_info_slab) | |
63 | goto err; | |
64 | ||
65 | raw_node_ref_slab = kmem_cache_create("jffs2_raw_node_ref", | |
66 | sizeof(struct jffs2_raw_node_ref), | |
67 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
68 | if (!raw_node_ref_slab) | |
69 | goto err; | |
70 | ||
71 | node_frag_slab = kmem_cache_create("jffs2_node_frag", | |
72 | sizeof(struct jffs2_node_frag), | |
73 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
74 | if (!node_frag_slab) | |
75 | goto err; | |
76 | ||
77 | inode_cache_slab = kmem_cache_create("jffs2_inode_cache", | |
78 | sizeof(struct jffs2_inode_cache), | |
79 | 0, JFFS2_SLAB_POISON, NULL, NULL); | |
80 | if (inode_cache_slab) | |
81 | return 0; | |
82 | err: | |
83 | jffs2_destroy_slab_caches(); | |
84 | return -ENOMEM; | |
85 | } | |
86 | ||
87 | void jffs2_destroy_slab_caches(void) | |
88 | { | |
89 | if(full_dnode_slab) | |
90 | kmem_cache_destroy(full_dnode_slab); | |
91 | if(raw_dirent_slab) | |
92 | kmem_cache_destroy(raw_dirent_slab); | |
93 | if(raw_inode_slab) | |
94 | kmem_cache_destroy(raw_inode_slab); | |
95 | if(tmp_dnode_info_slab) | |
96 | kmem_cache_destroy(tmp_dnode_info_slab); | |
97 | if(raw_node_ref_slab) | |
98 | kmem_cache_destroy(raw_node_ref_slab); | |
99 | if(node_frag_slab) | |
100 | kmem_cache_destroy(node_frag_slab); | |
101 | if(inode_cache_slab) | |
102 | kmem_cache_destroy(inode_cache_slab); | |
103 | } | |
104 | ||
105 | struct jffs2_full_dirent *jffs2_alloc_full_dirent(int namesize) | |
106 | { | |
107 | return kmalloc(sizeof(struct jffs2_full_dirent) + namesize, GFP_KERNEL); | |
108 | } | |
109 | ||
110 | void jffs2_free_full_dirent(struct jffs2_full_dirent *x) | |
111 | { | |
112 | kfree(x); | |
113 | } | |
114 | ||
115 | struct jffs2_full_dnode *jffs2_alloc_full_dnode(void) | |
116 | { | |
117 | struct jffs2_full_dnode *ret = kmem_cache_alloc(full_dnode_slab, GFP_KERNEL); | |
118 | D3 (printk (KERN_DEBUG "alloc_full_dnode at %p\n", ret)); | |
119 | return ret; | |
120 | } | |
121 | ||
122 | void jffs2_free_full_dnode(struct jffs2_full_dnode *x) | |
123 | { | |
124 | D3 (printk (KERN_DEBUG "free full_dnode at %p\n", x)); | |
125 | kmem_cache_free(full_dnode_slab, x); | |
126 | } | |
127 | ||
128 | struct jffs2_raw_dirent *jffs2_alloc_raw_dirent(void) | |
129 | { | |
130 | struct jffs2_raw_dirent *ret = kmem_cache_alloc(raw_dirent_slab, GFP_KERNEL); | |
131 | D3 (printk (KERN_DEBUG "alloc_raw_dirent\n", ret)); | |
132 | return ret; | |
133 | } | |
134 | ||
135 | void jffs2_free_raw_dirent(struct jffs2_raw_dirent *x) | |
136 | { | |
137 | D3 (printk (KERN_DEBUG "free_raw_dirent at %p\n", x)); | |
138 | kmem_cache_free(raw_dirent_slab, x); | |
139 | } | |
140 | ||
141 | struct jffs2_raw_inode *jffs2_alloc_raw_inode(void) | |
142 | { | |
143 | struct jffs2_raw_inode *ret = kmem_cache_alloc(raw_inode_slab, GFP_KERNEL); | |
144 | D3 (printk (KERN_DEBUG "alloc_raw_inode at %p\n", ret)); | |
145 | return ret; | |
146 | } | |
147 | ||
148 | void jffs2_free_raw_inode(struct jffs2_raw_inode *x) | |
149 | { | |
150 | D3 (printk (KERN_DEBUG "free_raw_inode at %p\n", x)); | |
151 | kmem_cache_free(raw_inode_slab, x); | |
152 | } | |
153 | ||
154 | struct jffs2_tmp_dnode_info *jffs2_alloc_tmp_dnode_info(void) | |
155 | { | |
156 | struct jffs2_tmp_dnode_info *ret = kmem_cache_alloc(tmp_dnode_info_slab, GFP_KERNEL); | |
157 | D3 (printk (KERN_DEBUG "alloc_tmp_dnode_info at %p\n", ret)); | |
158 | return ret; | |
159 | } | |
160 | ||
161 | void jffs2_free_tmp_dnode_info(struct jffs2_tmp_dnode_info *x) | |
162 | { | |
163 | D3 (printk (KERN_DEBUG "free_tmp_dnode_info at %p\n", x)); | |
164 | kmem_cache_free(tmp_dnode_info_slab, x); | |
165 | } | |
166 | ||
167 | struct jffs2_raw_node_ref *jffs2_alloc_raw_node_ref(void) | |
168 | { | |
169 | struct jffs2_raw_node_ref *ret = kmem_cache_alloc(raw_node_ref_slab, GFP_KERNEL); | |
170 | D3 (printk (KERN_DEBUG "alloc_raw_node_ref at %p\n", ret)); | |
171 | return ret; | |
172 | } | |
173 | ||
174 | void jffs2_free_raw_node_ref(struct jffs2_raw_node_ref *x) | |
175 | { | |
176 | D3 (printk (KERN_DEBUG "free_raw_node_ref at %p\n", x)); | |
177 | kmem_cache_free(raw_node_ref_slab, x); | |
178 | } | |
179 | ||
180 | struct jffs2_node_frag *jffs2_alloc_node_frag(void) | |
181 | { | |
182 | struct jffs2_node_frag *ret = kmem_cache_alloc(node_frag_slab, GFP_KERNEL); | |
183 | D3 (printk (KERN_DEBUG "alloc_node_frag at %p\n", ret)); | |
184 | return ret; | |
185 | } | |
186 | ||
187 | void jffs2_free_node_frag(struct jffs2_node_frag *x) | |
188 | { | |
189 | D3 (printk (KERN_DEBUG "free_node_frag at %p\n", x)); | |
190 | kmem_cache_free(node_frag_slab, x); | |
191 | } | |
192 | ||
193 | struct jffs2_inode_cache *jffs2_alloc_inode_cache(void) | |
194 | { | |
195 | struct jffs2_inode_cache *ret = kmem_cache_alloc(inode_cache_slab, GFP_KERNEL); | |
196 | D3 (printk(KERN_DEBUG "Allocated inocache at %p\n", ret)); | |
197 | return ret; | |
198 | } | |
199 | ||
200 | void jffs2_free_inode_cache(struct jffs2_inode_cache *x) | |
201 | { | |
202 | D3 (printk(KERN_DEBUG "Freeing inocache at %p\n", x)); | |
203 | kmem_cache_free(inode_cache_slab, x); | |
204 | } | |
205 |