1 /* $NetBSD: pool.h,v 1.54 2006/08/20 09:35:25 yamt Exp $ */
4 * Copyright (c) 1997, 1998, 1999, 2000 The NetBSD Foundation, Inc.
7 * This code is derived from software contributed to The NetBSD Foundation
8 * by Paul Kranenburg; by Jason R. Thorpe of the Numerical Aerospace
9 * Simulation Facility, NASA Ames Research Center.
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
14 * 1. Redistributions of source code must retain the above copyright
15 * notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 * notice, this list of conditions and the following disclaimer in the
18 * documentation and/or other materials provided with the distribution.
19 * 3. All advertising materials mentioning features or use of this software
20 * must display the following acknowledgement:
21 * This product includes software developed by the NetBSD
22 * Foundation, Inc. and its contributors.
23 * 4. Neither the name of The NetBSD Foundation nor the names of its
24 * contributors may be used to endorse or promote products derived
25 * from this software without specific prior written permission.
27 * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS
28 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED
29 * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
30 * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS
31 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
32 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
33 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
34 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
35 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
36 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
37 * POSSIBILITY OF SUCH DAMAGE.
47 #if defined(_KERNEL_OPT)
53 #include <sys/queue.h>
57 #include <sys/callback.h>
58 #endif /* defined(_KERNEL) */
61 #define PCG_NOBJECTS 16
63 #define POOL_PADDR_INVALID ((paddr_t) -1)
66 /* The pool cache group. */
67 struct pool_cache_group
{
68 LIST_ENTRY(pool_cache_group
)
69 pcg_list
; /* link in the pool cache's group list */
70 u_int pcg_avail
; /* # available objects */
71 /* pointers to the objects */
73 void *pcgo_va
; /* cache object virtual address */
74 paddr_t pcgo_pa
;/* cache object physical address */
75 } pcg_objects
[PCG_NOBJECTS
];
78 LIST_HEAD(pool_cache_grouplist
,pool_cache_group
);
80 LIST_ENTRY(pool_cache
)
81 pc_poollist
; /* entry on pool's group list */
82 struct pool_cache_grouplist
83 pc_emptygroups
; /* list of empty cache groups */
84 struct pool_cache_grouplist
85 pc_fullgroups
; /* list of full cache groups */
86 struct pool_cache_grouplist
87 pc_partgroups
; /* list of partial cache groups */
88 struct pool
*pc_pool
; /* parent pool */
89 struct simplelock pc_slock
; /* mutex */
91 int (*pc_ctor
)(void *, void *, int);
92 void (*pc_dtor
)(void *, void *);
96 unsigned long pc_hits
; /* cache hits */
97 unsigned long pc_misses
; /* cache misses */
99 unsigned long pc_ngroups
; /* # cache groups */
101 unsigned long pc_nitems
; /* # objects currently in cache */
104 struct pool_allocator
{
105 void *(*pa_alloc
)(struct pool
*, int);
106 void (*pa_free
)(struct pool
*, void *);
107 unsigned int pa_pagesz
;
109 /* The following fields are for internal use only. */
110 struct simplelock pa_slock
;
111 TAILQ_HEAD(, pool
) pa_list
; /* list of pools using this allocator */
113 #define PA_INITIALIZED 0x01
116 struct vm_map
*pa_backingmap
;
118 struct vm_map
**pa_backingmapptr
;
119 SLIST_ENTRY(pool_allocator
) pa_q
;
120 #endif /* defined(_KERNEL) */
123 LIST_HEAD(pool_pagelist
,pool_item_header
);
129 pr_emptypages
; /* Empty pages */
131 pr_fullpages
; /* Full pages */
133 pr_partpages
; /* Partially-allocated pages */
134 struct pool_item_header
*pr_curpage
;
135 struct pool
*pr_phpool
; /* Pool item header pool */
136 LIST_HEAD(,pool_cache
)
137 pr_cachelist
; /* Caches for this pool */
138 unsigned int pr_size
; /* Size of item */
139 unsigned int pr_align
; /* Requested alignment, must be 2^n */
140 unsigned int pr_itemoffset
; /* Align this offset in item */
141 unsigned int pr_minitems
; /* minimum # of items to keep */
142 unsigned int pr_minpages
; /* same in page units */
143 unsigned int pr_maxpages
; /* maximum # of pages to keep */
144 unsigned int pr_npages
; /* # of pages allocated */
145 unsigned int pr_itemsperpage
;/* # items that fit in a page */
146 unsigned int pr_slack
; /* unused space in a page */
147 unsigned int pr_nitems
; /* number of available items in pool */
148 unsigned int pr_nout
; /* # items currently allocated */
149 unsigned int pr_hardlimit
; /* hard limit to number of allocated
151 struct pool_allocator
*pr_alloc
;/* back-end allocator */
152 TAILQ_ENTRY(pool
) pr_alloc_list
;/* link on allocator's pool list */
155 void (*pr_drain_hook
)(void *, int);
156 void *pr_drain_hook_arg
;
158 const char *pr_wchan
; /* tsleep(9) identifier */
159 unsigned int pr_flags
; /* r/w flags */
160 unsigned int pr_roflags
; /* r/o flags */
161 #define PR_NOWAIT 0x00 /* for symmetry */
162 #define PR_WAITOK 0x02
163 #define PR_WANTED 0x04
164 #define PR_PHINPAGE 0x40
165 #define PR_LOGGING 0x80
166 #define PR_LIMITFAIL 0x100 /* even if waiting, fail if we hit limit */
167 #define PR_RECURSIVE 0x200 /* pool contains pools, for vmstat(8) */
168 #define PR_NOTOUCH 0x400 /* don't use free items to keep internal state*/
169 #define PR_NOALIGN 0x800 /* don't assume backend alignment */
172 * `pr_slock' protects the pool's data structures when removing
173 * items from or returning items to the pool, or when reading
174 * or updating read/write fields in the pool descriptor.
176 * We assume back-end page allocators provide their own locking
177 * scheme. They will be called with the pool descriptor _unlocked_,
178 * since the page allocators may block.
180 struct simplelock pr_slock
;
182 SPLAY_HEAD(phtree
, pool_item_header
) pr_phtree
;
184 int pr_maxcolor
; /* Cache colouring */
186 int pr_phoffset
; /* Offset in page of page header */
189 * Warning message to be issued, and a per-time-delta rate cap,
190 * if the hard limit is reached.
192 const char *pr_hardlimit_warning
;
193 struct timeval pr_hardlimit_ratecap
;
194 struct timeval pr_hardlimit_warning_last
;
199 unsigned long pr_nget
; /* # of successful requests */
200 unsigned long pr_nfail
; /* # of unsuccessful requests */
201 unsigned long pr_nput
; /* # of releases */
202 unsigned long pr_npagealloc
; /* # of pages allocated */
203 unsigned long pr_npagefree
; /* # of pages released */
204 unsigned int pr_hiwat
; /* max # of pages in pool */
205 unsigned long pr_nidle
; /* # of idle pages */
210 struct pool_log
*pr_log
;
214 const char *pr_entered_file
; /* reentrancy check */
215 long pr_entered_line
;
218 struct callback_entry pr_reclaimerentry
;
221 #endif /* __POOL_EXPOSE */
225 * pool_allocator_kmem is the default that all pools get unless
226 * otherwise specified. pool_allocator_nointr is provided for
227 * pools that know they will never be accessed in interrupt
230 extern struct pool_allocator pool_allocator_kmem
;
231 extern struct pool_allocator pool_allocator_nointr
;
233 /* The above are subpage allocators in this case. */
234 extern struct pool_allocator pool_allocator_kmem_fullpage
;
235 extern struct pool_allocator pool_allocator_nointr_fullpage
;
238 struct link_pool_init
{ /* same as args to pool_init() */
245 struct pool_allocator
*palloc
;
247 #define POOL_INIT(pp, size, align, align_offset, flags, wchan, palloc) \
249 static const struct link_pool_init _link_ ## pp[1] = { \
250 { &pp, size, align, align_offset, flags, wchan, palloc } \
252 __link_set_add_rodata(pools, _link_ ## pp)
254 void pool_subsystem_init(void);
256 void pool_init(struct pool
*, size_t, u_int
, u_int
,
257 int, const char *, struct pool_allocator
*);
258 void pool_destroy(struct pool
*);
260 void pool_set_drain_hook(struct pool
*,
261 void (*)(void *, int), void *);
263 void *pool_get(struct pool
*, int);
264 void pool_put(struct pool
*, void *);
265 int pool_reclaim(struct pool
*);
267 #ifdef POOL_DIAGNOSTIC
269 * These versions do reentrancy checking.
271 void *_pool_get(struct pool
*, int, const char *, long);
272 void _pool_put(struct pool
*, void *, const char *, long);
273 int _pool_reclaim(struct pool
*, const char *, long);
274 #define pool_get(h, f) _pool_get((h), (f), __FILE__, __LINE__)
275 #define pool_put(h, v) _pool_put((h), (v), __FILE__, __LINE__)
276 #define pool_reclaim(h) _pool_reclaim((h), __FILE__, __LINE__)
277 #endif /* POOL_DIAGNOSTIC */
279 int pool_prime(struct pool
*, int);
280 void pool_setlowat(struct pool
*, int);
281 void pool_sethiwat(struct pool
*, int);
282 void pool_sethardlimit(struct pool
*, int, const char *, int);
283 void pool_drain(void *);
286 * Debugging and diagnostic aides.
288 void pool_print(struct pool
*, const char *);
289 void pool_printit(struct pool
*, const char *,
290 void (*)(const char *, ...));
291 void pool_printall(const char *, void (*)(const char *, ...));
292 int pool_chk(struct pool
*, const char *);
295 * Pool cache routines.
297 void pool_cache_init(struct pool_cache
*, struct pool
*,
298 int (*)(void *, void *, int),
299 void (*)(void *, void *),
301 void pool_cache_destroy(struct pool_cache
*);
302 void *pool_cache_get_paddr(struct pool_cache
*, int, paddr_t
*);
303 #define pool_cache_get(pc, f) pool_cache_get_paddr((pc), (f), NULL)
304 void pool_cache_put_paddr(struct pool_cache
*, void *, paddr_t
);
305 #define pool_cache_put(pc, o) pool_cache_put_paddr((pc), (o), \
307 void pool_cache_destruct_object(struct pool_cache
*, void *);
308 void pool_cache_invalidate(struct pool_cache
*);
311 #endif /* _SYS_POOL_H_ */