]> git.proxmox.com Git - ceph.git/blame - ceph/src/spdk/dpdk/lib/librte_mempool/rte_mempool_ops_default.c
update source to Ceph Pacific 16.2.2
[ceph.git] / ceph / src / spdk / dpdk / lib / librte_mempool / rte_mempool_ops_default.c
CommitLineData
11fdf7f2
TL
1/* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2016 Intel Corporation.
3 * Copyright(c) 2016 6WIND S.A.
4 * Copyright(c) 2018 Solarflare Communications Inc.
5 */
6
7#include <rte_mempool.h>
8
9ssize_t
f67539c2
TL
10rte_mempool_op_calc_mem_size_helper(const struct rte_mempool *mp,
11 uint32_t obj_num, uint32_t pg_shift,
12 size_t chunk_reserve,
13 size_t *min_chunk_size, size_t *align)
11fdf7f2
TL
14{
15 size_t total_elt_sz;
f67539c2 16 size_t obj_per_page, pg_sz, objs_in_last_page;
11fdf7f2
TL
17 size_t mem_size;
18
19 total_elt_sz = mp->header_size + mp->elt_size + mp->trailer_size;
20 if (total_elt_sz == 0) {
21 mem_size = 0;
22 } else if (pg_shift == 0) {
f67539c2 23 mem_size = total_elt_sz * obj_num + chunk_reserve;
11fdf7f2
TL
24 } else {
25 pg_sz = (size_t)1 << pg_shift;
f67539c2
TL
26 if (chunk_reserve >= pg_sz)
27 return -EINVAL;
28 obj_per_page = (pg_sz - chunk_reserve) / total_elt_sz;
11fdf7f2
TL
29 if (obj_per_page == 0) {
30 /*
31 * Note that if object size is bigger than page size,
32 * then it is assumed that pages are grouped in subsets
33 * of physically continuous pages big enough to store
34 * at least one object.
35 */
f67539c2
TL
36 mem_size = RTE_ALIGN_CEIL(total_elt_sz + chunk_reserve,
37 pg_sz) * obj_num;
11fdf7f2 38 } else {
f67539c2
TL
39 /* In the best case, the allocator will return a
40 * page-aligned address. For example, with 5 objs,
41 * the required space is as below:
42 * | page0 | page1 | page2 (last) |
43 * |obj0 |obj1 |xxx|obj2 |obj3 |xxx|obj4|
44 * <------------- mem_size ------------->
45 */
46 objs_in_last_page = ((obj_num - 1) % obj_per_page) + 1;
47 /* room required for the last page */
48 mem_size = objs_in_last_page * total_elt_sz +
49 chunk_reserve;
50 /* room required for other pages */
51 mem_size += ((obj_num - objs_in_last_page) /
52 obj_per_page) << pg_shift;
53
54 /* In the worst case, the allocator returns a
55 * non-aligned pointer, wasting up to
56 * total_elt_sz. Add a margin for that.
57 */
58 mem_size += total_elt_sz - 1;
11fdf7f2
TL
59 }
60 }
61
f67539c2
TL
62 *min_chunk_size = total_elt_sz;
63 *align = RTE_MEMPOOL_ALIGN;
11fdf7f2
TL
64
65 return mem_size;
66}
67
f67539c2
TL
68ssize_t
69rte_mempool_op_calc_mem_size_default(const struct rte_mempool *mp,
70 uint32_t obj_num, uint32_t pg_shift,
71 size_t *min_chunk_size, size_t *align)
72{
73 return rte_mempool_op_calc_mem_size_helper(mp, obj_num, pg_shift,
74 0, min_chunk_size, align);
75}
76
77/* Returns -1 if object crosses a page boundary, else returns 0 */
78static int
79check_obj_bounds(char *obj, size_t pg_sz, size_t elt_sz)
80{
81 if (pg_sz == 0)
82 return 0;
83 if (elt_sz > pg_sz)
84 return 0;
85 if (RTE_PTR_ALIGN(obj, pg_sz) != RTE_PTR_ALIGN(obj + elt_sz - 1, pg_sz))
86 return -1;
87 return 0;
88}
89
11fdf7f2 90int
f67539c2
TL
91rte_mempool_op_populate_helper(struct rte_mempool *mp, unsigned int flags,
92 unsigned int max_objs, void *vaddr, rte_iova_t iova,
93 size_t len, rte_mempool_populate_obj_cb_t *obj_cb,
94 void *obj_cb_arg)
11fdf7f2 95{
f67539c2
TL
96 char *va = vaddr;
97 size_t total_elt_sz, pg_sz;
11fdf7f2
TL
98 size_t off;
99 unsigned int i;
100 void *obj;
f67539c2
TL
101 int ret;
102
103 ret = rte_mempool_get_page_size(mp, &pg_sz);
104 if (ret < 0)
105 return ret;
11fdf7f2
TL
106
107 total_elt_sz = mp->header_size + mp->elt_size + mp->trailer_size;
108
f67539c2
TL
109 if (flags & RTE_MEMPOOL_POPULATE_F_ALIGN_OBJ)
110 off = total_elt_sz - (((uintptr_t)(va - 1) % total_elt_sz) + 1);
111 else
112 off = 0;
113 for (i = 0; i < max_objs; i++) {
114 /* avoid objects to cross page boundaries */
115 if (check_obj_bounds(va + off, pg_sz, total_elt_sz) < 0) {
116 off += RTE_PTR_ALIGN_CEIL(va + off, pg_sz) - (va + off);
117 if (flags & RTE_MEMPOOL_POPULATE_F_ALIGN_OBJ)
118 off += total_elt_sz -
119 (((uintptr_t)(va + off - 1) %
120 total_elt_sz) + 1);
121 }
122
123 if (off + total_elt_sz > len)
124 break;
125
11fdf7f2 126 off += mp->header_size;
f67539c2 127 obj = va + off;
11fdf7f2
TL
128 obj_cb(mp, obj_cb_arg, obj,
129 (iova == RTE_BAD_IOVA) ? RTE_BAD_IOVA : (iova + off));
130 rte_mempool_ops_enqueue_bulk(mp, &obj, 1);
131 off += mp->elt_size + mp->trailer_size;
132 }
133
134 return i;
135}
f67539c2
TL
136
137int
138rte_mempool_op_populate_default(struct rte_mempool *mp, unsigned int max_objs,
139 void *vaddr, rte_iova_t iova, size_t len,
140 rte_mempool_populate_obj_cb_t *obj_cb,
141 void *obj_cb_arg)
142{
143 return rte_mempool_op_populate_helper(mp, 0, max_objs, vaddr, iova,
144 len, obj_cb, obj_cb_arg);
145}