]>
Commit | Line | Data |
---|---|---|
1da177e4 LT |
1 | /* |
2 | * Written by Pat Gaughen (gone@us.ibm.com) Mar 2002 | |
3 | * | |
4 | */ | |
5 | ||
6 | #ifndef _ASM_MMZONE_H_ | |
7 | #define _ASM_MMZONE_H_ | |
8 | ||
9 | #include <asm/smp.h> | |
10 | ||
11 | #ifdef CONFIG_DISCONTIGMEM | |
12 | ||
13 | #ifdef CONFIG_NUMA | |
14 | #ifdef CONFIG_X86_NUMAQ | |
15 | #include <asm/numaq.h> | |
16 | #else /* summit or generic arch */ | |
17 | #include <asm/srat.h> | |
18 | #endif | |
19 | #else /* !CONFIG_NUMA */ | |
20 | #define get_memcfg_numa get_memcfg_numa_flat | |
21 | #define get_zholes_size(n) (0) | |
22 | #endif /* CONFIG_NUMA */ | |
23 | ||
24 | extern struct pglist_data *node_data[]; | |
25 | #define NODE_DATA(nid) (node_data[nid]) | |
26 | ||
27 | /* | |
28 | * generic node memory support, the following assumptions apply: | |
29 | * | |
30 | * 1) memory comes in 256Mb contigious chunks which are either present or not | |
31 | * 2) we will not have more than 64Gb in total | |
32 | * | |
33 | * for now assume that 64Gb is max amount of RAM for whole system | |
34 | * 64Gb / 4096bytes/page = 16777216 pages | |
35 | */ | |
36 | #define MAX_NR_PAGES 16777216 | |
37 | #define MAX_ELEMENTS 256 | |
38 | #define PAGES_PER_ELEMENT (MAX_NR_PAGES/MAX_ELEMENTS) | |
39 | ||
40 | extern s8 physnode_map[]; | |
41 | ||
42 | static inline int pfn_to_nid(unsigned long pfn) | |
43 | { | |
44 | #ifdef CONFIG_NUMA | |
45 | return((int) physnode_map[(pfn) / PAGES_PER_ELEMENT]); | |
46 | #else | |
47 | return 0; | |
48 | #endif | |
49 | } | |
50 | ||
51 | /* | |
52 | * Following are macros that are specific to this numa platform. | |
53 | */ | |
54 | #define reserve_bootmem(addr, size) \ | |
55 | reserve_bootmem_node(NODE_DATA(0), (addr), (size)) | |
56 | #define alloc_bootmem(x) \ | |
57 | __alloc_bootmem_node(NODE_DATA(0), (x), SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS)) | |
58 | #define alloc_bootmem_low(x) \ | |
59 | __alloc_bootmem_node(NODE_DATA(0), (x), SMP_CACHE_BYTES, 0) | |
60 | #define alloc_bootmem_pages(x) \ | |
61 | __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, __pa(MAX_DMA_ADDRESS)) | |
62 | #define alloc_bootmem_low_pages(x) \ | |
63 | __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, 0) | |
64 | #define alloc_bootmem_node(ignore, x) \ | |
65 | __alloc_bootmem_node(NODE_DATA(0), (x), SMP_CACHE_BYTES, __pa(MAX_DMA_ADDRESS)) | |
66 | #define alloc_bootmem_pages_node(ignore, x) \ | |
67 | __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, __pa(MAX_DMA_ADDRESS)) | |
68 | #define alloc_bootmem_low_pages_node(ignore, x) \ | |
69 | __alloc_bootmem_node(NODE_DATA(0), (x), PAGE_SIZE, 0) | |
70 | ||
71 | #define node_localnr(pfn, nid) ((pfn) - node_data[nid]->node_start_pfn) | |
72 | ||
73 | /* | |
74 | * Following are macros that each numa implmentation must define. | |
75 | */ | |
76 | ||
77 | /* | |
78 | * Given a kernel address, find the home node of the underlying memory. | |
79 | */ | |
80 | #define kvaddr_to_nid(kaddr) pfn_to_nid(__pa(kaddr) >> PAGE_SHIFT) | |
81 | ||
82 | #define node_mem_map(nid) (NODE_DATA(nid)->node_mem_map) | |
83 | #define node_start_pfn(nid) (NODE_DATA(nid)->node_start_pfn) | |
84 | #define node_end_pfn(nid) \ | |
85 | ({ \ | |
86 | pg_data_t *__pgdat = NODE_DATA(nid); \ | |
87 | __pgdat->node_start_pfn + __pgdat->node_spanned_pages; \ | |
88 | }) | |
89 | ||
90 | #define local_mapnr(kvaddr) \ | |
91 | ({ \ | |
92 | unsigned long __pfn = __pa(kvaddr) >> PAGE_SHIFT; \ | |
93 | (__pfn - node_start_pfn(pfn_to_nid(__pfn))); \ | |
94 | }) | |
95 | ||
96 | /* XXX: FIXME -- wli */ | |
97 | #define kern_addr_valid(kaddr) (0) | |
98 | ||
99 | #define pfn_to_page(pfn) \ | |
100 | ({ \ | |
101 | unsigned long __pfn = pfn; \ | |
102 | int __node = pfn_to_nid(__pfn); \ | |
103 | &node_mem_map(__node)[node_localnr(__pfn,__node)]; \ | |
104 | }) | |
105 | ||
106 | #define page_to_pfn(pg) \ | |
107 | ({ \ | |
108 | struct page *__page = pg; \ | |
109 | struct zone *__zone = page_zone(__page); \ | |
110 | (unsigned long)(__page - __zone->zone_mem_map) \ | |
111 | + __zone->zone_start_pfn; \ | |
112 | }) | |
113 | ||
114 | #ifdef CONFIG_X86_NUMAQ /* we have contiguous memory on NUMA-Q */ | |
115 | #define pfn_valid(pfn) ((pfn) < num_physpages) | |
116 | #else | |
117 | static inline int pfn_valid(int pfn) | |
118 | { | |
119 | int nid = pfn_to_nid(pfn); | |
120 | ||
121 | if (nid >= 0) | |
122 | return (pfn < node_end_pfn(nid)); | |
123 | return 0; | |
124 | } | |
125 | #endif | |
126 | ||
127 | extern int get_memcfg_numa_flat(void ); | |
128 | /* | |
129 | * This allows any one NUMA architecture to be compiled | |
130 | * for, and still fall back to the flat function if it | |
131 | * fails. | |
132 | */ | |
133 | static inline void get_memcfg_numa(void) | |
134 | { | |
135 | #ifdef CONFIG_X86_NUMAQ | |
136 | if (get_memcfg_numaq()) | |
137 | return; | |
138 | #elif CONFIG_ACPI_SRAT | |
139 | if (get_memcfg_from_srat()) | |
140 | return; | |
141 | #endif | |
142 | ||
143 | get_memcfg_numa_flat(); | |
144 | } | |
145 | ||
146 | #endif /* CONFIG_DISCONTIGMEM */ | |
147 | #endif /* _ASM_MMZONE_H_ */ |