Root/
1 | #ifndef __LINUX_MEMORY_HOTPLUG_H |
2 | #define __LINUX_MEMORY_HOTPLUG_H |
3 | |
4 | #include <linux/mmzone.h> |
5 | #include <linux/spinlock.h> |
6 | #include <linux/notifier.h> |
7 | |
8 | struct page; |
9 | struct zone; |
10 | struct pglist_data; |
11 | struct mem_section; |
12 | |
13 | #ifdef CONFIG_MEMORY_HOTPLUG |
14 | |
15 | /* |
16 | * Types for free bootmem. |
17 | * The normal smallest mapcount is -1. Here is smaller value than it. |
18 | */ |
19 | #define SECTION_INFO (-1 - 1) |
20 | #define MIX_SECTION_INFO (-1 - 2) |
21 | #define NODE_INFO (-1 - 3) |
22 | |
23 | /* |
24 | * pgdat resizing functions |
25 | */ |
26 | static inline |
27 | void pgdat_resize_lock(struct pglist_data *pgdat, unsigned long *flags) |
28 | { |
29 | spin_lock_irqsave(&pgdat->node_size_lock, *flags); |
30 | } |
31 | static inline |
32 | void pgdat_resize_unlock(struct pglist_data *pgdat, unsigned long *flags) |
33 | { |
34 | spin_unlock_irqrestore(&pgdat->node_size_lock, *flags); |
35 | } |
36 | static inline |
37 | void pgdat_resize_init(struct pglist_data *pgdat) |
38 | { |
39 | spin_lock_init(&pgdat->node_size_lock); |
40 | } |
41 | /* |
42 | * Zone resizing functions |
43 | */ |
44 | static inline unsigned zone_span_seqbegin(struct zone *zone) |
45 | { |
46 | return read_seqbegin(&zone->span_seqlock); |
47 | } |
48 | static inline int zone_span_seqretry(struct zone *zone, unsigned iv) |
49 | { |
50 | return read_seqretry(&zone->span_seqlock, iv); |
51 | } |
52 | static inline void zone_span_writelock(struct zone *zone) |
53 | { |
54 | write_seqlock(&zone->span_seqlock); |
55 | } |
56 | static inline void zone_span_writeunlock(struct zone *zone) |
57 | { |
58 | write_sequnlock(&zone->span_seqlock); |
59 | } |
60 | static inline void zone_seqlock_init(struct zone *zone) |
61 | { |
62 | seqlock_init(&zone->span_seqlock); |
63 | } |
64 | extern int zone_grow_free_lists(struct zone *zone, unsigned long new_nr_pages); |
65 | extern int zone_grow_waitqueues(struct zone *zone, unsigned long nr_pages); |
66 | extern int add_one_highpage(struct page *page, int pfn, int bad_ppro); |
67 | /* need some defines for these for archs that don't support it */ |
68 | extern void online_page(struct page *page); |
69 | /* VM interface that may be used by firmware interface */ |
70 | extern int online_pages(unsigned long, unsigned long); |
71 | extern void __offline_isolated_pages(unsigned long, unsigned long); |
72 | |
73 | /* reasonably generic interface to expand the physical pages in a zone */ |
74 | extern int __add_pages(int nid, struct zone *zone, unsigned long start_pfn, |
75 | unsigned long nr_pages); |
76 | extern int __remove_pages(struct zone *zone, unsigned long start_pfn, |
77 | unsigned long nr_pages); |
78 | |
79 | #ifdef CONFIG_NUMA |
80 | extern int memory_add_physaddr_to_nid(u64 start); |
81 | #else |
82 | static inline int memory_add_physaddr_to_nid(u64 start) |
83 | { |
84 | return 0; |
85 | } |
86 | #endif |
87 | |
88 | #ifdef CONFIG_HAVE_ARCH_NODEDATA_EXTENSION |
89 | /* |
90 | * For supporting node-hotadd, we have to allocate a new pgdat. |
91 | * |
92 | * If an arch has generic style NODE_DATA(), |
93 | * node_data[nid] = kzalloc() works well. But it depends on the architecture. |
94 | * |
95 | * In general, generic_alloc_nodedata() is used. |
96 | * Now, arch_free_nodedata() is just defined for error path of node_hot_add. |
97 | * |
98 | */ |
99 | extern pg_data_t *arch_alloc_nodedata(int nid); |
100 | extern void arch_free_nodedata(pg_data_t *pgdat); |
101 | extern void arch_refresh_nodedata(int nid, pg_data_t *pgdat); |
102 | |
103 | #else /* CONFIG_HAVE_ARCH_NODEDATA_EXTENSION */ |
104 | |
105 | #define arch_alloc_nodedata(nid) generic_alloc_nodedata(nid) |
106 | #define arch_free_nodedata(pgdat) generic_free_nodedata(pgdat) |
107 | |
108 | #ifdef CONFIG_NUMA |
109 | /* |
110 | * If ARCH_HAS_NODEDATA_EXTENSION=n, this func is used to allocate pgdat. |
111 | * XXX: kmalloc_node() can't work well to get new node's memory at this time. |
112 | * Because, pgdat for the new node is not allocated/initialized yet itself. |
113 | * To use new node's memory, more consideration will be necessary. |
114 | */ |
115 | #define generic_alloc_nodedata(nid) \ |
116 | ({ \ |
117 | kzalloc(sizeof(pg_data_t), GFP_KERNEL); \ |
118 | }) |
119 | /* |
120 | * This definition is just for error path in node hotadd. |
121 | * For node hotremove, we have to replace this. |
122 | */ |
123 | #define generic_free_nodedata(pgdat) kfree(pgdat) |
124 | |
125 | extern pg_data_t *node_data[]; |
126 | static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) |
127 | { |
128 | node_data[nid] = pgdat; |
129 | } |
130 | |
131 | #else /* !CONFIG_NUMA */ |
132 | |
133 | /* never called */ |
134 | static inline pg_data_t *generic_alloc_nodedata(int nid) |
135 | { |
136 | BUG(); |
137 | return NULL; |
138 | } |
139 | static inline void generic_free_nodedata(pg_data_t *pgdat) |
140 | { |
141 | } |
142 | static inline void arch_refresh_nodedata(int nid, pg_data_t *pgdat) |
143 | { |
144 | } |
145 | #endif /* CONFIG_NUMA */ |
146 | #endif /* CONFIG_HAVE_ARCH_NODEDATA_EXTENSION */ |
147 | |
148 | #ifdef CONFIG_SPARSEMEM_VMEMMAP |
149 | static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) |
150 | { |
151 | } |
152 | static inline void put_page_bootmem(struct page *page) |
153 | { |
154 | } |
155 | #else |
156 | extern void register_page_bootmem_info_node(struct pglist_data *pgdat); |
157 | extern void put_page_bootmem(struct page *page); |
158 | #endif |
159 | |
160 | #else /* ! CONFIG_MEMORY_HOTPLUG */ |
161 | /* |
162 | * Stub functions for when hotplug is off |
163 | */ |
164 | static inline void pgdat_resize_lock(struct pglist_data *p, unsigned long *f) {} |
165 | static inline void pgdat_resize_unlock(struct pglist_data *p, unsigned long *f) {} |
166 | static inline void pgdat_resize_init(struct pglist_data *pgdat) {} |
167 | |
168 | static inline unsigned zone_span_seqbegin(struct zone *zone) |
169 | { |
170 | return 0; |
171 | } |
172 | static inline int zone_span_seqretry(struct zone *zone, unsigned iv) |
173 | { |
174 | return 0; |
175 | } |
176 | static inline void zone_span_writelock(struct zone *zone) {} |
177 | static inline void zone_span_writeunlock(struct zone *zone) {} |
178 | static inline void zone_seqlock_init(struct zone *zone) {} |
179 | |
180 | static inline int mhp_notimplemented(const char *func) |
181 | { |
182 | printk(KERN_WARNING "%s() called, with CONFIG_MEMORY_HOTPLUG disabled\n", func); |
183 | dump_stack(); |
184 | return -ENOSYS; |
185 | } |
186 | |
187 | static inline void register_page_bootmem_info_node(struct pglist_data *pgdat) |
188 | { |
189 | } |
190 | |
191 | #endif /* ! CONFIG_MEMORY_HOTPLUG */ |
192 | |
193 | #ifdef CONFIG_MEMORY_HOTREMOVE |
194 | |
195 | extern int is_mem_section_removable(unsigned long pfn, unsigned long nr_pages); |
196 | |
197 | #else |
198 | static inline int is_mem_section_removable(unsigned long pfn, |
199 | unsigned long nr_pages) |
200 | { |
201 | return 0; |
202 | } |
203 | #endif /* CONFIG_MEMORY_HOTREMOVE */ |
204 | |
205 | extern int add_memory(int nid, u64 start, u64 size); |
206 | extern int arch_add_memory(int nid, u64 start, u64 size); |
207 | extern int remove_memory(u64 start, u64 size); |
208 | extern int sparse_add_one_section(struct zone *zone, unsigned long start_pfn, |
209 | int nr_pages); |
210 | extern void sparse_remove_one_section(struct zone *zone, struct mem_section *ms); |
211 | extern struct page *sparse_decode_mem_map(unsigned long coded_mem_map, |
212 | unsigned long pnum); |
213 | |
214 | #endif /* __LINUX_MEMORY_HOTPLUG_H */ |
215 |
Branches:
ben-wpan
ben-wpan-stefan
javiroman/ks7010
jz-2.6.34
jz-2.6.34-rc5
jz-2.6.34-rc6
jz-2.6.34-rc7
jz-2.6.35
jz-2.6.36
jz-2.6.37
jz-2.6.38
jz-2.6.39
jz-3.0
jz-3.1
jz-3.11
jz-3.12
jz-3.13
jz-3.15
jz-3.16
jz-3.18-dt
jz-3.2
jz-3.3
jz-3.4
jz-3.5
jz-3.6
jz-3.6-rc2-pwm
jz-3.9
jz-3.9-clk
jz-3.9-rc8
jz47xx
jz47xx-2.6.38
master
Tags:
od-2011-09-04
od-2011-09-18
v2.6.34-rc5
v2.6.34-rc6
v2.6.34-rc7
v3.9