.. | .. |
---|
| 1 | +/* SPDX-License-Identifier: GPL-2.0-or-later */ |
---|
1 | 2 | /* memcontrol.h - Memory Controller |
---|
2 | 3 | * |
---|
3 | 4 | * Copyright IBM Corporation, 2007 |
---|
.. | .. |
---|
5 | 6 | * |
---|
6 | 7 | * Copyright 2007 OpenVZ SWsoft Inc |
---|
7 | 8 | * Author: Pavel Emelianov <xemul@openvz.org> |
---|
8 | | - * |
---|
9 | | - * This program is free software; you can redistribute it and/or modify |
---|
10 | | - * it under the terms of the GNU General Public License as published by |
---|
11 | | - * the Free Software Foundation; either version 2 of the License, or |
---|
12 | | - * (at your option) any later version. |
---|
13 | | - * |
---|
14 | | - * This program is distributed in the hope that it will be useful, |
---|
15 | | - * but WITHOUT ANY WARRANTY; without even the implied warranty of |
---|
16 | | - * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
---|
17 | | - * GNU General Public License for more details. |
---|
18 | 9 | */ |
---|
19 | 10 | |
---|
20 | 11 | #ifndef _LINUX_MEMCONTROL_H |
---|
.. | .. |
---|
32 | 23 | #include <linux/page-flags.h> |
---|
33 | 24 | |
---|
34 | 25 | struct mem_cgroup; |
---|
| 26 | +struct obj_cgroup; |
---|
35 | 27 | struct page; |
---|
36 | 28 | struct mm_struct; |
---|
37 | 29 | struct kmem_cache; |
---|
38 | 30 | |
---|
39 | 31 | /* Cgroup-specific page state, on top of universal node page state */ |
---|
40 | 32 | enum memcg_stat_item { |
---|
41 | | - MEMCG_CACHE = NR_VM_NODE_STAT_ITEMS, |
---|
42 | | - MEMCG_RSS, |
---|
43 | | - MEMCG_RSS_HUGE, |
---|
44 | | - MEMCG_SWAP, |
---|
| 33 | + MEMCG_SWAP = NR_VM_NODE_STAT_ITEMS, |
---|
45 | 34 | MEMCG_SOCK, |
---|
46 | | - /* XXX: why are these zone and not node counters? */ |
---|
47 | | - MEMCG_KERNEL_STACK_KB, |
---|
| 35 | + MEMCG_PERCPU_B, |
---|
48 | 36 | MEMCG_NR_STAT, |
---|
49 | 37 | }; |
---|
50 | 38 | |
---|
.. | .. |
---|
54 | 42 | MEMCG_MAX, |
---|
55 | 43 | MEMCG_OOM, |
---|
56 | 44 | MEMCG_OOM_KILL, |
---|
| 45 | + MEMCG_SWAP_HIGH, |
---|
57 | 46 | MEMCG_SWAP_MAX, |
---|
58 | 47 | MEMCG_SWAP_FAIL, |
---|
59 | 48 | MEMCG_NR_MEMORY_EVENTS, |
---|
60 | 49 | }; |
---|
61 | 50 | |
---|
62 | | -enum mem_cgroup_protection { |
---|
63 | | - MEMCG_PROT_NONE, |
---|
64 | | - MEMCG_PROT_LOW, |
---|
65 | | - MEMCG_PROT_MIN, |
---|
66 | | -}; |
---|
67 | | - |
---|
68 | 51 | struct mem_cgroup_reclaim_cookie { |
---|
69 | 52 | pg_data_t *pgdat; |
---|
70 | | - int priority; |
---|
71 | 53 | unsigned int generation; |
---|
72 | 54 | }; |
---|
73 | 55 | |
---|
.. | .. |
---|
78 | 60 | |
---|
79 | 61 | struct mem_cgroup_id { |
---|
80 | 62 | int id; |
---|
81 | | - atomic_t ref; |
---|
| 63 | + refcount_t ref; |
---|
82 | 64 | }; |
---|
83 | 65 | |
---|
84 | 66 | /* |
---|
85 | 67 | * Per memcg event counter is incremented at every pagein/pageout. With THP, |
---|
86 | | - * it will be incremated by the number of pages. This counter is used for |
---|
87 | | - * for trigger some periodic events. This is straightforward and better |
---|
| 68 | + * it will be incremented by the number of pages. This counter is used |
---|
| 69 | + * to trigger some periodic events. This is straightforward and better |
---|
88 | 70 | * than using jiffies etc. to handle periodic memcg event. |
---|
89 | 71 | */ |
---|
90 | 72 | enum mem_cgroup_events_target { |
---|
91 | 73 | MEM_CGROUP_TARGET_THRESH, |
---|
92 | 74 | MEM_CGROUP_TARGET_SOFTLIMIT, |
---|
93 | | - MEM_CGROUP_TARGET_NUMAINFO, |
---|
94 | 75 | MEM_CGROUP_NTARGETS, |
---|
95 | 76 | }; |
---|
96 | 77 | |
---|
97 | | -struct mem_cgroup_stat_cpu { |
---|
98 | | - long count[MEMCG_NR_STAT]; |
---|
| 78 | +struct memcg_vmstats_percpu { |
---|
| 79 | + long stat[MEMCG_NR_STAT]; |
---|
99 | 80 | unsigned long events[NR_VM_EVENT_ITEMS]; |
---|
100 | 81 | unsigned long nr_page_events; |
---|
101 | 82 | unsigned long targets[MEM_CGROUP_NTARGETS]; |
---|
.. | .. |
---|
117 | 98 | */ |
---|
118 | 99 | struct memcg_shrinker_map { |
---|
119 | 100 | struct rcu_head rcu; |
---|
120 | | - unsigned long map[0]; |
---|
| 101 | + unsigned long map[]; |
---|
121 | 102 | }; |
---|
122 | 103 | |
---|
123 | 104 | /* |
---|
124 | | - * per-zone information in memory controller. |
---|
| 105 | + * per-node information in memory controller. |
---|
125 | 106 | */ |
---|
126 | 107 | struct mem_cgroup_per_node { |
---|
127 | 108 | struct lruvec lruvec; |
---|
128 | 109 | |
---|
| 110 | + /* Legacy local VM stats */ |
---|
| 111 | + struct lruvec_stat __percpu *lruvec_stat_local; |
---|
| 112 | + |
---|
| 113 | + /* Subtree VM stats (batched updates) */ |
---|
129 | 114 | struct lruvec_stat __percpu *lruvec_stat_cpu; |
---|
130 | 115 | atomic_long_t lruvec_stat[NR_VM_NODE_STAT_ITEMS]; |
---|
131 | 116 | |
---|
132 | 117 | unsigned long lru_zone_size[MAX_NR_ZONES][NR_LRU_LISTS]; |
---|
133 | 118 | |
---|
134 | | - struct mem_cgroup_reclaim_iter iter[DEF_PRIORITY + 1]; |
---|
| 119 | + struct mem_cgroup_reclaim_iter iter; |
---|
135 | 120 | |
---|
136 | | -#ifdef CONFIG_MEMCG_KMEM |
---|
137 | 121 | struct memcg_shrinker_map __rcu *shrinker_map; |
---|
138 | | -#endif |
---|
| 122 | + |
---|
139 | 123 | struct rb_node tree_node; /* RB tree node */ |
---|
140 | 124 | unsigned long usage_in_excess;/* Set to the value by which */ |
---|
141 | 125 | /* the soft limit is exceeded*/ |
---|
142 | 126 | bool on_tree; |
---|
143 | | - bool congested; /* memcg has many dirty pages */ |
---|
144 | | - /* backed by a congested BDI */ |
---|
145 | | - |
---|
146 | 127 | struct mem_cgroup *memcg; /* Back pointer, we cannot */ |
---|
147 | 128 | /* use container_of */ |
---|
148 | 129 | }; |
---|
.. | .. |
---|
159 | 140 | /* Size of entries[] */ |
---|
160 | 141 | unsigned int size; |
---|
161 | 142 | /* Array of thresholds */ |
---|
162 | | - struct mem_cgroup_threshold entries[0]; |
---|
| 143 | + struct mem_cgroup_threshold entries[]; |
---|
163 | 144 | }; |
---|
164 | 145 | |
---|
165 | 146 | struct mem_cgroup_thresholds { |
---|
.. | .. |
---|
189 | 170 | #endif |
---|
190 | 171 | |
---|
191 | 172 | /* |
---|
| 173 | + * Remember four most recent foreign writebacks with dirty pages in this |
---|
| 174 | + * cgroup. Inode sharing is expected to be uncommon and, even if we miss |
---|
| 175 | + * one in a given round, we're likely to catch it later if it keeps |
---|
| 176 | + * foreign-dirtying, so a fairly low count should be enough. |
---|
| 177 | + * |
---|
| 178 | + * See mem_cgroup_track_foreign_dirty_slowpath() for details. |
---|
| 179 | + */ |
---|
| 180 | +#define MEMCG_CGWB_FRN_CNT 4 |
---|
| 181 | + |
---|
| 182 | +struct memcg_cgwb_frn { |
---|
| 183 | + u64 bdi_id; /* bdi->id of the foreign inode */ |
---|
| 184 | + int memcg_id; /* memcg->css.id of foreign inode */ |
---|
| 185 | + u64 at; /* jiffies_64 at the time of dirtying */ |
---|
| 186 | + struct wb_completion done; /* tracks in-flight foreign writebacks */ |
---|
| 187 | +}; |
---|
| 188 | + |
---|
| 189 | +/* |
---|
| 190 | + * Bucket for arbitrarily byte-sized objects charged to a memory |
---|
| 191 | + * cgroup. The bucket can be reparented in one piece when the cgroup |
---|
| 192 | + * is destroyed, without having to round up the individual references |
---|
| 193 | + * of all live memory objects in the wild. |
---|
| 194 | + */ |
---|
| 195 | +struct obj_cgroup { |
---|
| 196 | + struct percpu_ref refcnt; |
---|
| 197 | + struct mem_cgroup *memcg; |
---|
| 198 | + atomic_t nr_charged_bytes; |
---|
| 199 | + union { |
---|
| 200 | + struct list_head list; /* protected by objcg_lock */ |
---|
| 201 | + struct rcu_head rcu; |
---|
| 202 | + }; |
---|
| 203 | +}; |
---|
| 204 | + |
---|
| 205 | +/* |
---|
192 | 206 | * The memory controller data structure. The memory controller controls both |
---|
193 | 207 | * page cache and RSS per cgroup. We would eventually like to provide |
---|
194 | 208 | * statistics based on the statistics developed by Rik Van Riel for clock-pro, |
---|
.. | .. |
---|
201 | 215 | struct mem_cgroup_id id; |
---|
202 | 216 | |
---|
203 | 217 | /* Accounted resources */ |
---|
204 | | - struct page_counter memory; |
---|
205 | | - struct page_counter swap; |
---|
| 218 | + struct page_counter memory; /* Both v1 & v2 */ |
---|
| 219 | + |
---|
| 220 | + union { |
---|
| 221 | + struct page_counter swap; /* v2 only */ |
---|
| 222 | + struct page_counter memsw; /* v1 only */ |
---|
| 223 | + }; |
---|
206 | 224 | |
---|
207 | 225 | /* Legacy consumer-oriented counters */ |
---|
208 | | - struct page_counter memsw; |
---|
209 | | - struct page_counter kmem; |
---|
210 | | - struct page_counter tcpmem; |
---|
211 | | - |
---|
212 | | - /* Upper bound of normal memory consumption range */ |
---|
213 | | - unsigned long high; |
---|
| 226 | + struct page_counter kmem; /* v1 only */ |
---|
| 227 | + struct page_counter tcpmem; /* v1 only */ |
---|
214 | 228 | |
---|
215 | 229 | /* Range enforcement for interrupt charges */ |
---|
216 | 230 | struct work_struct high_work; |
---|
.. | .. |
---|
238 | 252 | /* OOM-Killer disable */ |
---|
239 | 253 | int oom_kill_disable; |
---|
240 | 254 | |
---|
241 | | - /* memory.events */ |
---|
| 255 | + /* memory.events and memory.events.local */ |
---|
242 | 256 | struct cgroup_file events_file; |
---|
| 257 | + struct cgroup_file events_local_file; |
---|
243 | 258 | |
---|
244 | 259 | /* handle for "memory.swap.events" */ |
---|
245 | 260 | struct cgroup_file swap_events_file; |
---|
.. | .. |
---|
267 | 282 | |
---|
268 | 283 | MEMCG_PADDING(_pad1_); |
---|
269 | 284 | |
---|
270 | | - /* |
---|
271 | | - * set > 0 if pages under this cgroup are moving to other cgroup. |
---|
272 | | - */ |
---|
273 | | - atomic_t moving_account; |
---|
274 | | - struct task_struct *move_lock_task; |
---|
| 285 | + atomic_long_t vmstats[MEMCG_NR_STAT]; |
---|
| 286 | + atomic_long_t vmevents[NR_VM_EVENT_ITEMS]; |
---|
275 | 287 | |
---|
276 | | - /* memory.stat */ |
---|
277 | | - struct mem_cgroup_stat_cpu __percpu *stat_cpu; |
---|
278 | | - |
---|
279 | | - MEMCG_PADDING(_pad2_); |
---|
280 | | - |
---|
281 | | - atomic_long_t stat[MEMCG_NR_STAT]; |
---|
282 | | - atomic_long_t events[NR_VM_EVENT_ITEMS]; |
---|
283 | | - atomic_long_t memory_events[MEMCG_NR_MEMORY_EVENTS]; |
---|
| 288 | + /* memory.events */ |
---|
| 289 | + atomic_long_t memory_events[MEMCG_NR_MEMORY_EVENTS]; |
---|
| 290 | + atomic_long_t memory_events_local[MEMCG_NR_MEMORY_EVENTS]; |
---|
284 | 291 | |
---|
285 | 292 | unsigned long socket_pressure; |
---|
286 | 293 | |
---|
.. | .. |
---|
292 | 299 | /* Index in the kmem_cache->memcg_params.memcg_caches array */ |
---|
293 | 300 | int kmemcg_id; |
---|
294 | 301 | enum memcg_kmem_state kmem_state; |
---|
295 | | - struct list_head kmem_caches; |
---|
| 302 | + struct obj_cgroup __rcu *objcg; |
---|
| 303 | + /* list of inherited objcgs, protected by objcg_lock */ |
---|
| 304 | + struct list_head objcg_list; |
---|
296 | 305 | #endif |
---|
297 | 306 | |
---|
298 | | - int last_scanned_node; |
---|
299 | | -#if MAX_NUMNODES > 1 |
---|
300 | | - nodemask_t scan_nodes; |
---|
301 | | - atomic_t numainfo_events; |
---|
302 | | - atomic_t numainfo_updating; |
---|
303 | | -#endif |
---|
| 307 | + MEMCG_PADDING(_pad2_); |
---|
| 308 | + |
---|
| 309 | + /* |
---|
| 310 | + * set > 0 if pages under this cgroup are moving to other cgroup. |
---|
| 311 | + */ |
---|
| 312 | + atomic_t moving_account; |
---|
| 313 | + struct task_struct *move_lock_task; |
---|
| 314 | + |
---|
| 315 | + /* Legacy local VM stats and events */ |
---|
| 316 | + struct memcg_vmstats_percpu __percpu *vmstats_local; |
---|
| 317 | + |
---|
| 318 | + /* Subtree VM stats and events (batched updates) */ |
---|
| 319 | + struct memcg_vmstats_percpu __percpu *vmstats_percpu; |
---|
304 | 320 | |
---|
305 | 321 | #ifdef CONFIG_CGROUP_WRITEBACK |
---|
306 | 322 | struct list_head cgwb_list; |
---|
307 | 323 | struct wb_domain cgwb_domain; |
---|
| 324 | + struct memcg_cgwb_frn cgwb_frn[MEMCG_CGWB_FRN_CNT]; |
---|
308 | 325 | #endif |
---|
309 | 326 | |
---|
310 | 327 | /* List of events which userspace want to receive */ |
---|
311 | 328 | struct list_head event_list; |
---|
312 | 329 | spinlock_t event_list_lock; |
---|
313 | 330 | |
---|
| 331 | +#ifdef CONFIG_TRANSPARENT_HUGEPAGE |
---|
| 332 | + struct deferred_split deferred_split_queue; |
---|
| 333 | +#endif |
---|
| 334 | + |
---|
| 335 | + ANDROID_OEM_DATA(1); |
---|
314 | 336 | struct mem_cgroup_per_node *nodeinfo[0]; |
---|
315 | 337 | /* WARNING: nodeinfo must be the last member here */ |
---|
316 | 338 | }; |
---|
.. | .. |
---|
323 | 345 | |
---|
324 | 346 | extern struct mem_cgroup *root_mem_cgroup; |
---|
325 | 347 | |
---|
| 348 | +struct lruvec *page_to_lruvec(struct page *page, pg_data_t *pgdat); |
---|
| 349 | +void do_traversal_all_lruvec(void); |
---|
| 350 | + |
---|
| 351 | +static __always_inline bool memcg_stat_item_in_bytes(int idx) |
---|
| 352 | +{ |
---|
| 353 | + if (idx == MEMCG_PERCPU_B) |
---|
| 354 | + return true; |
---|
| 355 | + return vmstat_item_in_bytes(idx); |
---|
| 356 | +} |
---|
| 357 | + |
---|
326 | 358 | static inline bool mem_cgroup_is_root(struct mem_cgroup *memcg) |
---|
327 | 359 | { |
---|
328 | 360 | return (memcg == root_mem_cgroup); |
---|
.. | .. |
---|
333 | 365 | return !cgroup_subsys_enabled(memory_cgrp_subsys); |
---|
334 | 366 | } |
---|
335 | 367 | |
---|
336 | | -enum mem_cgroup_protection mem_cgroup_protected(struct mem_cgroup *root, |
---|
337 | | - struct mem_cgroup *memcg); |
---|
| 368 | +static inline void mem_cgroup_protection(struct mem_cgroup *root, |
---|
| 369 | + struct mem_cgroup *memcg, |
---|
| 370 | + unsigned long *min, |
---|
| 371 | + unsigned long *low) |
---|
| 372 | +{ |
---|
| 373 | + *min = *low = 0; |
---|
338 | 374 | |
---|
339 | | -int mem_cgroup_try_charge(struct page *page, struct mm_struct *mm, |
---|
340 | | - gfp_t gfp_mask, struct mem_cgroup **memcgp, |
---|
341 | | - bool compound); |
---|
342 | | -int mem_cgroup_try_charge_delay(struct page *page, struct mm_struct *mm, |
---|
343 | | - gfp_t gfp_mask, struct mem_cgroup **memcgp, |
---|
344 | | - bool compound); |
---|
345 | | -void mem_cgroup_commit_charge(struct page *page, struct mem_cgroup *memcg, |
---|
346 | | - bool lrucare, bool compound); |
---|
347 | | -void mem_cgroup_cancel_charge(struct page *page, struct mem_cgroup *memcg, |
---|
348 | | - bool compound); |
---|
349 | | -void mem_cgroup_uncharge(struct page *page); |
---|
350 | | -void mem_cgroup_uncharge_list(struct list_head *page_list); |
---|
| 375 | + if (mem_cgroup_disabled()) |
---|
| 376 | + return; |
---|
| 377 | + |
---|
| 378 | + /* |
---|
| 379 | + * There is no reclaim protection applied to a targeted reclaim. |
---|
| 380 | + * We are special casing this specific case here because |
---|
| 381 | + * mem_cgroup_protected calculation is not robust enough to keep |
---|
| 382 | + * the protection invariant for calculated effective values for |
---|
| 383 | + * parallel reclaimers with different reclaim target. This is |
---|
| 384 | + * especially a problem for tail memcgs (as they have pages on LRU) |
---|
| 385 | + * which would want to have effective values 0 for targeted reclaim |
---|
| 386 | + * but a different value for external reclaim. |
---|
| 387 | + * |
---|
| 388 | + * Example |
---|
| 389 | + * Let's have global and A's reclaim in parallel: |
---|
| 390 | + * | |
---|
| 391 | + * A (low=2G, usage = 3G, max = 3G, children_low_usage = 1.5G) |
---|
| 392 | + * |\ |
---|
| 393 | + * | C (low = 1G, usage = 2.5G) |
---|
| 394 | + * B (low = 1G, usage = 0.5G) |
---|
| 395 | + * |
---|
| 396 | + * For the global reclaim |
---|
| 397 | + * A.elow = A.low |
---|
| 398 | + * B.elow = min(B.usage, B.low) because children_low_usage <= A.elow |
---|
| 399 | + * C.elow = min(C.usage, C.low) |
---|
| 400 | + * |
---|
| 401 | + * With the effective values resetting we have A reclaim |
---|
| 402 | + * A.elow = 0 |
---|
| 403 | + * B.elow = B.low |
---|
| 404 | + * C.elow = C.low |
---|
| 405 | + * |
---|
| 406 | + * If the global reclaim races with A's reclaim then |
---|
| 407 | + * B.elow = C.elow = 0 because children_low_usage > A.elow) |
---|
| 408 | + * is possible and reclaiming B would be violating the protection. |
---|
| 409 | + * |
---|
| 410 | + */ |
---|
| 411 | + if (root == memcg) |
---|
| 412 | + return; |
---|
| 413 | + |
---|
| 414 | + *min = READ_ONCE(memcg->memory.emin); |
---|
| 415 | + *low = READ_ONCE(memcg->memory.elow); |
---|
| 416 | +} |
---|
| 417 | + |
---|
| 418 | +void mem_cgroup_calculate_protection(struct mem_cgroup *root, |
---|
| 419 | + struct mem_cgroup *memcg); |
---|
| 420 | + |
---|
| 421 | +static inline bool mem_cgroup_supports_protection(struct mem_cgroup *memcg) |
---|
| 422 | +{ |
---|
| 423 | + /* |
---|
| 424 | + * The root memcg doesn't account charges, and doesn't support |
---|
| 425 | + * protection. |
---|
| 426 | + */ |
---|
| 427 | + return !mem_cgroup_disabled() && !mem_cgroup_is_root(memcg); |
---|
| 428 | + |
---|
| 429 | +} |
---|
| 430 | + |
---|
| 431 | +static inline bool mem_cgroup_below_low(struct mem_cgroup *memcg) |
---|
| 432 | +{ |
---|
| 433 | + if (!mem_cgroup_supports_protection(memcg)) |
---|
| 434 | + return false; |
---|
| 435 | + |
---|
| 436 | + return READ_ONCE(memcg->memory.elow) >= |
---|
| 437 | + page_counter_read(&memcg->memory); |
---|
| 438 | +} |
---|
| 439 | + |
---|
| 440 | +static inline bool mem_cgroup_below_min(struct mem_cgroup *memcg) |
---|
| 441 | +{ |
---|
| 442 | + if (!mem_cgroup_supports_protection(memcg)) |
---|
| 443 | + return false; |
---|
| 444 | + |
---|
| 445 | + return READ_ONCE(memcg->memory.emin) >= |
---|
| 446 | + page_counter_read(&memcg->memory); |
---|
| 447 | +} |
---|
| 448 | + |
---|
| 449 | +int __mem_cgroup_charge(struct page *page, struct mm_struct *mm, |
---|
| 450 | + gfp_t gfp_mask); |
---|
| 451 | +static inline int mem_cgroup_charge(struct page *page, struct mm_struct *mm, |
---|
| 452 | + gfp_t gfp_mask) |
---|
| 453 | +{ |
---|
| 454 | + if (mem_cgroup_disabled()) |
---|
| 455 | + return 0; |
---|
| 456 | + return __mem_cgroup_charge(page, mm, gfp_mask); |
---|
| 457 | +} |
---|
| 458 | + |
---|
| 459 | +void __mem_cgroup_uncharge(struct page *page); |
---|
| 460 | +static inline void mem_cgroup_uncharge(struct page *page) |
---|
| 461 | +{ |
---|
| 462 | + if (mem_cgroup_disabled()) |
---|
| 463 | + return; |
---|
| 464 | + __mem_cgroup_uncharge(page); |
---|
| 465 | +} |
---|
| 466 | + |
---|
| 467 | +void __mem_cgroup_uncharge_list(struct list_head *page_list); |
---|
| 468 | +static inline void mem_cgroup_uncharge_list(struct list_head *page_list) |
---|
| 469 | +{ |
---|
| 470 | + if (mem_cgroup_disabled()) |
---|
| 471 | + return; |
---|
| 472 | + __mem_cgroup_uncharge_list(page_list); |
---|
| 473 | +} |
---|
351 | 474 | |
---|
352 | 475 | void mem_cgroup_migrate(struct page *oldpage, struct page *newpage); |
---|
353 | 476 | |
---|
.. | .. |
---|
358 | 481 | } |
---|
359 | 482 | |
---|
360 | 483 | /** |
---|
361 | | - * mem_cgroup_lruvec - get the lru list vector for a node or a memcg zone |
---|
362 | | - * @node: node of the wanted lruvec |
---|
| 484 | + * mem_cgroup_lruvec - get the lru list vector for a memcg & node |
---|
363 | 485 | * @memcg: memcg of the wanted lruvec |
---|
364 | 486 | * |
---|
365 | | - * Returns the lru list vector holding pages for a given @node or a given |
---|
366 | | - * @memcg and @zone. This can be the node lruvec, if the memory controller |
---|
367 | | - * is disabled. |
---|
| 487 | + * Returns the lru list vector holding pages for a given @memcg & |
---|
| 488 | + * @node combination. This can be the node lruvec, if the memory |
---|
| 489 | + * controller is disabled. |
---|
368 | 490 | */ |
---|
369 | | -static inline struct lruvec *mem_cgroup_lruvec(struct pglist_data *pgdat, |
---|
370 | | - struct mem_cgroup *memcg) |
---|
| 491 | +static inline struct lruvec *mem_cgroup_lruvec(struct mem_cgroup *memcg, |
---|
| 492 | + struct pglist_data *pgdat) |
---|
371 | 493 | { |
---|
372 | 494 | struct mem_cgroup_per_node *mz; |
---|
373 | 495 | struct lruvec *lruvec; |
---|
374 | 496 | |
---|
375 | 497 | if (mem_cgroup_disabled()) { |
---|
376 | | - lruvec = node_lruvec(pgdat); |
---|
| 498 | + lruvec = &pgdat->__lruvec; |
---|
377 | 499 | goto out; |
---|
378 | 500 | } |
---|
| 501 | + |
---|
| 502 | + if (!memcg) |
---|
| 503 | + memcg = root_mem_cgroup; |
---|
379 | 504 | |
---|
380 | 505 | mz = mem_cgroup_nodeinfo(memcg, pgdat->node_id); |
---|
381 | 506 | lruvec = &mz->lruvec; |
---|
.. | .. |
---|
392 | 517 | |
---|
393 | 518 | struct lruvec *mem_cgroup_page_lruvec(struct page *, struct pglist_data *); |
---|
394 | 519 | |
---|
395 | | -bool task_in_mem_cgroup(struct task_struct *task, struct mem_cgroup *memcg); |
---|
396 | 520 | struct mem_cgroup *mem_cgroup_from_task(struct task_struct *p); |
---|
397 | 521 | |
---|
398 | 522 | struct mem_cgroup *get_mem_cgroup_from_mm(struct mm_struct *mm); |
---|
.. | .. |
---|
402 | 526 | static inline |
---|
403 | 527 | struct mem_cgroup *mem_cgroup_from_css(struct cgroup_subsys_state *css){ |
---|
404 | 528 | return css ? container_of(css, struct mem_cgroup, css) : NULL; |
---|
| 529 | +} |
---|
| 530 | + |
---|
| 531 | +static inline bool obj_cgroup_tryget(struct obj_cgroup *objcg) |
---|
| 532 | +{ |
---|
| 533 | + return percpu_ref_tryget(&objcg->refcnt); |
---|
| 534 | +} |
---|
| 535 | + |
---|
| 536 | +static inline void obj_cgroup_get(struct obj_cgroup *objcg) |
---|
| 537 | +{ |
---|
| 538 | + percpu_ref_get(&objcg->refcnt); |
---|
| 539 | +} |
---|
| 540 | + |
---|
| 541 | +static inline void obj_cgroup_put(struct obj_cgroup *objcg) |
---|
| 542 | +{ |
---|
| 543 | + percpu_ref_put(&objcg->refcnt); |
---|
| 544 | +} |
---|
| 545 | + |
---|
| 546 | +/* |
---|
| 547 | + * After the initialization objcg->memcg is always pointing at |
---|
| 548 | + * a valid memcg, but can be atomically swapped to the parent memcg. |
---|
| 549 | + * |
---|
| 550 | + * The caller must ensure that the returned memcg won't be released: |
---|
| 551 | + * e.g. acquire the rcu_read_lock or css_set_lock. |
---|
| 552 | + */ |
---|
| 553 | +static inline struct mem_cgroup *obj_cgroup_memcg(struct obj_cgroup *objcg) |
---|
| 554 | +{ |
---|
| 555 | + return READ_ONCE(objcg->memcg); |
---|
405 | 556 | } |
---|
406 | 557 | |
---|
407 | 558 | static inline void mem_cgroup_put(struct mem_cgroup *memcg) |
---|
.. | .. |
---|
428 | 579 | return memcg->id.id; |
---|
429 | 580 | } |
---|
430 | 581 | struct mem_cgroup *mem_cgroup_from_id(unsigned short id); |
---|
| 582 | + |
---|
| 583 | +static inline struct mem_cgroup *mem_cgroup_from_seq(struct seq_file *m) |
---|
| 584 | +{ |
---|
| 585 | + return mem_cgroup_from_css(seq_css(m)); |
---|
| 586 | +} |
---|
431 | 587 | |
---|
432 | 588 | static inline struct mem_cgroup *lruvec_memcg(struct lruvec *lruvec) |
---|
433 | 589 | { |
---|
.. | .. |
---|
496 | 652 | void mem_cgroup_update_lru_size(struct lruvec *lruvec, enum lru_list lru, |
---|
497 | 653 | int zid, int nr_pages); |
---|
498 | 654 | |
---|
499 | | -unsigned long mem_cgroup_node_nr_lru_pages(struct mem_cgroup *memcg, |
---|
500 | | - int nid, unsigned int lru_mask); |
---|
501 | | - |
---|
502 | | -static inline |
---|
503 | | -unsigned long mem_cgroup_get_lru_size(struct lruvec *lruvec, enum lru_list lru) |
---|
504 | | -{ |
---|
505 | | - struct mem_cgroup_per_node *mz; |
---|
506 | | - unsigned long nr_pages = 0; |
---|
507 | | - int zid; |
---|
508 | | - |
---|
509 | | - mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); |
---|
510 | | - for (zid = 0; zid < MAX_NR_ZONES; zid++) |
---|
511 | | - nr_pages += mz->lru_zone_size[zid][lru]; |
---|
512 | | - return nr_pages; |
---|
513 | | -} |
---|
514 | | - |
---|
515 | 655 | static inline |
---|
516 | 656 | unsigned long mem_cgroup_get_zone_lru_size(struct lruvec *lruvec, |
---|
517 | 657 | enum lru_list lru, int zone_idx) |
---|
.. | .. |
---|
519 | 659 | struct mem_cgroup_per_node *mz; |
---|
520 | 660 | |
---|
521 | 661 | mz = container_of(lruvec, struct mem_cgroup_per_node, lruvec); |
---|
522 | | - return mz->lru_zone_size[zone_idx][lru]; |
---|
| 662 | + return READ_ONCE(mz->lru_zone_size[zone_idx][lru]); |
---|
523 | 663 | } |
---|
524 | 664 | |
---|
525 | 665 | void mem_cgroup_handle_over_high(void); |
---|
526 | 666 | |
---|
527 | 667 | unsigned long mem_cgroup_get_max(struct mem_cgroup *memcg); |
---|
528 | 668 | |
---|
529 | | -void mem_cgroup_print_oom_info(struct mem_cgroup *memcg, |
---|
| 669 | +unsigned long mem_cgroup_size(struct mem_cgroup *memcg); |
---|
| 670 | + |
---|
| 671 | +void mem_cgroup_print_oom_context(struct mem_cgroup *memcg, |
---|
530 | 672 | struct task_struct *p); |
---|
| 673 | + |
---|
| 674 | +void mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg); |
---|
531 | 675 | |
---|
532 | 676 | static inline void mem_cgroup_enter_user_fault(void) |
---|
533 | 677 | { |
---|
.. | .. |
---|
552 | 696 | void mem_cgroup_print_oom_group(struct mem_cgroup *memcg); |
---|
553 | 697 | |
---|
554 | 698 | #ifdef CONFIG_MEMCG_SWAP |
---|
555 | | -extern int do_swap_account; |
---|
| 699 | +extern bool cgroup_memory_noswap; |
---|
556 | 700 | #endif |
---|
557 | 701 | |
---|
558 | 702 | struct mem_cgroup *lock_page_memcg(struct page *page); |
---|
.. | .. |
---|
563 | 707 | * idx can be of type enum memcg_stat_item or node_stat_item. |
---|
564 | 708 | * Keep in sync with memcg_exact_page_state(). |
---|
565 | 709 | */ |
---|
566 | | -static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, |
---|
567 | | - int idx) |
---|
| 710 | +static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx) |
---|
568 | 711 | { |
---|
569 | | - long x = atomic_long_read(&memcg->stat[idx]); |
---|
| 712 | + long x = atomic_long_read(&memcg->vmstats[idx]); |
---|
570 | 713 | #ifdef CONFIG_SMP |
---|
571 | 714 | if (x < 0) |
---|
572 | 715 | x = 0; |
---|
.. | .. |
---|
574 | 717 | return x; |
---|
575 | 718 | } |
---|
576 | 719 | |
---|
577 | | -/* idx can be of type enum memcg_stat_item or node_stat_item */ |
---|
578 | | -static inline void __mod_memcg_state(struct mem_cgroup *memcg, |
---|
579 | | - int idx, int val) |
---|
| 720 | +/* |
---|
| 721 | + * idx can be of type enum memcg_stat_item or node_stat_item. |
---|
| 722 | + * Keep in sync with memcg_exact_page_state(). |
---|
| 723 | + */ |
---|
| 724 | +static inline unsigned long memcg_page_state_local(struct mem_cgroup *memcg, |
---|
| 725 | + int idx) |
---|
580 | 726 | { |
---|
581 | | - long x; |
---|
| 727 | + long x = 0; |
---|
| 728 | + int cpu; |
---|
582 | 729 | |
---|
583 | | - if (mem_cgroup_disabled()) |
---|
584 | | - return; |
---|
585 | | - |
---|
586 | | - x = val + __this_cpu_read(memcg->stat_cpu->count[idx]); |
---|
587 | | - if (unlikely(abs(x) > MEMCG_CHARGE_BATCH)) { |
---|
588 | | - atomic_long_add(x, &memcg->stat[idx]); |
---|
| 730 | + for_each_possible_cpu(cpu) |
---|
| 731 | + x += per_cpu(memcg->vmstats_local->stat[idx], cpu); |
---|
| 732 | +#ifdef CONFIG_SMP |
---|
| 733 | + if (x < 0) |
---|
589 | 734 | x = 0; |
---|
590 | | - } |
---|
591 | | - __this_cpu_write(memcg->stat_cpu->count[idx], x); |
---|
| 735 | +#endif |
---|
| 736 | + return x; |
---|
592 | 737 | } |
---|
| 738 | + |
---|
| 739 | +void __mod_memcg_state(struct mem_cgroup *memcg, int idx, int val); |
---|
593 | 740 | |
---|
594 | 741 | /* idx can be of type enum memcg_stat_item or node_stat_item */ |
---|
595 | 742 | static inline void mod_memcg_state(struct mem_cgroup *memcg, |
---|
.. | .. |
---|
651 | 798 | return x; |
---|
652 | 799 | } |
---|
653 | 800 | |
---|
654 | | -static inline void __mod_lruvec_state(struct lruvec *lruvec, |
---|
655 | | - enum node_stat_item idx, int val) |
---|
| 801 | +static inline unsigned long lruvec_page_state_local(struct lruvec *lruvec, |
---|
| 802 | + enum node_stat_item idx) |
---|
656 | 803 | { |
---|
657 | 804 | struct mem_cgroup_per_node *pn; |
---|
658 | | - long x; |
---|
659 | | - |
---|
660 | | - /* Update node */ |
---|
661 | | - __mod_node_page_state(lruvec_pgdat(lruvec), idx, val); |
---|
| 805 | + long x = 0; |
---|
| 806 | + int cpu; |
---|
662 | 807 | |
---|
663 | 808 | if (mem_cgroup_disabled()) |
---|
664 | | - return; |
---|
| 809 | + return node_page_state(lruvec_pgdat(lruvec), idx); |
---|
665 | 810 | |
---|
666 | 811 | pn = container_of(lruvec, struct mem_cgroup_per_node, lruvec); |
---|
667 | | - |
---|
668 | | - /* Update memcg */ |
---|
669 | | - __mod_memcg_state(pn->memcg, idx, val); |
---|
670 | | - |
---|
671 | | - /* Update lruvec */ |
---|
672 | | - x = val + __this_cpu_read(pn->lruvec_stat_cpu->count[idx]); |
---|
673 | | - if (unlikely(abs(x) > MEMCG_CHARGE_BATCH)) { |
---|
674 | | - atomic_long_add(x, &pn->lruvec_stat[idx]); |
---|
| 812 | + for_each_possible_cpu(cpu) |
---|
| 813 | + x += per_cpu(pn->lruvec_stat_local->count[idx], cpu); |
---|
| 814 | +#ifdef CONFIG_SMP |
---|
| 815 | + if (x < 0) |
---|
675 | 816 | x = 0; |
---|
676 | | - } |
---|
677 | | - __this_cpu_write(pn->lruvec_stat_cpu->count[idx], x); |
---|
| 817 | +#endif |
---|
| 818 | + return x; |
---|
| 819 | +} |
---|
| 820 | + |
---|
| 821 | +void __mod_memcg_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, |
---|
| 822 | + int val); |
---|
| 823 | +void __mod_lruvec_state(struct lruvec *lruvec, enum node_stat_item idx, |
---|
| 824 | + int val); |
---|
| 825 | +void __mod_lruvec_slab_state(void *p, enum node_stat_item idx, int val); |
---|
| 826 | + |
---|
| 827 | +void mod_memcg_obj_state(void *p, int idx, int val); |
---|
| 828 | + |
---|
| 829 | +static inline void mod_lruvec_slab_state(void *p, enum node_stat_item idx, |
---|
| 830 | + int val) |
---|
| 831 | +{ |
---|
| 832 | + unsigned long flags; |
---|
| 833 | + |
---|
| 834 | + local_irq_save(flags); |
---|
| 835 | + __mod_lruvec_slab_state(p, idx, val); |
---|
| 836 | + local_irq_restore(flags); |
---|
| 837 | +} |
---|
| 838 | + |
---|
| 839 | +static inline void mod_memcg_lruvec_state(struct lruvec *lruvec, |
---|
| 840 | + enum node_stat_item idx, int val) |
---|
| 841 | +{ |
---|
| 842 | + unsigned long flags; |
---|
| 843 | + |
---|
| 844 | + local_irq_save(flags); |
---|
| 845 | + __mod_memcg_lruvec_state(lruvec, idx, val); |
---|
| 846 | + local_irq_restore(flags); |
---|
678 | 847 | } |
---|
679 | 848 | |
---|
680 | 849 | static inline void mod_lruvec_state(struct lruvec *lruvec, |
---|
.. | .. |
---|
690 | 859 | static inline void __mod_lruvec_page_state(struct page *page, |
---|
691 | 860 | enum node_stat_item idx, int val) |
---|
692 | 861 | { |
---|
| 862 | + struct page *head = compound_head(page); /* rmap on tail pages */ |
---|
693 | 863 | pg_data_t *pgdat = page_pgdat(page); |
---|
694 | 864 | struct lruvec *lruvec; |
---|
695 | 865 | |
---|
696 | 866 | /* Untracked pages have no memcg, no lruvec. Update only the node */ |
---|
697 | | - if (!page->mem_cgroup) { |
---|
| 867 | + if (!head->mem_cgroup) { |
---|
698 | 868 | __mod_node_page_state(pgdat, idx, val); |
---|
699 | 869 | return; |
---|
700 | 870 | } |
---|
701 | 871 | |
---|
702 | | - lruvec = mem_cgroup_lruvec(pgdat, page->mem_cgroup); |
---|
| 872 | + lruvec = mem_cgroup_lruvec(head->mem_cgroup, pgdat); |
---|
703 | 873 | __mod_lruvec_state(lruvec, idx, val); |
---|
704 | 874 | } |
---|
705 | 875 | |
---|
.. | .. |
---|
717 | 887 | gfp_t gfp_mask, |
---|
718 | 888 | unsigned long *total_scanned); |
---|
719 | 889 | |
---|
720 | | -static inline void __count_memcg_events(struct mem_cgroup *memcg, |
---|
721 | | - enum vm_event_item idx, |
---|
722 | | - unsigned long count) |
---|
723 | | -{ |
---|
724 | | - unsigned long x; |
---|
725 | | - |
---|
726 | | - if (mem_cgroup_disabled()) |
---|
727 | | - return; |
---|
728 | | - |
---|
729 | | - x = count + __this_cpu_read(memcg->stat_cpu->events[idx]); |
---|
730 | | - if (unlikely(x > MEMCG_CHARGE_BATCH)) { |
---|
731 | | - atomic_long_add(x, &memcg->events[idx]); |
---|
732 | | - x = 0; |
---|
733 | | - } |
---|
734 | | - __this_cpu_write(memcg->stat_cpu->events[idx], x); |
---|
735 | | -} |
---|
| 890 | +void __count_memcg_events(struct mem_cgroup *memcg, enum vm_event_item idx, |
---|
| 891 | + unsigned long count); |
---|
736 | 892 | |
---|
737 | 893 | static inline void count_memcg_events(struct mem_cgroup *memcg, |
---|
738 | 894 | enum vm_event_item idx, |
---|
.. | .. |
---|
770 | 926 | static inline void memcg_memory_event(struct mem_cgroup *memcg, |
---|
771 | 927 | enum memcg_memory_event event) |
---|
772 | 928 | { |
---|
773 | | - atomic_long_inc(&memcg->memory_events[event]); |
---|
774 | | - cgroup_file_notify(&memcg->events_file); |
---|
| 929 | + bool swap_event = event == MEMCG_SWAP_HIGH || event == MEMCG_SWAP_MAX || |
---|
| 930 | + event == MEMCG_SWAP_FAIL; |
---|
| 931 | + |
---|
| 932 | + atomic_long_inc(&memcg->memory_events_local[event]); |
---|
| 933 | + if (!swap_event) |
---|
| 934 | + cgroup_file_notify(&memcg->events_local_file); |
---|
| 935 | + |
---|
| 936 | + do { |
---|
| 937 | + atomic_long_inc(&memcg->memory_events[event]); |
---|
| 938 | + if (swap_event) |
---|
| 939 | + cgroup_file_notify(&memcg->swap_events_file); |
---|
| 940 | + else |
---|
| 941 | + cgroup_file_notify(&memcg->events_file); |
---|
| 942 | + |
---|
| 943 | + if (!cgroup_subsys_on_dfl(memory_cgrp_subsys)) |
---|
| 944 | + break; |
---|
| 945 | + if (cgrp_dfl_root.flags & CGRP_ROOT_MEMORY_LOCAL_EVENTS) |
---|
| 946 | + break; |
---|
| 947 | + } while ((memcg = parent_mem_cgroup(memcg)) && |
---|
| 948 | + !mem_cgroup_is_root(memcg)); |
---|
775 | 949 | } |
---|
776 | 950 | |
---|
777 | 951 | static inline void memcg_memory_event_mm(struct mm_struct *mm, |
---|
.. | .. |
---|
789 | 963 | rcu_read_unlock(); |
---|
790 | 964 | } |
---|
791 | 965 | |
---|
792 | | -#ifdef CONFIG_TRANSPARENT_HUGEPAGE |
---|
793 | | -void mem_cgroup_split_huge_fixup(struct page *head); |
---|
794 | | -#endif |
---|
| 966 | +void split_page_memcg(struct page *head, unsigned int nr); |
---|
795 | 967 | |
---|
796 | 968 | #else /* CONFIG_MEMCG */ |
---|
797 | 969 | |
---|
.. | .. |
---|
799 | 971 | #define MEM_CGROUP_ID_MAX 0 |
---|
800 | 972 | |
---|
801 | 973 | struct mem_cgroup; |
---|
| 974 | + |
---|
| 975 | +static inline struct lruvec *page_to_lruvec(struct page *page, pg_data_t *pgdat) |
---|
| 976 | +{ |
---|
| 977 | + return NULL; |
---|
| 978 | +} |
---|
| 979 | + |
---|
| 980 | +static inline void do_traversal_all_lruvec(void) |
---|
| 981 | +{ |
---|
| 982 | +} |
---|
802 | 983 | |
---|
803 | 984 | static inline bool mem_cgroup_is_root(struct mem_cgroup *memcg) |
---|
804 | 985 | { |
---|
.. | .. |
---|
820 | 1001 | { |
---|
821 | 1002 | } |
---|
822 | 1003 | |
---|
823 | | -static inline enum mem_cgroup_protection mem_cgroup_protected( |
---|
824 | | - struct mem_cgroup *root, struct mem_cgroup *memcg) |
---|
| 1004 | +static inline void mem_cgroup_protection(struct mem_cgroup *root, |
---|
| 1005 | + struct mem_cgroup *memcg, |
---|
| 1006 | + unsigned long *min, |
---|
| 1007 | + unsigned long *low) |
---|
825 | 1008 | { |
---|
826 | | - return MEMCG_PROT_NONE; |
---|
| 1009 | + *min = *low = 0; |
---|
827 | 1010 | } |
---|
828 | 1011 | |
---|
829 | | -static inline int mem_cgroup_try_charge(struct page *page, struct mm_struct *mm, |
---|
830 | | - gfp_t gfp_mask, |
---|
831 | | - struct mem_cgroup **memcgp, |
---|
832 | | - bool compound) |
---|
| 1012 | +static inline void mem_cgroup_calculate_protection(struct mem_cgroup *root, |
---|
| 1013 | + struct mem_cgroup *memcg) |
---|
833 | 1014 | { |
---|
834 | | - *memcgp = NULL; |
---|
| 1015 | +} |
---|
| 1016 | + |
---|
| 1017 | +static inline bool mem_cgroup_below_low(struct mem_cgroup *memcg) |
---|
| 1018 | +{ |
---|
| 1019 | + return false; |
---|
| 1020 | +} |
---|
| 1021 | + |
---|
| 1022 | +static inline bool mem_cgroup_below_min(struct mem_cgroup *memcg) |
---|
| 1023 | +{ |
---|
| 1024 | + return false; |
---|
| 1025 | +} |
---|
| 1026 | + |
---|
| 1027 | +static inline int mem_cgroup_charge(struct page *page, struct mm_struct *mm, |
---|
| 1028 | + gfp_t gfp_mask) |
---|
| 1029 | +{ |
---|
835 | 1030 | return 0; |
---|
836 | | -} |
---|
837 | | - |
---|
838 | | -static inline int mem_cgroup_try_charge_delay(struct page *page, |
---|
839 | | - struct mm_struct *mm, |
---|
840 | | - gfp_t gfp_mask, |
---|
841 | | - struct mem_cgroup **memcgp, |
---|
842 | | - bool compound) |
---|
843 | | -{ |
---|
844 | | - *memcgp = NULL; |
---|
845 | | - return 0; |
---|
846 | | -} |
---|
847 | | - |
---|
848 | | -static inline void mem_cgroup_commit_charge(struct page *page, |
---|
849 | | - struct mem_cgroup *memcg, |
---|
850 | | - bool lrucare, bool compound) |
---|
851 | | -{ |
---|
852 | | -} |
---|
853 | | - |
---|
854 | | -static inline void mem_cgroup_cancel_charge(struct page *page, |
---|
855 | | - struct mem_cgroup *memcg, |
---|
856 | | - bool compound) |
---|
857 | | -{ |
---|
858 | 1031 | } |
---|
859 | 1032 | |
---|
860 | 1033 | static inline void mem_cgroup_uncharge(struct page *page) |
---|
.. | .. |
---|
869 | 1042 | { |
---|
870 | 1043 | } |
---|
871 | 1044 | |
---|
872 | | -static inline struct lruvec *mem_cgroup_lruvec(struct pglist_data *pgdat, |
---|
873 | | - struct mem_cgroup *memcg) |
---|
| 1045 | +static inline struct lruvec *mem_cgroup_lruvec(struct mem_cgroup *memcg, |
---|
| 1046 | + struct pglist_data *pgdat) |
---|
874 | 1047 | { |
---|
875 | | - return node_lruvec(pgdat); |
---|
| 1048 | + return &pgdat->__lruvec; |
---|
876 | 1049 | } |
---|
877 | 1050 | |
---|
878 | 1051 | static inline struct lruvec *mem_cgroup_page_lruvec(struct page *page, |
---|
879 | 1052 | struct pglist_data *pgdat) |
---|
880 | 1053 | { |
---|
881 | | - return &pgdat->lruvec; |
---|
| 1054 | + return &pgdat->__lruvec; |
---|
| 1055 | +} |
---|
| 1056 | + |
---|
| 1057 | +static inline struct mem_cgroup *parent_mem_cgroup(struct mem_cgroup *memcg) |
---|
| 1058 | +{ |
---|
| 1059 | + return NULL; |
---|
882 | 1060 | } |
---|
883 | 1061 | |
---|
884 | 1062 | static inline bool mm_match_cgroup(struct mm_struct *mm, |
---|
885 | 1063 | struct mem_cgroup *memcg) |
---|
886 | | -{ |
---|
887 | | - return true; |
---|
888 | | -} |
---|
889 | | - |
---|
890 | | -static inline bool task_in_mem_cgroup(struct task_struct *task, |
---|
891 | | - const struct mem_cgroup *memcg) |
---|
892 | 1064 | { |
---|
893 | 1065 | return true; |
---|
894 | 1066 | } |
---|
.. | .. |
---|
938 | 1110 | return NULL; |
---|
939 | 1111 | } |
---|
940 | 1112 | |
---|
| 1113 | +static inline struct mem_cgroup *mem_cgroup_from_seq(struct seq_file *m) |
---|
| 1114 | +{ |
---|
| 1115 | + return NULL; |
---|
| 1116 | +} |
---|
| 1117 | + |
---|
941 | 1118 | static inline struct mem_cgroup *lruvec_memcg(struct lruvec *lruvec) |
---|
942 | 1119 | { |
---|
943 | 1120 | return NULL; |
---|
.. | .. |
---|
948 | 1125 | return true; |
---|
949 | 1126 | } |
---|
950 | 1127 | |
---|
951 | | -static inline unsigned long |
---|
952 | | -mem_cgroup_get_lru_size(struct lruvec *lruvec, enum lru_list lru) |
---|
953 | | -{ |
---|
954 | | - return 0; |
---|
955 | | -} |
---|
956 | 1128 | static inline |
---|
957 | 1129 | unsigned long mem_cgroup_get_zone_lru_size(struct lruvec *lruvec, |
---|
958 | 1130 | enum lru_list lru, int zone_idx) |
---|
959 | | -{ |
---|
960 | | - return 0; |
---|
961 | | -} |
---|
962 | | - |
---|
963 | | -static inline unsigned long |
---|
964 | | -mem_cgroup_node_nr_lru_pages(struct mem_cgroup *memcg, |
---|
965 | | - int nid, unsigned int lru_mask) |
---|
966 | 1131 | { |
---|
967 | 1132 | return 0; |
---|
968 | 1133 | } |
---|
.. | .. |
---|
972 | 1137 | return 0; |
---|
973 | 1138 | } |
---|
974 | 1139 | |
---|
| 1140 | +static inline unsigned long mem_cgroup_size(struct mem_cgroup *memcg) |
---|
| 1141 | +{ |
---|
| 1142 | + return 0; |
---|
| 1143 | +} |
---|
| 1144 | + |
---|
975 | 1145 | static inline void |
---|
976 | | -mem_cgroup_print_oom_info(struct mem_cgroup *memcg, struct task_struct *p) |
---|
| 1146 | +mem_cgroup_print_oom_context(struct mem_cgroup *memcg, struct task_struct *p) |
---|
| 1147 | +{ |
---|
| 1148 | +} |
---|
| 1149 | + |
---|
| 1150 | +static inline void |
---|
| 1151 | +mem_cgroup_print_oom_meminfo(struct mem_cgroup *memcg) |
---|
977 | 1152 | { |
---|
978 | 1153 | } |
---|
979 | 1154 | |
---|
.. | .. |
---|
1022 | 1197 | { |
---|
1023 | 1198 | } |
---|
1024 | 1199 | |
---|
1025 | | -static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, |
---|
1026 | | - int idx) |
---|
| 1200 | +static inline unsigned long memcg_page_state(struct mem_cgroup *memcg, int idx) |
---|
| 1201 | +{ |
---|
| 1202 | + return 0; |
---|
| 1203 | +} |
---|
| 1204 | + |
---|
| 1205 | +static inline unsigned long memcg_page_state_local(struct mem_cgroup *memcg, |
---|
| 1206 | + int idx) |
---|
1027 | 1207 | { |
---|
1028 | 1208 | return 0; |
---|
1029 | 1209 | } |
---|
.. | .. |
---|
1058 | 1238 | return node_page_state(lruvec_pgdat(lruvec), idx); |
---|
1059 | 1239 | } |
---|
1060 | 1240 | |
---|
| 1241 | +static inline unsigned long lruvec_page_state_local(struct lruvec *lruvec, |
---|
| 1242 | + enum node_stat_item idx) |
---|
| 1243 | +{ |
---|
| 1244 | + return node_page_state(lruvec_pgdat(lruvec), idx); |
---|
| 1245 | +} |
---|
| 1246 | + |
---|
| 1247 | +static inline void __mod_memcg_lruvec_state(struct lruvec *lruvec, |
---|
| 1248 | + enum node_stat_item idx, int val) |
---|
| 1249 | +{ |
---|
| 1250 | +} |
---|
| 1251 | + |
---|
1061 | 1252 | static inline void __mod_lruvec_state(struct lruvec *lruvec, |
---|
1062 | 1253 | enum node_stat_item idx, int val) |
---|
1063 | 1254 | { |
---|
.. | .. |
---|
1082 | 1273 | mod_node_page_state(page_pgdat(page), idx, val); |
---|
1083 | 1274 | } |
---|
1084 | 1275 | |
---|
| 1276 | +static inline void __mod_lruvec_slab_state(void *p, enum node_stat_item idx, |
---|
| 1277 | + int val) |
---|
| 1278 | +{ |
---|
| 1279 | + struct page *page = virt_to_head_page(p); |
---|
| 1280 | + |
---|
| 1281 | + __mod_node_page_state(page_pgdat(page), idx, val); |
---|
| 1282 | +} |
---|
| 1283 | + |
---|
| 1284 | +static inline void mod_lruvec_slab_state(void *p, enum node_stat_item idx, |
---|
| 1285 | + int val) |
---|
| 1286 | +{ |
---|
| 1287 | + struct page *page = virt_to_head_page(p); |
---|
| 1288 | + |
---|
| 1289 | + mod_node_page_state(page_pgdat(page), idx, val); |
---|
| 1290 | +} |
---|
| 1291 | + |
---|
| 1292 | +static inline void mod_memcg_obj_state(void *p, int idx, int val) |
---|
| 1293 | +{ |
---|
| 1294 | +} |
---|
| 1295 | + |
---|
1085 | 1296 | static inline |
---|
1086 | 1297 | unsigned long mem_cgroup_soft_limit_reclaim(pg_data_t *pgdat, int order, |
---|
1087 | 1298 | gfp_t gfp_mask, |
---|
.. | .. |
---|
1090 | 1301 | return 0; |
---|
1091 | 1302 | } |
---|
1092 | 1303 | |
---|
1093 | | -static inline void mem_cgroup_split_huge_fixup(struct page *head) |
---|
| 1304 | +static inline void split_page_memcg(struct page *head, unsigned int nr) |
---|
1094 | 1305 | { |
---|
1095 | 1306 | } |
---|
1096 | 1307 | |
---|
1097 | 1308 | static inline void count_memcg_events(struct mem_cgroup *memcg, |
---|
1098 | 1309 | enum vm_event_item idx, |
---|
1099 | 1310 | unsigned long count) |
---|
| 1311 | +{ |
---|
| 1312 | +} |
---|
| 1313 | + |
---|
| 1314 | +static inline void __count_memcg_events(struct mem_cgroup *memcg, |
---|
| 1315 | + enum vm_event_item idx, |
---|
| 1316 | + unsigned long count) |
---|
1100 | 1317 | { |
---|
1101 | 1318 | } |
---|
1102 | 1319 | |
---|
.. | .. |
---|
1163 | 1380 | __mod_lruvec_page_state(page, idx, -1); |
---|
1164 | 1381 | } |
---|
1165 | 1382 | |
---|
| 1383 | +static inline void __inc_lruvec_slab_state(void *p, enum node_stat_item idx) |
---|
| 1384 | +{ |
---|
| 1385 | + __mod_lruvec_slab_state(p, idx, 1); |
---|
| 1386 | +} |
---|
| 1387 | + |
---|
| 1388 | +static inline void __dec_lruvec_slab_state(void *p, enum node_stat_item idx) |
---|
| 1389 | +{ |
---|
| 1390 | + __mod_lruvec_slab_state(p, idx, -1); |
---|
| 1391 | +} |
---|
| 1392 | + |
---|
1166 | 1393 | /* idx can be of type enum memcg_stat_item or node_stat_item */ |
---|
1167 | 1394 | static inline void inc_memcg_state(struct mem_cgroup *memcg, |
---|
1168 | 1395 | int idx) |
---|
.. | .. |
---|
1215 | 1442 | mod_lruvec_page_state(page, idx, -1); |
---|
1216 | 1443 | } |
---|
1217 | 1444 | |
---|
| 1445 | +static inline struct lruvec *parent_lruvec(struct lruvec *lruvec) |
---|
| 1446 | +{ |
---|
| 1447 | + struct mem_cgroup *memcg; |
---|
| 1448 | + |
---|
| 1449 | + memcg = lruvec_memcg(lruvec); |
---|
| 1450 | + if (!memcg) |
---|
| 1451 | + return NULL; |
---|
| 1452 | + memcg = parent_mem_cgroup(memcg); |
---|
| 1453 | + if (!memcg) |
---|
| 1454 | + return NULL; |
---|
| 1455 | + return mem_cgroup_lruvec(memcg, lruvec_pgdat(lruvec)); |
---|
| 1456 | +} |
---|
| 1457 | + |
---|
1218 | 1458 | #ifdef CONFIG_CGROUP_WRITEBACK |
---|
1219 | 1459 | |
---|
1220 | 1460 | struct wb_domain *mem_cgroup_wb_domain(struct bdi_writeback *wb); |
---|
1221 | 1461 | void mem_cgroup_wb_stats(struct bdi_writeback *wb, unsigned long *pfilepages, |
---|
1222 | 1462 | unsigned long *pheadroom, unsigned long *pdirty, |
---|
1223 | 1463 | unsigned long *pwriteback); |
---|
| 1464 | + |
---|
| 1465 | +void mem_cgroup_track_foreign_dirty_slowpath(struct page *page, |
---|
| 1466 | + struct bdi_writeback *wb); |
---|
| 1467 | + |
---|
| 1468 | +static inline void mem_cgroup_track_foreign_dirty(struct page *page, |
---|
| 1469 | + struct bdi_writeback *wb) |
---|
| 1470 | +{ |
---|
| 1471 | + if (mem_cgroup_disabled()) |
---|
| 1472 | + return; |
---|
| 1473 | + |
---|
| 1474 | + if (unlikely(&page->mem_cgroup->css != wb->memcg_css)) |
---|
| 1475 | + mem_cgroup_track_foreign_dirty_slowpath(page, wb); |
---|
| 1476 | +} |
---|
| 1477 | + |
---|
| 1478 | +void mem_cgroup_flush_foreign(struct bdi_writeback *wb); |
---|
1224 | 1479 | |
---|
1225 | 1480 | #else /* CONFIG_CGROUP_WRITEBACK */ |
---|
1226 | 1481 | |
---|
.. | .. |
---|
1234 | 1489 | unsigned long *pheadroom, |
---|
1235 | 1490 | unsigned long *pdirty, |
---|
1236 | 1491 | unsigned long *pwriteback) |
---|
| 1492 | +{ |
---|
| 1493 | +} |
---|
| 1494 | + |
---|
| 1495 | +static inline void mem_cgroup_track_foreign_dirty(struct page *page, |
---|
| 1496 | + struct bdi_writeback *wb) |
---|
| 1497 | +{ |
---|
| 1498 | +} |
---|
| 1499 | + |
---|
| 1500 | +static inline void mem_cgroup_flush_foreign(struct bdi_writeback *wb) |
---|
1237 | 1501 | { |
---|
1238 | 1502 | } |
---|
1239 | 1503 | |
---|
.. | .. |
---|
1257 | 1521 | } while ((memcg = parent_mem_cgroup(memcg))); |
---|
1258 | 1522 | return false; |
---|
1259 | 1523 | } |
---|
| 1524 | + |
---|
| 1525 | +extern int memcg_expand_shrinker_maps(int new_id); |
---|
| 1526 | + |
---|
| 1527 | +extern void memcg_set_shrinker_bit(struct mem_cgroup *memcg, |
---|
| 1528 | + int nid, int shrinker_id); |
---|
1260 | 1529 | #else |
---|
1261 | 1530 | #define mem_cgroup_sockets_enabled 0 |
---|
1262 | 1531 | static inline void mem_cgroup_sk_alloc(struct sock *sk) { }; |
---|
.. | .. |
---|
1265 | 1534 | { |
---|
1266 | 1535 | return false; |
---|
1267 | 1536 | } |
---|
| 1537 | + |
---|
| 1538 | +static inline void memcg_set_shrinker_bit(struct mem_cgroup *memcg, |
---|
| 1539 | + int nid, int shrinker_id) |
---|
| 1540 | +{ |
---|
| 1541 | +} |
---|
1268 | 1542 | #endif |
---|
1269 | 1543 | |
---|
1270 | | -struct kmem_cache *memcg_kmem_get_cache(struct kmem_cache *cachep); |
---|
1271 | | -void memcg_kmem_put_cache(struct kmem_cache *cachep); |
---|
1272 | | -int memcg_kmem_charge_memcg(struct page *page, gfp_t gfp, int order, |
---|
1273 | | - struct mem_cgroup *memcg); |
---|
1274 | | -int memcg_kmem_charge(struct page *page, gfp_t gfp, int order); |
---|
1275 | | -void memcg_kmem_uncharge(struct page *page, int order); |
---|
1276 | | - |
---|
1277 | 1544 | #ifdef CONFIG_MEMCG_KMEM |
---|
| 1545 | +int __memcg_kmem_charge(struct mem_cgroup *memcg, gfp_t gfp, |
---|
| 1546 | + unsigned int nr_pages); |
---|
| 1547 | +void __memcg_kmem_uncharge(struct mem_cgroup *memcg, unsigned int nr_pages); |
---|
| 1548 | +int __memcg_kmem_charge_page(struct page *page, gfp_t gfp, int order); |
---|
| 1549 | +void __memcg_kmem_uncharge_page(struct page *page, int order); |
---|
| 1550 | + |
---|
| 1551 | +struct obj_cgroup *get_obj_cgroup_from_current(void); |
---|
| 1552 | + |
---|
| 1553 | +int obj_cgroup_charge(struct obj_cgroup *objcg, gfp_t gfp, size_t size); |
---|
| 1554 | +void obj_cgroup_uncharge(struct obj_cgroup *objcg, size_t size); |
---|
| 1555 | + |
---|
1278 | 1556 | extern struct static_key_false memcg_kmem_enabled_key; |
---|
1279 | | -extern struct workqueue_struct *memcg_kmem_cache_wq; |
---|
1280 | 1557 | |
---|
1281 | 1558 | extern int memcg_nr_cache_ids; |
---|
1282 | 1559 | void memcg_get_cache_ids(void); |
---|
.. | .. |
---|
1292 | 1569 | |
---|
1293 | 1570 | static inline bool memcg_kmem_enabled(void) |
---|
1294 | 1571 | { |
---|
1295 | | - return static_branch_unlikely(&memcg_kmem_enabled_key); |
---|
| 1572 | + return static_branch_likely(&memcg_kmem_enabled_key); |
---|
| 1573 | +} |
---|
| 1574 | + |
---|
| 1575 | +static inline int memcg_kmem_charge_page(struct page *page, gfp_t gfp, |
---|
| 1576 | + int order) |
---|
| 1577 | +{ |
---|
| 1578 | + if (memcg_kmem_enabled()) |
---|
| 1579 | + return __memcg_kmem_charge_page(page, gfp, order); |
---|
| 1580 | + return 0; |
---|
| 1581 | +} |
---|
| 1582 | + |
---|
| 1583 | +static inline void memcg_kmem_uncharge_page(struct page *page, int order) |
---|
| 1584 | +{ |
---|
| 1585 | + if (memcg_kmem_enabled()) |
---|
| 1586 | + __memcg_kmem_uncharge_page(page, order); |
---|
| 1587 | +} |
---|
| 1588 | + |
---|
| 1589 | +static inline int memcg_kmem_charge(struct mem_cgroup *memcg, gfp_t gfp, |
---|
| 1590 | + unsigned int nr_pages) |
---|
| 1591 | +{ |
---|
| 1592 | + if (memcg_kmem_enabled()) |
---|
| 1593 | + return __memcg_kmem_charge(memcg, gfp, nr_pages); |
---|
| 1594 | + return 0; |
---|
| 1595 | +} |
---|
| 1596 | + |
---|
| 1597 | +static inline void memcg_kmem_uncharge(struct mem_cgroup *memcg, |
---|
| 1598 | + unsigned int nr_pages) |
---|
| 1599 | +{ |
---|
| 1600 | + if (memcg_kmem_enabled()) |
---|
| 1601 | + __memcg_kmem_uncharge(memcg, nr_pages); |
---|
1296 | 1602 | } |
---|
1297 | 1603 | |
---|
1298 | 1604 | /* |
---|
.. | .. |
---|
1305 | 1611 | return memcg ? memcg->kmemcg_id : -1; |
---|
1306 | 1612 | } |
---|
1307 | 1613 | |
---|
1308 | | -extern int memcg_expand_shrinker_maps(int new_id); |
---|
| 1614 | +struct mem_cgroup *mem_cgroup_from_obj(void *p); |
---|
1309 | 1615 | |
---|
1310 | | -extern void memcg_set_shrinker_bit(struct mem_cgroup *memcg, |
---|
1311 | | - int nid, int shrinker_id); |
---|
1312 | 1616 | #else |
---|
| 1617 | + |
---|
| 1618 | +static inline int memcg_kmem_charge_page(struct page *page, gfp_t gfp, |
---|
| 1619 | + int order) |
---|
| 1620 | +{ |
---|
| 1621 | + return 0; |
---|
| 1622 | +} |
---|
| 1623 | + |
---|
| 1624 | +static inline void memcg_kmem_uncharge_page(struct page *page, int order) |
---|
| 1625 | +{ |
---|
| 1626 | +} |
---|
| 1627 | + |
---|
| 1628 | +static inline int __memcg_kmem_charge_page(struct page *page, gfp_t gfp, |
---|
| 1629 | + int order) |
---|
| 1630 | +{ |
---|
| 1631 | + return 0; |
---|
| 1632 | +} |
---|
| 1633 | + |
---|
| 1634 | +static inline void __memcg_kmem_uncharge_page(struct page *page, int order) |
---|
| 1635 | +{ |
---|
| 1636 | +} |
---|
| 1637 | + |
---|
1313 | 1638 | #define for_each_memcg_cache_index(_idx) \ |
---|
1314 | 1639 | for (; NULL; ) |
---|
1315 | 1640 | |
---|
.. | .. |
---|
1331 | 1656 | { |
---|
1332 | 1657 | } |
---|
1333 | 1658 | |
---|
1334 | | -static inline void memcg_set_shrinker_bit(struct mem_cgroup *memcg, |
---|
1335 | | - int nid, int shrinker_id) { } |
---|
| 1659 | +static inline struct mem_cgroup *mem_cgroup_from_obj(void *p) |
---|
| 1660 | +{ |
---|
| 1661 | + return NULL; |
---|
| 1662 | +} |
---|
| 1663 | + |
---|
1336 | 1664 | #endif /* CONFIG_MEMCG_KMEM */ |
---|
1337 | 1665 | |
---|
1338 | 1666 | #endif /* _LINUX_MEMCONTROL_H */ |
---|