summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorAndrew Morton <akpm@linux-foundation.org>2024-04-29 11:21:14 -0700
committerAndrew Morton <akpm@linux-foundation.org>2024-04-29 11:21:14 -0700
commit4022ed83d8dafb36a2afdf25bbc615281ed59e31 (patch)
tree388f0f2e39a9121262412b48edc17820fd550e14
parentf6eb4258da2e239693ef9aa5d9d931c3413b1674 (diff)
download25-new-4022ed83d8dafb36a2afdf25bbc615281ed59e31.tar.gz
foo
-rw-r--r--patches/__mod_memcg_lruvec_state-enhance-diagnostics-fix.patch2
-rw-r--r--patches/memcg-cleanup-__mod_memcg_lruvec_state.patch2
-rw-r--r--patches/memcg-reduce-memory-size-of-mem_cgroup_events_index.patch2
-rw-r--r--patches/mm-add-defines-for-min-max-swappiness.patch2
-rw-r--r--patches/mm-add-swapiness=-arg-to-memoryreclaim.patch16
-rw-r--r--patches/mm-cleanup-workingset_nodes-in-workingset.patch2
-rw-r--r--patches/mm-pagemap-make-trylock_page-return-bool.patch2
-rw-r--r--patches/mm-remove-page_cache_alloc.patch2
-rw-r--r--patches/mm-rmap-do-not-add-fully-unmapped-large-folio-to-deferred-split-list.patch4
-rw-r--r--pc/devel-series2
10 files changed, 18 insertions, 18 deletions
diff --git a/patches/__mod_memcg_lruvec_state-enhance-diagnostics-fix.patch b/patches/__mod_memcg_lruvec_state-enhance-diagnostics-fix.patch
index 02f36b5d0..d7ec2a748 100644
--- a/patches/__mod_memcg_lruvec_state-enhance-diagnostics-fix.patch
+++ b/patches/__mod_memcg_lruvec_state-enhance-diagnostics-fix.patch
@@ -12,7 +12,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
--- a/mm/memcontrol.c~__mod_memcg_lruvec_state-enhance-diagnostics-fix
+++ a/mm/memcontrol.c
-@@ -867,7 +867,8 @@ static void __mod_memcg_lruvec_state(str
+@@ -1029,7 +1029,8 @@ static void __mod_memcg_lruvec_state(str
pr_warn("stat item index: %d\n", idx);
break;
default:
diff --git a/patches/memcg-cleanup-__mod_memcg_lruvec_state.patch b/patches/memcg-cleanup-__mod_memcg_lruvec_state.patch
index 70d309894..2f84aa814 100644
--- a/patches/memcg-cleanup-__mod_memcg_lruvec_state.patch
+++ b/patches/memcg-cleanup-__mod_memcg_lruvec_state.patch
@@ -8,10 +8,10 @@ NR_FILE_PMDMAPPED. Let's remove them.
Link: https://lkml.kernel.org/r/20240427003733.3898961-5-shakeel.butt@linux.dev
Signed-off-by: Shakeel Butt <shakeel.butt@linux.dev>
Reviewed-by: Yosry Ahmed <yosryahmed@google.com>
+Reviewed-by: Roman Gushchin <roman.gushchin@linux.dev>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Muchun Song <muchun.song@linux.dev>
-Cc: Roman Gushchin <roman.gushchin@linux.dev>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
diff --git a/patches/memcg-reduce-memory-size-of-mem_cgroup_events_index.patch b/patches/memcg-reduce-memory-size-of-mem_cgroup_events_index.patch
index 2e6ea26e9..cc0596c3d 100644
--- a/patches/memcg-reduce-memory-size-of-mem_cgroup_events_index.patch
+++ b/patches/memcg-reduce-memory-size-of-mem_cgroup_events_index.patch
@@ -29,10 +29,10 @@ BUILD_BUG_ON() and will switch to short once NR_VM_EVENT_ITEMS touches
Link: https://lkml.kernel.org/r/20240427003733.3898961-1-shakeel.butt@linux.dev
Link: https://lkml.kernel.org/r/20240427003733.3898961-2-shakeel.butt@linux.dev
Signed-off-by: Shakeel Butt <shakeel.butt@linux.dev>
+Reviewed-by: Roman Gushchin <roman.gushchin@linux.dev>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Muchun Song <muchun.song@linux.dev>
-Cc: Roman Gushchin <roman.gushchin@linux.dev>
Cc: Yosry Ahmed <yosryahmed@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
diff --git a/patches/mm-add-defines-for-min-max-swappiness.patch b/patches/mm-add-defines-for-min-max-swappiness.patch
index d3c49e1a5..49914cea9 100644
--- a/patches/mm-add-defines-for-min-max-swappiness.patch
+++ b/patches/mm-add-defines-for-min-max-swappiness.patch
@@ -130,7 +130,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
gfp_t gfp_mask,
--- a/mm/memcontrol.c~mm-add-defines-for-min-max-swappiness
+++ a/mm/memcontrol.c
-@@ -4431,7 +4431,7 @@ static int mem_cgroup_swappiness_write(s
+@@ -4603,7 +4603,7 @@ static int mem_cgroup_swappiness_write(s
{
struct mem_cgroup *memcg = mem_cgroup_from_css(css);
diff --git a/patches/mm-add-swapiness=-arg-to-memoryreclaim.patch b/patches/mm-add-swapiness=-arg-to-memoryreclaim.patch
index b460b3fe3..3d6aa6a41 100644
--- a/patches/mm-add-swapiness=-arg-to-memoryreclaim.patch
+++ b/patches/mm-add-swapiness=-arg-to-memoryreclaim.patch
@@ -144,7 +144,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
#include <linux/vmpressure.h>
#include <linux/memremap.h>
#include <linux/mm_inline.h>
-@@ -2477,7 +2478,8 @@ static unsigned long reclaim_high(struct
+@@ -2649,7 +2650,8 @@ static unsigned long reclaim_high(struct
psi_memstall_enter(&pflags);
nr_reclaimed += try_to_free_mem_cgroup_pages(memcg, nr_pages,
gfp_mask,
@@ -154,7 +154,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
psi_memstall_leave(&pflags);
} while ((memcg = parent_mem_cgroup(memcg)) &&
!mem_cgroup_is_root(memcg));
-@@ -2783,7 +2785,7 @@ retry:
+@@ -2955,7 +2957,7 @@ retry:
psi_memstall_enter(&pflags);
nr_reclaimed = try_to_free_mem_cgroup_pages(mem_over_limit, nr_pages,
@@ -163,7 +163,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
psi_memstall_leave(&pflags);
if (mem_cgroup_margin(mem_over_limit) >= nr_pages)
-@@ -3754,7 +3756,7 @@ static int mem_cgroup_resize_max(struct
+@@ -3926,7 +3928,7 @@ static int mem_cgroup_resize_max(struct
}
if (!try_to_free_mem_cgroup_pages(memcg, 1, GFP_KERNEL,
@@ -172,7 +172,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
ret = -EBUSY;
break;
}
-@@ -3868,7 +3870,7 @@ static int mem_cgroup_force_empty(struct
+@@ -4040,7 +4042,7 @@ static int mem_cgroup_force_empty(struct
return -EINTR;
if (!try_to_free_mem_cgroup_pages(memcg, 1, GFP_KERNEL,
@@ -181,7 +181,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
nr_retries--;
}
-@@ -6832,7 +6834,7 @@ static ssize_t memory_high_write(struct
+@@ -7014,7 +7016,7 @@ static ssize_t memory_high_write(struct
}
reclaimed = try_to_free_mem_cgroup_pages(memcg, nr_pages - high,
@@ -190,7 +190,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
if (!reclaimed && !nr_retries--)
break;
-@@ -6881,7 +6883,7 @@ static ssize_t memory_max_write(struct k
+@@ -7063,7 +7065,7 @@ static ssize_t memory_max_write(struct k
if (nr_reclaims) {
if (!try_to_free_mem_cgroup_pages(memcg, nr_pages - max,
@@ -199,7 +199,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
nr_reclaims--;
continue;
}
-@@ -7011,19 +7013,50 @@ static ssize_t memory_oom_group_write(st
+@@ -7193,19 +7195,50 @@ static ssize_t memory_oom_group_write(st
return nbytes;
}
@@ -254,7 +254,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
reclaim_options = MEMCG_RECLAIM_MAY_SWAP | MEMCG_RECLAIM_PROACTIVE;
while (nr_reclaimed < nr_to_reclaim) {
-@@ -7043,7 +7076,9 @@ static ssize_t memory_reclaim(struct ker
+@@ -7225,7 +7258,9 @@ static ssize_t memory_reclaim(struct ker
lru_add_drain_all();
reclaimed = try_to_free_mem_cgroup_pages(memcg,
diff --git a/patches/mm-cleanup-workingset_nodes-in-workingset.patch b/patches/mm-cleanup-workingset_nodes-in-workingset.patch
index e24fd9eae..8dad71032 100644
--- a/patches/mm-cleanup-workingset_nodes-in-workingset.patch
+++ b/patches/mm-cleanup-workingset_nodes-in-workingset.patch
@@ -9,10 +9,10 @@ this patch.
Link: https://lkml.kernel.org/r/20240427003733.3898961-8-shakeel.butt@linux.dev
Signed-off-by: Shakeel Butt <shakeel.butt@linux.dev>
+Reviewed-by: Roman Gushchin <roman.gushchin@linux.dev>
Cc: Johannes Weiner <hannes@cmpxchg.org>
Cc: Michal Hocko <mhocko@kernel.org>
Cc: Muchun Song <muchun.song@linux.dev>
-Cc: Roman Gushchin <roman.gushchin@linux.dev>
Cc: Yosry Ahmed <yosryahmed@google.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
---
diff --git a/patches/mm-pagemap-make-trylock_page-return-bool.patch b/patches/mm-pagemap-make-trylock_page-return-bool.patch
index 845e5d1dc..ddfac8d20 100644
--- a/patches/mm-pagemap-make-trylock_page-return-bool.patch
+++ b/patches/mm-pagemap-make-trylock_page-return-bool.patch
@@ -16,7 +16,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
--- a/include/linux/pagemap.h~mm-pagemap-make-trylock_page-return-bool
+++ a/include/linux/pagemap.h
-@@ -1011,7 +1011,7 @@ static inline bool folio_trylock(struct
+@@ -1012,7 +1012,7 @@ static inline bool folio_trylock(struct
/*
* Return true if the page was successfully locked
*/
diff --git a/patches/mm-remove-page_cache_alloc.patch b/patches/mm-remove-page_cache_alloc.patch
index af27330a7..ac65d874a 100644
--- a/patches/mm-remove-page_cache_alloc.patch
+++ b/patches/mm-remove-page_cache_alloc.patch
@@ -22,7 +22,7 @@ Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
--- a/include/linux/pagemap.h~mm-remove-page_cache_alloc
+++ a/include/linux/pagemap.h
-@@ -557,11 +557,6 @@ static inline struct page *__page_cache_
+@@ -558,11 +558,6 @@ static inline struct page *__page_cache_
return &filemap_alloc_folio(gfp, 0)->page;
}
diff --git a/patches/mm-rmap-do-not-add-fully-unmapped-large-folio-to-deferred-split-list.patch b/patches/mm-rmap-do-not-add-fully-unmapped-large-folio-to-deferred-split-list.patch
index aaef2d19c..cf1572626 100644
--- a/patches/mm-rmap-do-not-add-fully-unmapped-large-folio-to-deferred-split-list.patch
+++ b/patches/mm-rmap-do-not-add-fully-unmapped-large-folio-to-deferred-split-list.patch
@@ -42,8 +42,8 @@ Signed-off-by: Zi Yan <ziy@nvidia.com>
Suggested-by: David Hildenbrand <david@redhat.com>
Reviewed-by: Yang Shi <shy828301@gmail.com>
Reviewed-by: David Hildenbrand <david@redhat.com>
-Cc: Barry Song <21cnbao@gmail.com>
-Cc: Lance Yang <ioworker0@gmail.com>
+Reviewed-by: Barry Song <baohua@kernel.org>
+Reviewed-by: Lance Yang <ioworker0@gmail.com>
Cc: Matthew Wilcox (Oracle) <willy@infradead.org>
Cc: Ryan Roberts <ryan.roberts@arm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
diff --git a/pc/devel-series b/pc/devel-series
index e541e4acc..04198603b 100644
--- a/pc/devel-series
+++ b/pc/devel-series
@@ -743,7 +743,7 @@ mm-swapfile-mark-racy-access-on-si-highest_bit.patch
memcg-reduce-memory-size-of-mem_cgroup_events_index.patch
#memcg-dynamically-allocate-lruvec_stats.patch: https://lkml.kernel.org/r/Zi_BsznrtoC1N_lq@P9FQF9L96D
memcg-dynamically-allocate-lruvec_stats.patch
-#memcg-reduce-memory-for-the-lruvec-and-memcg-stats.patch: https://lkml.kernel.org/r/Zi_EEOUS_iCh2Nfh@P9FQF9L96D
+#memcg-reduce-memory-for-the-lruvec-and-memcg-stats.patch: https://lkml.kernel.org/r/Zi_EEOUS_iCh2Nfh@P9FQF9L96D https://lkml.kernel.org/r/CABdmKX0Mo74f-AjY46cyPwd2qSpwmj4CLvYVCEywq_PEsVmd9w@mail.gmail.com
memcg-reduce-memory-for-the-lruvec-and-memcg-stats.patch
memcg-cleanup-__mod_memcg_lruvec_state.patch
#memcg-pr_warn_once-for-unexpected-events-and-stats.patch: https://lkml.kernel.org/r/Zi_Ff_iAs_PTph4l@P9FQF9L96D