On 12.11.25 15:50, Deepanshu Kartikey wrote:
When allocating hugetlb folios for memfd, three initialization steps are missing:
- Folios are not zeroed, leading to kernel memory disclosure to userspace
- Folios are not marked uptodate before adding to page cache
- hugetlb_fault_mutex is not taken before hugetlb_add_to_page_cache()
The memfd allocation path bypasses the normal page fault handler (hugetlb_no_page) which would handle all of these initialization steps. This is problematic especially for udmabuf use cases where folios are pinned and directly accessed by userspace via DMA.
Fix by matching the initialization pattern used in hugetlb_no_page():
- Zero the folio using folio_zero_user() which is optimized for huge pages
- Mark it uptodate with folio_mark_uptodate()
- Take hugetlb_fault_mutex before adding to page cache to prevent races
The folio_zero_user() change also fixes a potential security issue where uninitialized kernel memory could be disclosed to userspace through read() or mmap() operations on the memfd.
Reported-by: syzbot+f64019ba229e3a5c411b@syzkaller.appspotmail.com Link: https://lore.kernel.org/all/20251112031631.2315651-1-kartikey406@gmail.com/ [v1] Closes: https://syzkaller.appspot.com/bug?extid=f64019ba229e3a5c411b Fixes: 89c1905d9c14 ("mm/gup: introduce memfd_pin_folios() for pinning memfd folios") Cc: stable@vger.kernel.org Suggested-by: Oscar Salvador osalvador@suse.de Suggested-by: David Hildenbrand david@redhat.com Tested-by: syzbot+f64019ba229e3a5c411b@syzkaller.appspotmail.com Signed-off-by: Deepanshu Kartikey kartikey406@gmail.com
v1 -> v2:
- Use folio_zero_user() instead of folio_zero_range() (optimized for huge pages)
- Add folio_mark_uptodate() before adding to page cache
- Add hugetlb_fault_mutex locking around hugetlb_add_to_page_cache()
- Add Fixes: tag and Cc: stable for backporting
- Add Suggested-by: tags for Oscar and David
mm/memfd.c | 27 +++++++++++++++++++++++++++ 1 file changed, 27 insertions(+)
diff --git a/mm/memfd.c b/mm/memfd.c index 1d109c1acf21..d32eef58d154 100644 --- a/mm/memfd.c +++ b/mm/memfd.c @@ -96,9 +96,36 @@ struct folio *memfd_alloc_folio(struct file *memfd, pgoff_t idx) NULL, gfp_mask); if (folio) {
u32 hash;/** Zero the folio to prevent information leaks to userspace.* Use folio_zero_user() which is optimized for huge/gigantic* pages. Pass 0 as addr_hint since this is not a faulting path* and we don't have a user virtual address yet.*/folio_zero_user(folio, 0);
Staring at hugetlbfs_fallocate(), we see, to pass the offset within the file.
I think it shouldn't make a difference here (I don't see how the offset in the file would be better than 0: it's in both cases not the user address).
/** Mark the folio uptodate before adding to page cache,* as required by filemap.c and other hugetlb paths.*/__folio_mark_uptodate(folio);
Personally, I'd drop this comment as it is really just doing what we do everywhere else :)
Hoping we can factor that out into hugetlb code properly.
Acked-by: David Hildenbrand (Red Hat) david@kernel.org