kernel/alloc.rs
1// SPDX-License-Identifier: GPL-2.0
2
3//! Implementation of the kernel's memory allocation infrastructure.
4
5pub mod allocator;
6pub mod kbox;
7pub mod kvec;
8pub mod layout;
9
10pub use self::kbox::Box;
11pub use self::kbox::KBox;
12pub use self::kbox::KVBox;
13pub use self::kbox::VBox;
14
15pub use self::kvec::IntoIter;
16pub use self::kvec::KVVec;
17pub use self::kvec::KVec;
18pub use self::kvec::VVec;
19pub use self::kvec::Vec;
20
21/// Indicates an allocation error.
22#[derive(Copy, Clone, PartialEq, Eq, Debug)]
23pub struct AllocError;
24
25use crate::error::{code::EINVAL, Result};
26use core::{alloc::Layout, ptr::NonNull};
27
28/// Flags to be used when allocating memory.
29///
30/// They can be combined with the operators `|`, `&`, and `!`.
31///
32/// Values can be used from the [`flags`] module.
33#[derive(Clone, Copy, PartialEq)]
34pub struct Flags(u32);
35
36impl Flags {
37 /// Get the raw representation of this flag.
38 pub(crate) fn as_raw(self) -> u32 {
39 self.0
40 }
41
42 /// Check whether `flags` is contained in `self`.
43 pub fn contains(self, flags: Flags) -> bool {
44 (self & flags) == flags
45 }
46}
47
48impl core::ops::BitOr for Flags {
49 type Output = Self;
50 fn bitor(self, rhs: Self) -> Self::Output {
51 Self(self.0 | rhs.0)
52 }
53}
54
55impl core::ops::BitAnd for Flags {
56 type Output = Self;
57 fn bitand(self, rhs: Self) -> Self::Output {
58 Self(self.0 & rhs.0)
59 }
60}
61
62impl core::ops::Not for Flags {
63 type Output = Self;
64 fn not(self) -> Self::Output {
65 Self(!self.0)
66 }
67}
68
69/// Allocation flags.
70///
71/// These are meant to be used in functions that can allocate memory.
72pub mod flags {
73 use super::Flags;
74
75 /// Zeroes out the allocated memory.
76 ///
77 /// This is normally or'd with other flags.
78 pub const __GFP_ZERO: Flags = Flags(bindings::__GFP_ZERO);
79
80 /// Allow the allocation to be in high memory.
81 ///
82 /// Allocations in high memory may not be mapped into the kernel's address space, so this can't
83 /// be used with `kmalloc` and other similar methods.
84 ///
85 /// This is normally or'd with other flags.
86 pub const __GFP_HIGHMEM: Flags = Flags(bindings::__GFP_HIGHMEM);
87
88 /// Users can not sleep and need the allocation to succeed.
89 ///
90 /// A lower watermark is applied to allow access to "atomic reserves". The current
91 /// implementation doesn't support NMI and few other strict non-preemptive contexts (e.g.
92 /// `raw_spin_lock`). The same applies to [`GFP_NOWAIT`].
93 pub const GFP_ATOMIC: Flags = Flags(bindings::GFP_ATOMIC);
94
95 /// Typical for kernel-internal allocations. The caller requires `ZONE_NORMAL` or a lower zone
96 /// for direct access but can direct reclaim.
97 pub const GFP_KERNEL: Flags = Flags(bindings::GFP_KERNEL);
98
99 /// The same as [`GFP_KERNEL`], except the allocation is accounted to kmemcg.
100 pub const GFP_KERNEL_ACCOUNT: Flags = Flags(bindings::GFP_KERNEL_ACCOUNT);
101
102 /// For kernel allocations that should not stall for direct reclaim, start physical IO or
103 /// use any filesystem callback. It is very likely to fail to allocate memory, even for very
104 /// small allocations.
105 pub const GFP_NOWAIT: Flags = Flags(bindings::GFP_NOWAIT);
106
107 /// Suppresses allocation failure reports.
108 ///
109 /// This is normally or'd with other flags.
110 pub const __GFP_NOWARN: Flags = Flags(bindings::__GFP_NOWARN);
111}
112
113/// Non Uniform Memory Access (NUMA) node identifier.
114#[derive(Clone, Copy, PartialEq)]
115pub struct NumaNode(i32);
116
117impl NumaNode {
118 /// Create a new NUMA node identifier (non-negative integer).
119 ///
120 /// Returns [`EINVAL`] if a negative id or an id exceeding [`bindings::MAX_NUMNODES`] is
121 /// specified.
122 pub fn new(node: i32) -> Result<Self> {
123 // MAX_NUMNODES never exceeds 2**10 because NODES_SHIFT is 0..10.
124 if node < 0 || node >= bindings::MAX_NUMNODES as i32 {
125 return Err(EINVAL);
126 }
127 Ok(Self(node))
128 }
129}
130
131/// Specify necessary constant to pass the information to Allocator that the caller doesn't care
132/// about the NUMA node to allocate memory from.
133impl NumaNode {
134 /// No node preference.
135 pub const NO_NODE: NumaNode = NumaNode(bindings::NUMA_NO_NODE);
136}
137
138/// The kernel's [`Allocator`] trait.
139///
140/// An implementation of [`Allocator`] can allocate, re-allocate and free memory buffers described
141/// via [`Layout`].
142///
143/// [`Allocator`] is designed to be implemented as a ZST; [`Allocator`] functions do not operate on
144/// an object instance.
145///
146/// In order to be able to support `#[derive(CoercePointee)]` later on, we need to avoid a design
147/// that requires an `Allocator` to be instantiated, hence its functions must not contain any kind
148/// of `self` parameter.
149///
150/// # Safety
151///
152/// - A memory allocation returned from an allocator must remain valid until it is explicitly freed.
153///
154/// - Any pointer to a valid memory allocation must be valid to be passed to any other [`Allocator`]
155/// function of the same type.
156///
157/// - Implementers must ensure that all trait functions abide by the guarantees documented in the
158/// `# Guarantees` sections.
159pub unsafe trait Allocator {
160 /// The minimum alignment satisfied by all allocations from this allocator.
161 ///
162 /// # Guarantees
163 ///
164 /// Any pointer allocated by this allocator is guaranteed to be aligned to `MIN_ALIGN` even if
165 /// the requested layout has a smaller alignment.
166 const MIN_ALIGN: usize;
167
168 /// Allocate memory based on `layout`, `flags` and `nid`.
169 ///
170 /// On success, returns a buffer represented as `NonNull<[u8]>` that satisfies the layout
171 /// constraints (i.e. minimum size and alignment as specified by `layout`).
172 ///
173 /// This function is equivalent to `realloc` when called with `None`.
174 ///
175 /// # Guarantees
176 ///
177 /// When the return value is `Ok(ptr)`, then `ptr` is
178 /// - valid for reads and writes for `layout.size()` bytes, until it is passed to
179 /// [`Allocator::free`] or [`Allocator::realloc`],
180 /// - aligned to `layout.align()`,
181 ///
182 /// Additionally, `Flags` are honored as documented in
183 /// <https://docs.kernel.org/core-api/mm-api.html#mm-api-gfp-flags>.
184 fn alloc(layout: Layout, flags: Flags, nid: NumaNode) -> Result<NonNull<[u8]>, AllocError> {
185 // SAFETY: Passing `None` to `realloc` is valid by its safety requirements and asks for a
186 // new memory allocation.
187 unsafe { Self::realloc(None, layout, Layout::new::<()>(), flags, nid) }
188 }
189
190 /// Re-allocate an existing memory allocation to satisfy the requested `layout` and
191 /// a specific NUMA node request to allocate the memory for.
192 ///
193 /// Systems employing a Non Uniform Memory Access (NUMA) architecture contain collections of
194 /// hardware resources including processors, memory, and I/O buses, that comprise what is
195 /// commonly known as a NUMA node.
196 ///
197 /// `nid` stands for NUMA id, i. e. NUMA node identifier, which is a non-negative integer
198 /// if a node needs to be specified, or [`NumaNode::NO_NODE`] if the caller doesn't care.
199 ///
200 /// If the requested size is zero, `realloc` behaves equivalent to `free`.
201 ///
202 /// If the requested size is larger than the size of the existing allocation, a successful call
203 /// to `realloc` guarantees that the new or grown buffer has at least `Layout::size` bytes, but
204 /// may also be larger.
205 ///
206 /// If the requested size is smaller than the size of the existing allocation, `realloc` may or
207 /// may not shrink the buffer; this is implementation specific to the allocator.
208 ///
209 /// On allocation failure, the existing buffer, if any, remains valid.
210 ///
211 /// The buffer is represented as `NonNull<[u8]>`.
212 ///
213 /// # Safety
214 ///
215 /// - If `ptr == Some(p)`, then `p` must point to an existing and valid memory allocation
216 /// created by this [`Allocator`]; if `old_layout` is zero-sized `p` does not need to be a
217 /// pointer returned by this [`Allocator`].
218 /// - `ptr` is allowed to be `None`; in this case a new memory allocation is created and
219 /// `old_layout` is ignored.
220 /// - `old_layout` must match the `Layout` the allocation has been created with.
221 ///
222 /// # Guarantees
223 ///
224 /// This function has the same guarantees as [`Allocator::alloc`]. When `ptr == Some(p)`, then
225 /// it additionally guarantees that:
226 /// - the contents of the memory pointed to by `p` are preserved up to the lesser of the new
227 /// and old size, i.e. `ret_ptr[0..min(layout.size(), old_layout.size())] ==
228 /// p[0..min(layout.size(), old_layout.size())]`.
229 /// - when the return value is `Err(AllocError)`, then `ptr` is still valid.
230 unsafe fn realloc(
231 ptr: Option<NonNull<u8>>,
232 layout: Layout,
233 old_layout: Layout,
234 flags: Flags,
235 nid: NumaNode,
236 ) -> Result<NonNull<[u8]>, AllocError>;
237
238 /// Free an existing memory allocation.
239 ///
240 /// # Safety
241 ///
242 /// - `ptr` must point to an existing and valid memory allocation created by this [`Allocator`];
243 /// if `old_layout` is zero-sized `p` does not need to be a pointer returned by this
244 /// [`Allocator`].
245 /// - `layout` must match the `Layout` the allocation has been created with.
246 /// - The memory allocation at `ptr` must never again be read from or written to.
247 unsafe fn free(ptr: NonNull<u8>, layout: Layout) {
248 // SAFETY: The caller guarantees that `ptr` points at a valid allocation created by this
249 // allocator. We are passing a `Layout` with the smallest possible alignment, so it is
250 // smaller than or equal to the alignment previously used with this allocation.
251 let _ = unsafe {
252 Self::realloc(
253 Some(ptr),
254 Layout::new::<()>(),
255 layout,
256 Flags(0),
257 NumaNode::NO_NODE,
258 )
259 };
260 }
261}
262
263/// Returns a properly aligned dangling pointer from the given `layout`.
264pub(crate) fn dangling_from_layout(layout: Layout) -> NonNull<u8> {
265 let ptr = layout.align() as *mut u8;
266
267 // SAFETY: `layout.align()` (and hence `ptr`) is guaranteed to be non-zero.
268 unsafe { NonNull::new_unchecked(ptr) }
269}