Linux kernel mirror (for testing) git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel os linux
1
fork

Configure Feed

Select the types of activity you want to include in your feed.

rust: iommu: add io_pgtable abstraction

This will be used by the Tyr driver to create and modify the page table
of each address space on the GPU. Each time a mapping gets created or
removed by userspace, Tyr will call into GPUVM, which will figure out
which calls to map_pages and unmap_pages are required to map the data in
question in the page table so that the GPU may access those pages when
using that address space.

The Rust type wraps the struct using a raw pointer rather than the usual
Opaque+ARef approach because Opaque+ARef requires the target type to be
refcounted.

Signed-off-by: Asahi Lina <lina+kernel@asahilina.net>
Acked-by: Boris Brezillon <boris.brezillon@collabora.com>
Reviewed-by: Daniel Almeida <daniel.almeida@collabora.com>
Tested-by: Deborah Brouwer <deborah.brouwer@collabora.com>
Co-developed-by: Alice Ryhl <aliceryhl@google.com>
Signed-off-by: Alice Ryhl <aliceryhl@google.com>
Reviewed-by: Gary Guo <gary@garyguo.net>
Reviewed-by: Danilo Krummrich <dakr@kernel.org>
[joro: Fixed up Rust import style]
Signed-off-by: Joerg Roedel <joerg.roedel@amd.com>

authored by

Asahi Lina and committed by
Joerg Roedel
2e2f6b0e e2692c4e

+288 -1
+1
MAINTAINERS
··· 13241 13241 F: include/linux/iommu.h 13242 13242 F: include/linux/iova.h 13243 13243 F: include/linux/of_iommu.h 13244 + F: rust/kernel/iommu/ 13244 13245 13245 13246 IOMMUFD 13246 13247 M: Jason Gunthorpe <jgg@nvidia.com>
+2 -1
rust/bindings/bindings_helper.h
··· 56 56 #include <linux/fdtable.h> 57 57 #include <linux/file.h> 58 58 #include <linux/firmware.h> 59 - #include <linux/interrupt.h> 60 59 #include <linux/fs.h> 61 60 #include <linux/i2c.h> 61 + #include <linux/interrupt.h> 62 + #include <linux/io-pgtable.h> 62 63 #include <linux/ioport.h> 63 64 #include <linux/jiffies.h> 64 65 #include <linux/jump_label.h>
+5
rust/kernel/iommu/mod.rs
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + 3 + //! Rust support related to IOMMU. 4 + 5 + pub mod pgtable;
+279
rust/kernel/iommu/pgtable.rs
··· 1 + // SPDX-License-Identifier: GPL-2.0 2 + 3 + //! IOMMU page table management. 4 + //! 5 + //! C header: [`include/io-pgtable.h`](srctree/include/io-pgtable.h) 6 + 7 + use core::{ 8 + marker::PhantomData, 9 + ptr::NonNull, // 10 + }; 11 + 12 + use crate::{ 13 + alloc, 14 + bindings, 15 + device::{ 16 + Bound, 17 + Device // 18 + }, 19 + devres::Devres, 20 + error::to_result, 21 + io::PhysAddr, 22 + prelude::*, // 23 + }; 24 + 25 + use bindings::io_pgtable_fmt; 26 + 27 + /// Protection flags used with IOMMU mappings. 28 + pub mod prot { 29 + /// Read access. 30 + pub const READ: u32 = bindings::IOMMU_READ; 31 + /// Write access. 32 + pub const WRITE: u32 = bindings::IOMMU_WRITE; 33 + /// Request cache coherency. 34 + pub const CACHE: u32 = bindings::IOMMU_CACHE; 35 + /// Request no-execute permission. 36 + pub const NOEXEC: u32 = bindings::IOMMU_NOEXEC; 37 + /// MMIO peripheral mapping. 38 + pub const MMIO: u32 = bindings::IOMMU_MMIO; 39 + /// Privileged mapping. 40 + pub const PRIVILEGED: u32 = bindings::IOMMU_PRIV; 41 + } 42 + 43 + /// Represents a requested `io_pgtable` configuration. 44 + pub struct Config { 45 + /// Quirk bitmask (type-specific). 46 + pub quirks: usize, 47 + /// Valid page sizes, as a bitmask of powers of two. 48 + pub pgsize_bitmap: usize, 49 + /// Input address space size in bits. 50 + pub ias: u32, 51 + /// Output address space size in bits. 52 + pub oas: u32, 53 + /// IOMMU uses coherent accesses for page table walks. 54 + pub coherent_walk: bool, 55 + } 56 + 57 + /// An io page table using a specific format. 58 + /// 59 + /// # Invariants 60 + /// 61 + /// The pointer references a valid io page table. 62 + pub struct IoPageTable<F: IoPageTableFmt> { 63 + ptr: NonNull<bindings::io_pgtable_ops>, 64 + _marker: PhantomData<F>, 65 + } 66 + 67 + // SAFETY: `struct io_pgtable_ops` is not restricted to a single thread. 68 + unsafe impl<F: IoPageTableFmt> Send for IoPageTable<F> {} 69 + // SAFETY: `struct io_pgtable_ops` may be accessed concurrently. 70 + unsafe impl<F: IoPageTableFmt> Sync for IoPageTable<F> {} 71 + 72 + /// The format used by this page table. 73 + pub trait IoPageTableFmt: 'static { 74 + /// The value representing this format. 75 + const FORMAT: io_pgtable_fmt; 76 + } 77 + 78 + impl<F: IoPageTableFmt> IoPageTable<F> { 79 + /// Create a new `IoPageTable` as a device resource. 80 + #[inline] 81 + pub fn new( 82 + dev: &Device<Bound>, 83 + config: Config, 84 + ) -> impl PinInit<Devres<IoPageTable<F>>, Error> + '_ { 85 + // SAFETY: Devres ensures that the value is dropped during device unbind. 86 + Devres::new(dev, unsafe { Self::new_raw(dev, config) }) 87 + } 88 + 89 + /// Create a new `IoPageTable`. 90 + /// 91 + /// # Safety 92 + /// 93 + /// If successful, then the returned `IoPageTable` must be dropped before the device is 94 + /// unbound. 95 + #[inline] 96 + pub unsafe fn new_raw(dev: &Device<Bound>, config: Config) -> Result<IoPageTable<F>> { 97 + let mut raw_cfg = bindings::io_pgtable_cfg { 98 + quirks: config.quirks, 99 + pgsize_bitmap: config.pgsize_bitmap, 100 + ias: config.ias, 101 + oas: config.oas, 102 + coherent_walk: config.coherent_walk, 103 + tlb: &raw const NOOP_FLUSH_OPS, 104 + iommu_dev: dev.as_raw(), 105 + // SAFETY: All zeroes is a valid value for `struct io_pgtable_cfg`. 106 + ..unsafe { core::mem::zeroed() } 107 + }; 108 + 109 + // SAFETY: 110 + // * The raw_cfg pointer is valid for the duration of this call. 111 + // * The provided `FLUSH_OPS` contains valid function pointers that accept a null pointer 112 + // as cookie. 113 + // * The caller ensures that the io pgtable does not outlive the device. 114 + let ops = unsafe { 115 + bindings::alloc_io_pgtable_ops(F::FORMAT, &mut raw_cfg, core::ptr::null_mut()) 116 + }; 117 + 118 + // INVARIANT: We successfully created a valid page table. 119 + Ok(IoPageTable { 120 + ptr: NonNull::new(ops).ok_or(ENOMEM)?, 121 + _marker: PhantomData, 122 + }) 123 + } 124 + 125 + /// Obtain a raw pointer to the underlying `struct io_pgtable_ops`. 126 + #[inline] 127 + pub fn raw_ops(&self) -> *mut bindings::io_pgtable_ops { 128 + self.ptr.as_ptr() 129 + } 130 + 131 + /// Obtain a raw pointer to the underlying `struct io_pgtable`. 132 + #[inline] 133 + pub fn raw_pgtable(&self) -> *mut bindings::io_pgtable { 134 + // SAFETY: The io_pgtable_ops of an io-pgtable is always the ops field of a io_pgtable. 135 + unsafe { kernel::container_of!(self.raw_ops(), bindings::io_pgtable, ops) } 136 + } 137 + 138 + /// Obtain a raw pointer to the underlying `struct io_pgtable_cfg`. 139 + #[inline] 140 + pub fn raw_cfg(&self) -> *mut bindings::io_pgtable_cfg { 141 + // SAFETY: The `raw_pgtable()` method returns a valid pointer. 142 + unsafe { &raw mut (*self.raw_pgtable()).cfg } 143 + } 144 + 145 + /// Map a physically contiguous range of pages of the same size. 146 + /// 147 + /// Even if successful, this operation may not map the entire range. In that case, only a 148 + /// prefix of the range is mapped, and the returned integer indicates its length in bytes. In 149 + /// this case, the caller will usually call `map_pages` again for the remaining range. 150 + /// 151 + /// The returned [`Result`] indicates whether an error was encountered while mapping pages. 152 + /// Note that this may return a non-zero length even if an error was encountered. The caller 153 + /// will usually [unmap the relevant pages](Self::unmap_pages) on error. 154 + /// 155 + /// The caller must flush the TLB before using the pgtable to access the newly created mapping. 156 + /// 157 + /// # Safety 158 + /// 159 + /// * No other io-pgtable operation may access the range `iova .. iova+pgsize*pgcount` while 160 + /// this `map_pages` operation executes. 161 + /// * This page table must not contain any mapping that overlaps with the mapping created by 162 + /// this call. 163 + /// * If this page table is live, then the caller must ensure that it's okay to access the 164 + /// physical address being mapped for the duration in which it is mapped. 165 + #[inline] 166 + pub unsafe fn map_pages( 167 + &self, 168 + iova: usize, 169 + paddr: PhysAddr, 170 + pgsize: usize, 171 + pgcount: usize, 172 + prot: u32, 173 + flags: alloc::Flags, 174 + ) -> (usize, Result) { 175 + let mut mapped: usize = 0; 176 + 177 + // SAFETY: The `map_pages` function in `io_pgtable_ops` is never null. 178 + let map_pages = unsafe { (*self.raw_ops()).map_pages.unwrap_unchecked() }; 179 + 180 + // SAFETY: The safety requirements of this method are sufficient to call `map_pages`. 181 + let ret = to_result(unsafe { 182 + (map_pages)( 183 + self.raw_ops(), 184 + iova, 185 + paddr, 186 + pgsize, 187 + pgcount, 188 + prot as i32, 189 + flags.as_raw(), 190 + &mut mapped, 191 + ) 192 + }); 193 + 194 + (mapped, ret) 195 + } 196 + 197 + /// Unmap a range of virtually contiguous pages of the same size. 198 + /// 199 + /// This may not unmap the entire range, and returns the length of the unmapped prefix in 200 + /// bytes. 201 + /// 202 + /// # Safety 203 + /// 204 + /// * No other io-pgtable operation may access the range `iova .. iova+pgsize*pgcount` while 205 + /// this `unmap_pages` operation executes. 206 + /// * This page table must contain one or more consecutive mappings starting at `iova` whose 207 + /// total size is `pgcount * pgsize`. 208 + #[inline] 209 + #[must_use] 210 + pub unsafe fn unmap_pages(&self, iova: usize, pgsize: usize, pgcount: usize) -> usize { 211 + // SAFETY: The `unmap_pages` function in `io_pgtable_ops` is never null. 212 + let unmap_pages = unsafe { (*self.raw_ops()).unmap_pages.unwrap_unchecked() }; 213 + 214 + // SAFETY: The safety requirements of this method are sufficient to call `unmap_pages`. 215 + unsafe { (unmap_pages)(self.raw_ops(), iova, pgsize, pgcount, core::ptr::null_mut()) } 216 + } 217 + } 218 + 219 + // For the initial users of these rust bindings, the GPU FW is managing the IOTLB and performs all 220 + // required invalidations using a range. There is no need for it get ARM style invalidation 221 + // instructions from the page table code. 222 + // 223 + // Support for flushing the TLB with ARM style invalidation instructions may be added in the 224 + // future. 225 + static NOOP_FLUSH_OPS: bindings::iommu_flush_ops = bindings::iommu_flush_ops { 226 + tlb_flush_all: Some(rust_tlb_flush_all_noop), 227 + tlb_flush_walk: Some(rust_tlb_flush_walk_noop), 228 + tlb_add_page: None, 229 + }; 230 + 231 + #[no_mangle] 232 + extern "C" fn rust_tlb_flush_all_noop(_cookie: *mut core::ffi::c_void) {} 233 + 234 + #[no_mangle] 235 + extern "C" fn rust_tlb_flush_walk_noop( 236 + _iova: usize, 237 + _size: usize, 238 + _granule: usize, 239 + _cookie: *mut core::ffi::c_void, 240 + ) { 241 + } 242 + 243 + impl<F: IoPageTableFmt> Drop for IoPageTable<F> { 244 + fn drop(&mut self) { 245 + // SAFETY: The caller of `Self::ttbr()` promised that the page table is not live when this 246 + // destructor runs. 247 + unsafe { bindings::free_io_pgtable_ops(self.raw_ops()) }; 248 + } 249 + } 250 + 251 + /// The `ARM_64_LPAE_S1` page table format. 252 + pub enum ARM64LPAES1 {} 253 + 254 + impl IoPageTableFmt for ARM64LPAES1 { 255 + const FORMAT: io_pgtable_fmt = bindings::io_pgtable_fmt_ARM_64_LPAE_S1 as io_pgtable_fmt; 256 + } 257 + 258 + impl IoPageTable<ARM64LPAES1> { 259 + /// Access the `ttbr` field of the configuration. 260 + /// 261 + /// This is the physical address of the page table, which may be passed to the device that 262 + /// needs to use it. 263 + /// 264 + /// # Safety 265 + /// 266 + /// The caller must ensure that the device stops using the page table before dropping it. 267 + #[inline] 268 + pub unsafe fn ttbr(&self) -> u64 { 269 + // SAFETY: `arm_lpae_s1_cfg` is the right cfg type for `ARM64LPAES1`. 270 + unsafe { (*self.raw_cfg()).__bindgen_anon_1.arm_lpae_s1_cfg.ttbr } 271 + } 272 + 273 + /// Access the `mair` field of the configuration. 274 + #[inline] 275 + pub fn mair(&self) -> u64 { 276 + // SAFETY: `arm_lpae_s1_cfg` is the right cfg type for `ARM64LPAES1`. 277 + unsafe { (*self.raw_cfg()).__bindgen_anon_1.arm_lpae_s1_cfg.mair } 278 + } 279 + }
+1
rust/kernel/lib.rs
··· 103 103 pub mod init; 104 104 pub mod io; 105 105 pub mod ioctl; 106 + pub mod iommu; 106 107 pub mod iov; 107 108 pub mod irq; 108 109 pub mod jump_label;