dma-mapping fixes for Linux 6.17
- another small fix relevant to arm64 systems with memory encryption (Shanker Donthineni) - fix relevant to arm32 systems with non-standard CMA configuration (Oreoluwa Babatunde) -----BEGIN PGP SIGNATURE----- iHUEABYIAB0WIQSrngzkoBtlA8uaaJ+Jp1EFxbsSRAUCaLAK7gAKCRCJp1EFxbsS REDxAQC+5hLiyzc/1rR5EQb1D6Xr1f/0VN3IFz3creHp3juFBAEApi1iFMdmahO7 0YKG4KkzHpcNkGrxaXKP0VNtQsDLwww= =fVrB -----END PGP SIGNATURE----- Merge tag 'dma-mapping-6.17-2025-08-28' of git://git.kernel.org/pub/scm/linux/kernel/git/mszyprowski/linux Pull dma-mapping fixes from Marek Szyprowski: - another small fix for arm64 systems with memory encryption (Shanker Donthineni) - fix for arm32 systems with non-standard CMA configuration (Oreoluwa Babatunde) * tag 'dma-mapping-6.17-2025-08-28' of git://git.kernel.org/pub/scm/linux/kernel/git/mszyprowski/linux: dma/pool: Ensure DMA_DIRECT_REMAP allocations are decrypted of: reserved_mem: Restructure call site for dma_contiguous_early_fixup()pull/1334/head
commit
4d28e28098
|
|
@ -25,6 +25,7 @@
|
||||||
#include <linux/memblock.h>
|
#include <linux/memblock.h>
|
||||||
#include <linux/kmemleak.h>
|
#include <linux/kmemleak.h>
|
||||||
#include <linux/cma.h>
|
#include <linux/cma.h>
|
||||||
|
#include <linux/dma-map-ops.h>
|
||||||
|
|
||||||
#include "of_private.h"
|
#include "of_private.h"
|
||||||
|
|
||||||
|
|
@ -175,13 +176,17 @@ static int __init __reserved_mem_reserve_reg(unsigned long node,
|
||||||
base = dt_mem_next_cell(dt_root_addr_cells, &prop);
|
base = dt_mem_next_cell(dt_root_addr_cells, &prop);
|
||||||
size = dt_mem_next_cell(dt_root_size_cells, &prop);
|
size = dt_mem_next_cell(dt_root_size_cells, &prop);
|
||||||
|
|
||||||
if (size &&
|
if (size && early_init_dt_reserve_memory(base, size, nomap) == 0) {
|
||||||
early_init_dt_reserve_memory(base, size, nomap) == 0)
|
/* Architecture specific contiguous memory fixup. */
|
||||||
|
if (of_flat_dt_is_compatible(node, "shared-dma-pool") &&
|
||||||
|
of_get_flat_dt_prop(node, "reusable", NULL))
|
||||||
|
dma_contiguous_early_fixup(base, size);
|
||||||
pr_debug("Reserved memory: reserved region for node '%s': base %pa, size %lu MiB\n",
|
pr_debug("Reserved memory: reserved region for node '%s': base %pa, size %lu MiB\n",
|
||||||
uname, &base, (unsigned long)(size / SZ_1M));
|
uname, &base, (unsigned long)(size / SZ_1M));
|
||||||
else
|
} else {
|
||||||
pr_err("Reserved memory: failed to reserve memory for node '%s': base %pa, size %lu MiB\n",
|
pr_err("Reserved memory: failed to reserve memory for node '%s': base %pa, size %lu MiB\n",
|
||||||
uname, &base, (unsigned long)(size / SZ_1M));
|
uname, &base, (unsigned long)(size / SZ_1M));
|
||||||
|
}
|
||||||
|
|
||||||
len -= t_len;
|
len -= t_len;
|
||||||
}
|
}
|
||||||
|
|
@ -472,7 +477,10 @@ static int __init __reserved_mem_alloc_size(unsigned long node, const char *unam
|
||||||
uname, (unsigned long)(size / SZ_1M));
|
uname, (unsigned long)(size / SZ_1M));
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
|
/* Architecture specific contiguous memory fixup. */
|
||||||
|
if (of_flat_dt_is_compatible(node, "shared-dma-pool") &&
|
||||||
|
of_get_flat_dt_prop(node, "reusable", NULL))
|
||||||
|
dma_contiguous_early_fixup(base, size);
|
||||||
/* Save region in the reserved_mem array */
|
/* Save region in the reserved_mem array */
|
||||||
fdt_reserved_mem_save_node(node, uname, base, size);
|
fdt_reserved_mem_save_node(node, uname, base, size);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
||||||
|
|
@ -153,6 +153,9 @@ static inline void dma_free_contiguous(struct device *dev, struct page *page,
|
||||||
{
|
{
|
||||||
__free_pages(page, get_order(size));
|
__free_pages(page, get_order(size));
|
||||||
}
|
}
|
||||||
|
static inline void dma_contiguous_early_fixup(phys_addr_t base, unsigned long size)
|
||||||
|
{
|
||||||
|
}
|
||||||
#endif /* CONFIG_DMA_CMA*/
|
#endif /* CONFIG_DMA_CMA*/
|
||||||
|
|
||||||
#ifdef CONFIG_DMA_DECLARE_COHERENT
|
#ifdef CONFIG_DMA_DECLARE_COHERENT
|
||||||
|
|
|
||||||
|
|
@ -483,8 +483,6 @@ static int __init rmem_cma_setup(struct reserved_mem *rmem)
|
||||||
pr_err("Reserved memory: unable to setup CMA region\n");
|
pr_err("Reserved memory: unable to setup CMA region\n");
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
/* Architecture specific contiguous memory fixup. */
|
|
||||||
dma_contiguous_early_fixup(rmem->base, rmem->size);
|
|
||||||
|
|
||||||
if (default_cma)
|
if (default_cma)
|
||||||
dma_contiguous_default_area = cma;
|
dma_contiguous_default_area = cma;
|
||||||
|
|
|
||||||
|
|
@ -102,8 +102,8 @@ static int atomic_pool_expand(struct gen_pool *pool, size_t pool_size,
|
||||||
|
|
||||||
#ifdef CONFIG_DMA_DIRECT_REMAP
|
#ifdef CONFIG_DMA_DIRECT_REMAP
|
||||||
addr = dma_common_contiguous_remap(page, pool_size,
|
addr = dma_common_contiguous_remap(page, pool_size,
|
||||||
pgprot_dmacoherent(PAGE_KERNEL),
|
pgprot_decrypted(pgprot_dmacoherent(PAGE_KERNEL)),
|
||||||
__builtin_return_address(0));
|
__builtin_return_address(0));
|
||||||
if (!addr)
|
if (!addr)
|
||||||
goto free_page;
|
goto free_page;
|
||||||
#else
|
#else
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue