ZONE_DMA32 is enabled by default, yet it is not
needed for all devices, nor is it desirable to have if not needed. For
instance, if a partner in GKI 1.0 did not use ZONE_DMA32, memory can
be lower for ZONE_NORMAL relative to older targets, such that memory
would run out more quickly in ZONE_NORMAL leading kswapd to be invoked
unnecessarily.
Correspondingly, provide a means of making ZONE_DMA32 empty via the
kernel command line when it is compiled in via CONFIG_ZONE_DMA32.
P.S. The following two patches are squashed into this one,
1. bf96382 ("ANDROID: dma-direct: Make DMA32 disablement work for CONFIG_NUMA")
2. 135406c ("ANDROID: dma-direct: Document disable_dma32")
Additionally, for kernels 6.6+, ensure that arm64_dma_phys_limit
is set correctly by letting it be set to the highest possible
physical address when disable_dma32 is present on the kernel
command line.
This allows SWIOTLB allocations to be satisfied from anywhere within
the physical memory space, as opposed to restricting these
allocations to the lower 32-bit addressable physical memory space.
It is important that SWIOTLB allocations use the correct addressing
limits, as SWIOTLB buffers can now be allocated at runtime--not just
boot. Using the correct addressing limits can help increase the amount
of memory available for SWIOTLB allocations, increasing the likelihood
of success.
This also aligns the behavior of disable_dma32 to what would happen if
CONFIG_ZONE_DMA32 were disabled.
Bug: 199917449
Bug: 268587627
Bug: 296282294
Bug: 372371001
Change-Id: I70ec76914b92e518d61a61072f0b3cb41cb28646
Signed-off-by: Chris Goldsworthy <quic_cgoldswo@quicinc.com>
Signed-off-by: Sudarshan Rajagopalan <quic_sudaraja@quicinc.com>
Signed-off-by: Chinwen Chang <chinwen.chang@mediatek.com>
[isaacmanjarres: Updated commit message to note that this
version of the patch treats arm64_phys_dma_limit differently
than the original version of the patch.]
Signed-off-by: Isaac J. Manjarres <isaacmanjarres@google.com>
Signed-off-by: Chinwen Chang <chinwen.chang@mediatek.com>
178 lines
4.4 KiB
C
178 lines
4.4 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* Internals of the DMA direct mapping implementation. Only for use by the
|
|
* DMA mapping code and IOMMU drivers.
|
|
*/
|
|
#ifndef _LINUX_DMA_DIRECT_H
|
|
#define _LINUX_DMA_DIRECT_H 1
|
|
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/dma-map-ops.h>
|
|
#include <linux/memblock.h> /* for min_low_pfn */
|
|
#include <linux/mem_encrypt.h>
|
|
#include <linux/swiotlb.h>
|
|
|
|
extern u64 zone_dma_limit;
|
|
|
|
/*
|
|
* Record the mapping of CPU physical to DMA addresses for a given region.
|
|
*/
|
|
struct bus_dma_region {
|
|
phys_addr_t cpu_start;
|
|
dma_addr_t dma_start;
|
|
u64 size;
|
|
};
|
|
|
|
static inline bool zone_dma32_is_empty(int node)
|
|
{
|
|
#ifdef CONFIG_ZONE_DMA32
|
|
pg_data_t *pgdat = NODE_DATA(node);
|
|
|
|
return zone_is_empty(&pgdat->node_zones[ZONE_DMA32]);
|
|
#else
|
|
return true;
|
|
#endif
|
|
}
|
|
|
|
static inline bool zone_dma32_are_empty(void)
|
|
{
|
|
#ifdef CONFIG_NUMA
|
|
int node;
|
|
|
|
for_each_node(node)
|
|
if (!zone_dma32_is_empty(node))
|
|
return false;
|
|
#else
|
|
if (!zone_dma32_is_empty(numa_node_id()))
|
|
return false;
|
|
#endif
|
|
|
|
return true;
|
|
}
|
|
|
|
static inline dma_addr_t translate_phys_to_dma(struct device *dev,
|
|
phys_addr_t paddr)
|
|
{
|
|
const struct bus_dma_region *m;
|
|
|
|
for (m = dev->dma_range_map; m->size; m++) {
|
|
u64 offset = paddr - m->cpu_start;
|
|
|
|
if (paddr >= m->cpu_start && offset < m->size)
|
|
return m->dma_start + offset;
|
|
}
|
|
|
|
/* make sure dma_capable fails when no translation is available */
|
|
return DMA_MAPPING_ERROR;
|
|
}
|
|
|
|
static inline phys_addr_t translate_dma_to_phys(struct device *dev,
|
|
dma_addr_t dma_addr)
|
|
{
|
|
const struct bus_dma_region *m;
|
|
|
|
for (m = dev->dma_range_map; m->size; m++) {
|
|
u64 offset = dma_addr - m->dma_start;
|
|
|
|
if (dma_addr >= m->dma_start && offset < m->size)
|
|
return m->cpu_start + offset;
|
|
}
|
|
|
|
return (phys_addr_t)-1;
|
|
}
|
|
|
|
static inline dma_addr_t dma_range_map_min(const struct bus_dma_region *map)
|
|
{
|
|
dma_addr_t ret = (dma_addr_t)U64_MAX;
|
|
|
|
for (; map->size; map++)
|
|
ret = min(ret, map->dma_start);
|
|
return ret;
|
|
}
|
|
|
|
static inline dma_addr_t dma_range_map_max(const struct bus_dma_region *map)
|
|
{
|
|
dma_addr_t ret = 0;
|
|
|
|
for (; map->size; map++)
|
|
ret = max(ret, map->dma_start + map->size - 1);
|
|
return ret;
|
|
}
|
|
|
|
#ifdef CONFIG_ARCH_HAS_PHYS_TO_DMA
|
|
#include <asm/dma-direct.h>
|
|
#ifndef phys_to_dma_unencrypted
|
|
#define phys_to_dma_unencrypted phys_to_dma
|
|
#endif
|
|
#else
|
|
static inline dma_addr_t phys_to_dma_unencrypted(struct device *dev,
|
|
phys_addr_t paddr)
|
|
{
|
|
if (dev->dma_range_map)
|
|
return translate_phys_to_dma(dev, paddr);
|
|
return paddr;
|
|
}
|
|
|
|
/*
|
|
* If memory encryption is supported, phys_to_dma will set the memory encryption
|
|
* bit in the DMA address, and dma_to_phys will clear it.
|
|
* phys_to_dma_unencrypted is for use on special unencrypted memory like swiotlb
|
|
* buffers.
|
|
*/
|
|
static inline dma_addr_t phys_to_dma(struct device *dev, phys_addr_t paddr)
|
|
{
|
|
return __sme_set(phys_to_dma_unencrypted(dev, paddr));
|
|
}
|
|
|
|
static inline phys_addr_t dma_to_phys(struct device *dev, dma_addr_t dma_addr)
|
|
{
|
|
phys_addr_t paddr;
|
|
|
|
if (dev->dma_range_map)
|
|
paddr = translate_dma_to_phys(dev, dma_addr);
|
|
else
|
|
paddr = dma_addr;
|
|
|
|
return __sme_clr(paddr);
|
|
}
|
|
#endif /* !CONFIG_ARCH_HAS_PHYS_TO_DMA */
|
|
|
|
#ifdef CONFIG_ARCH_HAS_FORCE_DMA_UNENCRYPTED
|
|
bool force_dma_unencrypted(struct device *dev);
|
|
#else
|
|
static inline bool force_dma_unencrypted(struct device *dev)
|
|
{
|
|
return false;
|
|
}
|
|
#endif /* CONFIG_ARCH_HAS_FORCE_DMA_UNENCRYPTED */
|
|
|
|
static inline bool dma_capable(struct device *dev, dma_addr_t addr, size_t size,
|
|
bool is_ram)
|
|
{
|
|
dma_addr_t end = addr + size - 1;
|
|
|
|
if (addr == DMA_MAPPING_ERROR)
|
|
return false;
|
|
if (is_ram && !IS_ENABLED(CONFIG_ARCH_DMA_ADDR_T_64BIT) &&
|
|
min(addr, end) < phys_to_dma(dev, PFN_PHYS(min_low_pfn)))
|
|
return false;
|
|
|
|
return end <= min_not_zero(*dev->dma_mask, dev->bus_dma_limit);
|
|
}
|
|
|
|
u64 dma_direct_get_required_mask(struct device *dev);
|
|
void *dma_direct_alloc(struct device *dev, size_t size, dma_addr_t *dma_handle,
|
|
gfp_t gfp, unsigned long attrs);
|
|
void dma_direct_free(struct device *dev, size_t size, void *cpu_addr,
|
|
dma_addr_t dma_addr, unsigned long attrs);
|
|
struct page *dma_direct_alloc_pages(struct device *dev, size_t size,
|
|
dma_addr_t *dma_handle, enum dma_data_direction dir, gfp_t gfp);
|
|
void dma_direct_free_pages(struct device *dev, size_t size,
|
|
struct page *page, dma_addr_t dma_addr,
|
|
enum dma_data_direction dir);
|
|
int dma_direct_supported(struct device *dev, u64 mask);
|
|
dma_addr_t dma_direct_map_resource(struct device *dev, phys_addr_t paddr,
|
|
size_t size, enum dma_data_direction dir, unsigned long attrs);
|
|
|
|
#endif /* _LINUX_DMA_DIRECT_H */
|