diff --git a/drivers/iommu/msm_dma_iommu_mapping.c b/drivers/iommu/msm_dma_iommu_mapping.c index 6b972350f0b8..a23aa9a48c40 100644 --- a/drivers/iommu/msm_dma_iommu_mapping.c +++ b/drivers/iommu/msm_dma_iommu_mapping.c @@ -1,6 +1,6 @@ // SPDX-License-Identifier: GPL-2.0-only /* - * Copyright (c) 2015-2019, The Linux Foundation. All rights reserved. + * Copyright (c) 2015-2020, The Linux Foundation. All rights reserved. */ #include @@ -251,7 +251,8 @@ static inline int __msm_dma_map_sg(struct device *dev, struct scatterlist *sg, dma_sync_sg_for_device(dev, iommu_map->sgl, iommu_map->nents, iommu_map->dir); - if (dev_is_dma_coherent(dev)) + if (dev_is_dma_coherent(dev) || + (attrs & DMA_ATTR_FORCE_COHERENT)) /* * Ensure all outstanding changes for coherent * buffers are applied to the cache before any diff --git a/drivers/of/address.c b/drivers/of/address.c index 978427a9d5e6..6c3f22d1aa54 100644 --- a/drivers/of/address.c +++ b/drivers/of/address.c @@ -1015,3 +1015,28 @@ bool of_dma_is_coherent(struct device_node *np) return false; } EXPORT_SYMBOL_GPL(of_dma_is_coherent); + +#if defined(CONFIG_DMA_COHERENT_HINT_CACHED) +/** + * of_dma_is_coherent_hint_cached - Check if device is coherent hint cached + * @np: device node + * + * It returns true if "dma-coherent-hint-cached" property was found + * for this device in DT. + */ +bool of_dma_is_coherent_hint_cached(struct device_node *np) +{ + struct device_node *node = of_node_get(np); + + while (node) { + if (of_property_read_bool(node, "dma-coherent-hint-cached")) { + of_node_put(node); + return true; + } + node = of_get_next_parent(node); + } + of_node_put(node); + return false; +} +EXPORT_SYMBOL(of_dma_is_coherent_hint_cached); +#endif /* CONFIG_DMA_COHERENT_HINT_CACHED */ diff --git a/drivers/of/device.c b/drivers/of/device.c index da8158392010..5bc25fa17222 100644 --- a/drivers/of/device.c +++ b/drivers/of/device.c @@ -90,7 +90,7 @@ int of_dma_configure(struct device *dev, struct device_node *np, bool force_dma) { u64 dma_addr, paddr, size = 0; int ret; - bool coherent; + bool coherent, coherent_hint_cached; unsigned long offset; const struct iommu_ops *iommu; u64 mask; @@ -159,6 +159,13 @@ int of_dma_configure(struct device *dev, struct device_node *np, bool force_dma) dev_dbg(dev, "device is%sdma coherent\n", coherent ? " " : " not "); + coherent_hint_cached = of_dma_is_coherent_hint_cached(np); + dev_dbg(dev, "device is%sdma coherent_hint_cached\n", + coherent_hint_cached ? " " : " not "); + dma_set_coherent_hint_cached(dev, coherent_hint_cached); + WARN(coherent && coherent_hint_cached, + "Should not set both dma-coherent and dma-coherent-hint-cached on the same device"); + iommu = of_iommu_configure(dev, np); if (IS_ERR(iommu) && PTR_ERR(iommu) == -EPROBE_DEFER) return -EPROBE_DEFER; diff --git a/drivers/staging/android/ion/heaps/msm_ion_dma_buf.c b/drivers/staging/android/ion/heaps/msm_ion_dma_buf.c index 29f64c45562a..eb7402f1fa85 100644 --- a/drivers/staging/android/ion/heaps/msm_ion_dma_buf.c +++ b/drivers/staging/android/ion/heaps/msm_ion_dma_buf.c @@ -1,11 +1,12 @@ // SPDX-License-Identifier: GPL-2.0-only /* * Copyright (C) 2011 Google, Inc. - * Copyright (c) 2011-2019, The Linux Foundation. All rights reserved. + * Copyright (c) 2011-2020, The Linux Foundation. All rights reserved. */ #include #include +#include #include #include #include @@ -163,6 +164,20 @@ static struct sg_table !hlos_accessible_buffer(buffer)) map_attrs |= DMA_ATTR_SKIP_CPU_SYNC; + if ((buffer->flags & ION_FLAG_CACHED) && + hlos_accessible_buffer(buffer) && + dev_is_dma_coherent_hint_cached(attachment->dev)) + map_attrs |= DMA_ATTR_FORCE_COHERENT; + + if (((dev_is_dma_coherent(attachment->dev) && + !(map_attrs & DMA_ATTR_FORCE_NON_COHERENT)) || + (map_attrs & DMA_ATTR_FORCE_COHERENT)) && + !(buffer->flags & ION_FLAG_CACHED)) { + pr_warn_ratelimited("dev:%s Cannot DMA map uncached buffer as IO-coherent attrs:0x%lx\n", + dev_name(attachment->dev), map_attrs); + return ERR_PTR(-EINVAL); + } + mutex_lock(&buffer->lock); if (map_attrs & DMA_ATTR_SKIP_CPU_SYNC) trace_ion_dma_map_cmo_skip(attachment->dev, @@ -233,6 +248,11 @@ static void msm_ion_unmap_dma_buf(struct dma_buf_attachment *attachment, !hlos_accessible_buffer(buffer)) map_attrs |= DMA_ATTR_SKIP_CPU_SYNC; + if ((buffer->flags & ION_FLAG_CACHED) && + hlos_accessible_buffer(buffer) && + dev_is_dma_coherent_hint_cached(attachment->dev)) + map_attrs |= DMA_ATTR_FORCE_COHERENT; + mutex_lock(&buffer->lock); if (map_attrs & DMA_ATTR_SKIP_CPU_SYNC) trace_ion_dma_unmap_cmo_skip(attachment->dev, diff --git a/include/linux/device.h b/include/linux/device.h index 3f4aadf3d2cc..5ae8491915fe 100644 --- a/include/linux/device.h +++ b/include/linux/device.h @@ -1245,7 +1245,9 @@ struct dev_links_info { * sync_state() callback. * @dma_coherent: this particular device is dma coherent, even if the * architecture supports non-coherent devices. - * + * @dma_coherent_hint_cached: Tell the framework to try and treat the device + * as DMA coherent when working with CPU cached + * buffers. * At the lowest level, every device in a Linux system is represented by an * instance of struct device. The device structure contains the information * that the device model core needs to model the system. Most subsystems, @@ -1345,6 +1347,10 @@ struct device { defined(CONFIG_ARCH_HAS_SYNC_DMA_FOR_CPU_ALL) bool dma_coherent:1; #endif +#if defined(CONFIG_DMA_COHERENT_HINT_CACHED) + bool dma_coherent_hint_cached:1; +#endif + }; static inline struct device *kobj_to_dev(struct kobject *kobj) diff --git a/include/linux/dma-mapping.h b/include/linux/dma-mapping.h index 07014778b5dd..e2d33885abe2 100644 --- a/include/linux/dma-mapping.h +++ b/include/linux/dma-mapping.h @@ -305,6 +305,19 @@ static inline void dma_direct_sync_sg_for_cpu(struct device *dev, size_t dma_direct_max_mapping_size(struct device *dev); +#ifdef CONFIG_DMA_COHERENT_HINT_CACHED +static inline void dma_set_coherent_hint_cached(struct device *dev, + bool hint_cached) +{ + dev->dma_coherent_hint_cached = hint_cached; +} +#else +static inline void dma_set_coherent_hint_cached(struct device *dev, + bool hint_cached) +{ +} +#endif + #ifdef CONFIG_HAS_DMA #include diff --git a/include/linux/dma-noncoherent.h b/include/linux/dma-noncoherent.h index 7c36b7966708..9b7140cc921b 100644 --- a/include/linux/dma-noncoherent.h +++ b/include/linux/dma-noncoherent.h @@ -21,6 +21,18 @@ static inline bool dev_is_dma_coherent(struct device *dev) } #endif /* CONFIG_ARCH_HAS_DMA_COHERENCE_H */ +#ifdef CONFIG_DMA_COHERENT_HINT_CACHED +static inline bool dev_is_dma_coherent_hint_cached(struct device *dev) +{ + return dev->dma_coherent_hint_cached; +} +#else +static inline bool dev_is_dma_coherent_hint_cached(struct device *dev) +{ + return false; +} +#endif + /* * Check if an allocation needs to be marked uncached to be coherent. */ diff --git a/include/linux/of_address.h b/include/linux/of_address.h index 30e40fb6936b..56cce0f91768 100644 --- a/include/linux/of_address.h +++ b/include/linux/of_address.h @@ -58,6 +58,16 @@ extern struct of_pci_range *of_pci_range_parser_one( extern int of_dma_get_range(struct device_node *np, u64 *dma_addr, u64 *paddr, u64 *size); extern bool of_dma_is_coherent(struct device_node *np); + +#if defined(CONFIG_DMA_COHERENT_HINT_CACHED) +extern bool of_dma_is_coherent_hint_cached(struct device_node *np); +#else /* CONFIG_DMA_COHERENT_HINT_CACHED */ +static inline bool of_dma_is_coherent_hint_cached(struct device_node *np) +{ + return false; +} +#endif /* CONFIG_DMA_COHERENT_HINT_CACHED */ + #else /* CONFIG_OF_ADDRESS */ static inline void __iomem *of_io_request_and_map(struct device_node *device, int index, const char *name) @@ -114,6 +124,12 @@ static inline bool of_dma_is_coherent(struct device_node *np) { return false; } + +static inline bool of_dma_is_coherent_hint_cached(struct device_node *np) +{ + return false; +} + #endif /* CONFIG_OF_ADDRESS */ #ifdef CONFIG_OF diff --git a/kernel/dma/Kconfig b/kernel/dma/Kconfig index 73c5c2b8e824..1fe0d1583478 100644 --- a/kernel/dma/Kconfig +++ b/kernel/dma/Kconfig @@ -77,6 +77,16 @@ config DMA_DIRECT_REMAP bool select DMA_REMAP +config DMA_COHERENT_HINT_CACHED + bool "DMA coherent hint cached" + help + Enabling this feature allows client to configure their device so + that the memory framework will only try to DMA map buffers to the device + as IO-coherent if the framework is confident that the buffers are mapped + to the CPU as cached. + + If unsure, say "n". + config DMA_CMA bool "DMA Contiguous Memory Allocator" depends on HAVE_DMA_CONTIGUOUS && CMA diff --git a/kernel/dma/mapping.c b/kernel/dma/mapping.c index 7f982ac3125a..ede39a2b9057 100644 --- a/kernel/dma/mapping.c +++ b/kernel/dma/mapping.c @@ -46,6 +46,18 @@ static int dmam_match(struct device *dev, void *res, void *match_data) return 0; } +static bool is_dma_coherent(struct device *dev, unsigned long attrs) +{ + if (attrs & DMA_ATTR_FORCE_COHERENT) + return true; + else if (attrs & DMA_ATTR_FORCE_NON_COHERENT) + return false; + else if (dev_is_dma_coherent(dev)) + return true; + else + return false; +} + /** * dmam_free_coherent - Managed dma_free_coherent() * @dev: Device to free coherent memory for @@ -116,7 +128,7 @@ int dma_common_get_sgtable(struct device *dev, struct sg_table *sgt, struct page *page; int ret; - if (!dev_is_dma_coherent(dev)) { + if (!is_dma_coherent(dev, attrs)) { unsigned long pfn; if (!IS_ENABLED(CONFIG_ARCH_HAS_DMA_COHERENT_TO_PFN)) @@ -154,6 +166,9 @@ int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, { const struct dma_map_ops *ops = get_dma_ops(dev); + if (dev_is_dma_coherent_hint_cached(dev)) + attrs |= DMA_ATTR_FORCE_COHERENT; + if (dma_is_direct(ops)) return dma_common_get_sgtable(dev, sgt, cpu_addr, dma_addr, size, attrs); @@ -164,18 +179,6 @@ int dma_get_sgtable_attrs(struct device *dev, struct sg_table *sgt, EXPORT_SYMBOL(dma_get_sgtable_attrs); #ifdef CONFIG_MMU -static bool is_dma_coherent(struct device *dev, unsigned long attrs) -{ - if (attrs & DMA_ATTR_FORCE_COHERENT) - return true; - else if (attrs & DMA_ATTR_FORCE_NON_COHERENT) - return false; - else if (dev_is_dma_coherent(dev)) - return true; - else - return false; -} - /* * Return the page attributes used for mapping dma_alloc_* memory, either in * kernel space if remapping is needed, or to userspace through dma_mmap_*. @@ -216,7 +219,7 @@ int dma_common_mmap(struct device *dev, struct vm_area_struct *vma, if (off >= count || user_count > count - off) return -ENXIO; - if (!dev_is_dma_coherent(dev)) { + if (!is_dma_coherent(dev, attrs)) { if (!IS_ENABLED(CONFIG_ARCH_HAS_DMA_COHERENT_TO_PFN)) return -ENXIO; @@ -275,6 +278,9 @@ int dma_mmap_attrs(struct device *dev, struct vm_area_struct *vma, { const struct dma_map_ops *ops = get_dma_ops(dev); + if (dev_is_dma_coherent_hint_cached(dev)) + attrs |= DMA_ATTR_FORCE_COHERENT; + if (dma_is_direct(ops)) return dma_common_mmap(dev, vma, cpu_addr, dma_addr, size, attrs); @@ -315,6 +321,9 @@ void *dma_alloc_attrs(struct device *dev, size_t size, dma_addr_t *dma_handle, WARN_ON(!of_reserved_mem_device_is_init(dev)); + if (dev_is_dma_coherent_hint_cached(dev)) + attrs |= DMA_ATTR_FORCE_COHERENT; + if (dma_alloc_from_dev_coherent(dev, size, dma_handle, &cpu_addr)) return cpu_addr; @@ -338,6 +347,9 @@ void dma_free_attrs(struct device *dev, size_t size, void *cpu_addr, { const struct dma_map_ops *ops = get_dma_ops(dev); + if (dev_is_dma_coherent_hint_cached(dev)) + attrs |= DMA_ATTR_FORCE_COHERENT; + if (dma_release_from_dev_coherent(dev, get_order(size), cpu_addr)) return; /*