|  | // SPDX-License-Identifier: GPL-2.0 | 
|  | /* | 
|  | * Copyright (C) 2021 Intel Corporation | 
|  | * Author: Johannes Berg <johannes@sipsolutions.net> | 
|  | */ | 
|  | #include <linux/types.h> | 
|  | #include <linux/slab.h> | 
|  | #include <linux/logic_iomem.h> | 
|  | #include <asm/io.h> | 
|  |  | 
|  | struct logic_iomem_region { | 
|  | const struct resource *res; | 
|  | const struct logic_iomem_region_ops *ops; | 
|  | struct list_head list; | 
|  | }; | 
|  |  | 
|  | struct logic_iomem_area { | 
|  | const struct logic_iomem_ops *ops; | 
|  | void *priv; | 
|  | }; | 
|  |  | 
|  | #define AREA_SHIFT	24 | 
|  | #define MAX_AREA_SIZE	(1 << AREA_SHIFT) | 
|  | #define MAX_AREAS	((1U << 31) / MAX_AREA_SIZE) | 
|  | #define AREA_BITS	((MAX_AREAS - 1) << AREA_SHIFT) | 
|  | #define AREA_MASK	(MAX_AREA_SIZE - 1) | 
|  | #ifdef CONFIG_64BIT | 
|  | #define IOREMAP_BIAS	0xDEAD000000000000UL | 
|  | #define IOREMAP_MASK	0xFFFFFFFF00000000UL | 
|  | #else | 
|  | #define IOREMAP_BIAS	0x80000000UL | 
|  | #define IOREMAP_MASK	0x80000000UL | 
|  | #endif | 
|  |  | 
|  | static DEFINE_MUTEX(regions_mtx); | 
|  | static LIST_HEAD(regions_list); | 
|  | static struct logic_iomem_area mapped_areas[MAX_AREAS]; | 
|  |  | 
|  | int logic_iomem_add_region(struct resource *resource, | 
|  | const struct logic_iomem_region_ops *ops) | 
|  | { | 
|  | struct logic_iomem_region *rreg; | 
|  | int err; | 
|  |  | 
|  | if (WARN_ON(!resource || !ops)) | 
|  | return -EINVAL; | 
|  |  | 
|  | if (WARN_ON((resource->flags & IORESOURCE_TYPE_BITS) != IORESOURCE_MEM)) | 
|  | return -EINVAL; | 
|  |  | 
|  | rreg = kzalloc(sizeof(*rreg), GFP_KERNEL); | 
|  | if (!rreg) | 
|  | return -ENOMEM; | 
|  |  | 
|  | err = request_resource(&iomem_resource, resource); | 
|  | if (err) { | 
|  | kfree(rreg); | 
|  | return -ENOMEM; | 
|  | } | 
|  |  | 
|  | mutex_lock(®ions_mtx); | 
|  | rreg->res = resource; | 
|  | rreg->ops = ops; | 
|  | list_add_tail(&rreg->list, ®ions_list); | 
|  | mutex_unlock(®ions_mtx); | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | EXPORT_SYMBOL(logic_iomem_add_region); | 
|  |  | 
|  | #ifndef CONFIG_INDIRECT_IOMEM_FALLBACK | 
|  | static void __iomem *real_ioremap(phys_addr_t offset, size_t size) | 
|  | { | 
|  | WARN(1, "invalid ioremap(0x%llx, 0x%zx)\n", | 
|  | (unsigned long long)offset, size); | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | static void real_iounmap(volatile void __iomem *addr) | 
|  | { | 
|  | WARN(1, "invalid iounmap for addr 0x%llx\n", | 
|  | (unsigned long long)(uintptr_t __force)addr); | 
|  | } | 
|  | #endif /* CONFIG_INDIRECT_IOMEM_FALLBACK */ | 
|  |  | 
|  | void __iomem *ioremap(phys_addr_t offset, size_t size) | 
|  | { | 
|  | void __iomem *ret = NULL; | 
|  | struct logic_iomem_region *rreg, *found = NULL; | 
|  | int i; | 
|  |  | 
|  | mutex_lock(®ions_mtx); | 
|  | list_for_each_entry(rreg, ®ions_list, list) { | 
|  | if (rreg->res->start > offset) | 
|  | continue; | 
|  | if (rreg->res->end < offset + size - 1) | 
|  | continue; | 
|  | found = rreg; | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (!found) | 
|  | goto out; | 
|  |  | 
|  | for (i = 0; i < MAX_AREAS; i++) { | 
|  | long offs; | 
|  |  | 
|  | if (mapped_areas[i].ops) | 
|  | continue; | 
|  |  | 
|  | offs = rreg->ops->map(offset - found->res->start, | 
|  | size, &mapped_areas[i].ops, | 
|  | &mapped_areas[i].priv); | 
|  | if (offs < 0) { | 
|  | mapped_areas[i].ops = NULL; | 
|  | break; | 
|  | } | 
|  |  | 
|  | if (WARN_ON(!mapped_areas[i].ops)) { | 
|  | mapped_areas[i].ops = NULL; | 
|  | break; | 
|  | } | 
|  |  | 
|  | ret = (void __iomem *)(IOREMAP_BIAS + (i << AREA_SHIFT) + offs); | 
|  | break; | 
|  | } | 
|  | out: | 
|  | mutex_unlock(®ions_mtx); | 
|  | if (ret) | 
|  | return ret; | 
|  | return real_ioremap(offset, size); | 
|  | } | 
|  | EXPORT_SYMBOL(ioremap); | 
|  |  | 
|  | static inline struct logic_iomem_area * | 
|  | get_area(const volatile void __iomem *addr) | 
|  | { | 
|  | unsigned long a = (unsigned long)addr; | 
|  | unsigned int idx; | 
|  |  | 
|  | if (WARN_ON((a & IOREMAP_MASK) != IOREMAP_BIAS)) | 
|  | return NULL; | 
|  |  | 
|  | idx = (a & AREA_BITS) >> AREA_SHIFT; | 
|  |  | 
|  | if (mapped_areas[idx].ops) | 
|  | return &mapped_areas[idx]; | 
|  |  | 
|  | return NULL; | 
|  | } | 
|  |  | 
|  | void iounmap(volatile void __iomem *addr) | 
|  | { | 
|  | struct logic_iomem_area *area = get_area(addr); | 
|  |  | 
|  | if (!area) { | 
|  | real_iounmap(addr); | 
|  | return; | 
|  | } | 
|  |  | 
|  | if (area->ops->unmap) | 
|  | area->ops->unmap(area->priv); | 
|  |  | 
|  | mutex_lock(®ions_mtx); | 
|  | area->ops = NULL; | 
|  | area->priv = NULL; | 
|  | mutex_unlock(®ions_mtx); | 
|  | } | 
|  | EXPORT_SYMBOL(iounmap); | 
|  |  | 
|  | #ifndef CONFIG_INDIRECT_IOMEM_FALLBACK | 
|  | #define MAKE_FALLBACK(op, sz) 						\ | 
|  | static u##sz real_raw_read ## op(const volatile void __iomem *addr)	\ | 
|  | {									\ | 
|  | WARN(1, "Invalid read" #op " at address %llx\n",		\ | 
|  | (unsigned long long)(uintptr_t __force)addr);		\ | 
|  | return (u ## sz)~0ULL;						\ | 
|  | }									\ | 
|  | \ | 
|  | static void real_raw_write ## op(u ## sz val,				\ | 
|  | volatile void __iomem *addr)		\ | 
|  | {									\ | 
|  | WARN(1, "Invalid writeq" #op " of 0x%llx at address %llx\n",	\ | 
|  | (unsigned long long)val,					\ | 
|  | (unsigned long long)(uintptr_t __force)addr);\ | 
|  | }									\ | 
|  |  | 
|  | MAKE_FALLBACK(b, 8); | 
|  | MAKE_FALLBACK(w, 16); | 
|  | MAKE_FALLBACK(l, 32); | 
|  | #ifdef CONFIG_64BIT | 
|  | MAKE_FALLBACK(q, 64); | 
|  | #endif | 
|  |  | 
|  | static void real_memset_io(volatile void __iomem *addr, int value, size_t size) | 
|  | { | 
|  | WARN(1, "Invalid memset_io at address 0x%llx\n", | 
|  | (unsigned long long)(uintptr_t __force)addr); | 
|  | } | 
|  |  | 
|  | static void real_memcpy_fromio(void *buffer, const volatile void __iomem *addr, | 
|  | size_t size) | 
|  | { | 
|  | WARN(1, "Invalid memcpy_fromio at address 0x%llx\n", | 
|  | (unsigned long long)(uintptr_t __force)addr); | 
|  |  | 
|  | memset(buffer, 0xff, size); | 
|  | } | 
|  |  | 
|  | static void real_memcpy_toio(volatile void __iomem *addr, const void *buffer, | 
|  | size_t size) | 
|  | { | 
|  | WARN(1, "Invalid memcpy_toio at address 0x%llx\n", | 
|  | (unsigned long long)(uintptr_t __force)addr); | 
|  | } | 
|  | #endif /* CONFIG_INDIRECT_IOMEM_FALLBACK */ | 
|  |  | 
|  | #define MAKE_OP(op, sz) 						\ | 
|  | u##sz __raw_read ## op(const volatile void __iomem *addr)		\ | 
|  | {									\ | 
|  | struct logic_iomem_area *area = get_area(addr);			\ | 
|  | \ | 
|  | if (!area)							\ | 
|  | return real_raw_read ## op(addr);			\ | 
|  | \ | 
|  | return (u ## sz) area->ops->read(area->priv,			\ | 
|  | (unsigned long)addr & AREA_MASK,\ | 
|  | sz / 8);			\ | 
|  | }									\ | 
|  | EXPORT_SYMBOL(__raw_read ## op);					\ | 
|  | \ | 
|  | void __raw_write ## op(u ## sz val, volatile void __iomem *addr)	\ | 
|  | {									\ | 
|  | struct logic_iomem_area *area = get_area(addr);			\ | 
|  | \ | 
|  | if (!area) {							\ | 
|  | real_raw_write ## op(val, addr);			\ | 
|  | return;							\ | 
|  | }								\ | 
|  | \ | 
|  | area->ops->write(area->priv,					\ | 
|  | (unsigned long)addr & AREA_MASK,		\ | 
|  | sz / 8, val);					\ | 
|  | }									\ | 
|  | EXPORT_SYMBOL(__raw_write ## op) | 
|  |  | 
|  | MAKE_OP(b, 8); | 
|  | MAKE_OP(w, 16); | 
|  | MAKE_OP(l, 32); | 
|  | #ifdef CONFIG_64BIT | 
|  | MAKE_OP(q, 64); | 
|  | #endif | 
|  |  | 
|  | void memset_io(volatile void __iomem *addr, int value, size_t size) | 
|  | { | 
|  | struct logic_iomem_area *area = get_area(addr); | 
|  | unsigned long offs, start; | 
|  |  | 
|  | if (!area) { | 
|  | real_memset_io(addr, value, size); | 
|  | return; | 
|  | } | 
|  |  | 
|  | start = (unsigned long)addr & AREA_MASK; | 
|  |  | 
|  | if (area->ops->set) { | 
|  | area->ops->set(area->priv, start, value, size); | 
|  | return; | 
|  | } | 
|  |  | 
|  | for (offs = 0; offs < size; offs++) | 
|  | area->ops->write(area->priv, start + offs, 1, value); | 
|  | } | 
|  | EXPORT_SYMBOL(memset_io); | 
|  |  | 
|  | void memcpy_fromio(void *buffer, const volatile void __iomem *addr, | 
|  | size_t size) | 
|  | { | 
|  | struct logic_iomem_area *area = get_area(addr); | 
|  | u8 *buf = buffer; | 
|  | unsigned long offs, start; | 
|  |  | 
|  | if (!area) { | 
|  | real_memcpy_fromio(buffer, addr, size); | 
|  | return; | 
|  | } | 
|  |  | 
|  | start = (unsigned long)addr & AREA_MASK; | 
|  |  | 
|  | if (area->ops->copy_from) { | 
|  | area->ops->copy_from(area->priv, buffer, start, size); | 
|  | return; | 
|  | } | 
|  |  | 
|  | for (offs = 0; offs < size; offs++) | 
|  | buf[offs] = area->ops->read(area->priv, start + offs, 1); | 
|  | } | 
|  | EXPORT_SYMBOL(memcpy_fromio); | 
|  |  | 
|  | void memcpy_toio(volatile void __iomem *addr, const void *buffer, size_t size) | 
|  | { | 
|  | struct logic_iomem_area *area = get_area(addr); | 
|  | const u8 *buf = buffer; | 
|  | unsigned long offs, start; | 
|  |  | 
|  | if (!area) { | 
|  | real_memcpy_toio(addr, buffer, size); | 
|  | return; | 
|  | } | 
|  |  | 
|  | start = (unsigned long)addr & AREA_MASK; | 
|  |  | 
|  | if (area->ops->copy_to) { | 
|  | area->ops->copy_to(area->priv, start, buffer, size); | 
|  | return; | 
|  | } | 
|  |  | 
|  | for (offs = 0; offs < size; offs++) | 
|  | area->ops->write(area->priv, start + offs, 1, buf[offs]); | 
|  | } | 
|  | EXPORT_SYMBOL(memcpy_toio); |