blob: 25efcc8724cfe433c93cf53a34804d0ee0e7ad4f [file] [log] [blame]
/* SPDX-License-Identifier: GPL-2.0-only */
#ifndef _REGION_H_
#define _REGION_H_
#include <sys/types.h>
#include <stddef.h>
#include <stdbool.h>
#include <commonlib/bsd/helpers.h>
#include <commonlib/mem_pool.h>
/*
* Region support.
*
* Regions are intended to abstract away the access mechanisms for blocks of
* data. This could be SPI, eMMC, or a memory region as the backing store.
* They are accessed through a region_device. Subregions can be made by
* chaining together multiple region_devices.
*/
struct region_device;
/*
* Returns NULL on error otherwise a buffer is returned with the contents of
* the requested data at offset of size.
*/
void *rdev_mmap(const struct region_device *rd, size_t offset, size_t size);
/* Unmap a previously mapped area. Returns 0 on success, < 0 on error. */
int rdev_munmap(const struct region_device *rd, void *mapping);
/*
* Returns < 0 on error otherwise returns size of data read at provided
* offset filling in the buffer passed.
*/
ssize_t rdev_readat(const struct region_device *rd, void *b, size_t offset,
size_t size);
/*
* Returns < 0 on error otherwise returns size of data wrote at provided
* offset from the buffer passed.
*/
ssize_t rdev_writeat(const struct region_device *rd, const void *b,
size_t offset, size_t size);
/*
* Returns < 0 on error otherwise returns size of data erased.
* If eraseat ops is not defined it returns size which indicates
* that operation was successful.
*/
ssize_t rdev_eraseat(const struct region_device *rd, size_t offset,
size_t size);
/****************************************
* Implementation of a region device *
****************************************/
/*
* Create a child region of the parent provided the sub-region is within
* the parent's region. Returns < 0 on error otherwise 0 on success. Note
* that the child device only calls through the parent's operations.
*/
int rdev_chain(struct region_device *child, const struct region_device *parent,
size_t offset, size_t size);
/* A region_device operations. */
struct region_device_ops {
void *(*mmap)(const struct region_device *, size_t, size_t);
int (*munmap)(const struct region_device *, void *);
ssize_t (*readat)(const struct region_device *, void *, size_t, size_t);
ssize_t (*writeat)(const struct region_device *, const void *, size_t,
size_t);
ssize_t (*eraseat)(const struct region_device *, size_t, size_t);
};
struct region {
size_t offset;
size_t size;
};
struct region_device {
const struct region_device *root;
const struct region_device_ops *ops;
struct region region;
};
#define REGION_DEV_INIT(ops_, offset_, size_) \
{ \
.root = NULL, \
.ops = (ops_), \
.region = { \
.offset = (offset_), \
.size = (size_), \
}, \
}
/* Helper to dynamically initialize region device. */
void region_device_init(struct region_device *rdev,
const struct region_device_ops *ops, size_t offset,
size_t size);
/* Return 1 if child is subregion of parent, else 0. */
int region_is_subregion(const struct region *p, const struct region *c);
static inline size_t region_offset(const struct region *r)
{
return r->offset;
}
static inline size_t region_sz(const struct region *r)
{
return r->size;
}
static inline size_t region_end(const struct region *r)
{
return region_offset(r) + region_sz(r);
}
static inline bool region_overlap(const struct region *r1, const struct region *r2)
{
return (region_end(r1) > region_offset(r2)) &&
(region_offset(r1) < region_end(r2));
}
static inline const struct region *region_device_region(
const struct region_device *rdev)
{
return &rdev->region;
}
static inline size_t region_device_sz(const struct region_device *rdev)
{
return region_sz(region_device_region(rdev));
}
static inline size_t region_device_offset(const struct region_device *rdev)
{
return region_offset(region_device_region(rdev));
}
static inline size_t region_device_end(const struct region_device *rdev)
{
return region_end(region_device_region(rdev));
}
/* Memory map entire region device. Same semantics as rdev_mmap() above. */
static inline void *rdev_mmap_full(const struct region_device *rd)
{
return rdev_mmap(rd, 0, region_device_sz(rd));
}
static inline int rdev_chain_full(struct region_device *child,
const struct region_device *parent)
{
/* Chain full size of parent. */
return rdev_chain(child, parent, 0, region_device_sz(parent));
}
/*
* Returns < 0 on error otherwise returns size of data read at provided
* offset filling in the buffer passed.
*
* You must ensure the buffer is large enough to hold the full region_device.
*/
static inline ssize_t rdev_read_full(const struct region_device *rd, void *b)
{
return rdev_readat(rd, b, 0, region_device_sz(rd));
}
/*
* Compute relative offset of the child (c) w.r.t. the parent (p). Returns < 0
* when child is not within the parent's region.
*/
ssize_t rdev_relative_offset(const struct region_device *p,
const struct region_device *c);
/* Helper functions to create an rdev that represents memory. */
int rdev_chain_mem(struct region_device *child, const void *base, size_t size);
int rdev_chain_mem_rw(struct region_device *child, void *base, size_t size);
struct mem_region_device {
char *base;
struct region_device rdev;
};
/* Initialize at runtime a mem_region_device. Should only be used for mappings
that need to fit right up to the edge of the physical address space. Most use
cases will want to use rdev_chain_mem() instead. */
void mem_region_device_ro_init(struct mem_region_device *mdev, void *base,
size_t size);
void mem_region_device_rw_init(struct mem_region_device *mdev, void *base,
size_t size);
extern const struct region_device_ops mem_rdev_ro_ops;
extern const struct region_device_ops mem_rdev_rw_ops;
/* Statically initialize mem_region_device. Should normally only be used for
const globals. Most use cases will want to use rdev_chain_mem() instead. */
#define MEM_REGION_DEV_INIT(base_, size_, ops_) \
{ \
.base = (void *)(base_), \
.rdev = REGION_DEV_INIT((ops_), 0, (size_)), \
}
#define MEM_REGION_DEV_RO_INIT(base_, size_) \
MEM_REGION_DEV_INIT(base_, size_, &mem_rdev_ro_ops) \
#define MEM_REGION_DEV_RW_INIT(base_, size_) \
MEM_REGION_DEV_INIT(base_, size_, &mem_rdev_rw_ops) \
struct mmap_helper_region_device {
struct mem_pool *pool;
struct region_device rdev;
};
#define MMAP_HELPER_DEV_INIT(ops_, offset_, size_, mpool_) \
{ \
.rdev = REGION_DEV_INIT((ops_), (offset_), (size_)), \
.pool = (mpool_), \
}
void *mmap_helper_rdev_mmap(const struct region_device *, size_t, size_t);
int mmap_helper_rdev_munmap(const struct region_device *, void *);
/*
* A translated region device provides the ability to publish a region device in one address
* space and use an access mechanism within another address space. The sub region is the window
* within the 1st address space and the request is modified prior to accessing the second
* address space provided by access_dev.
*
* Each xlate_region_device can support multiple translation windows described using
* xlate_window structure. The windows need not be contiguous in either address space. However,
* this poses restrictions on the operations being performed i.e. callers cannot perform
* operations across multiple windows of a translated region device. It is possible to support
* readat/writeat/eraseat by translating them into multiple calls one to access device in each
* window. However, mmap support is tricky because the caller expects that the memory mapped
* region is contiguous in both address spaces. Thus, to keep the semantics consistent for all
* region ops, xlate_region_device does not support any operations across the window
* boundary.
*
* Note: The platform is expected to ensure that the fmap description does not place any
* section (that will be operated using the translated region device) across multiple windows.
*/
struct xlate_window {
const struct region_device *access_dev;
struct region sub_region;
};
struct xlate_region_device {
size_t window_count;
const struct xlate_window *window_arr;
struct region_device rdev;
};
extern const struct region_device_ops xlate_rdev_ro_ops;
extern const struct region_device_ops xlate_rdev_rw_ops;
#define XLATE_REGION_DEV_INIT(window_arr_, parent_sz_, ops_) \
{ \
.window_count = ARRAY_SIZE(window_arr_), \
.window_arr = window_arr_, \
.rdev = REGION_DEV_INIT((ops_), 0, (parent_sz_)), \
}
#define XLATE_REGION_DEV_RO_INIT(window_arr_, parent_sz_) \
XLATE_REGION_DEV_INIT(window_arr_, parent_sz_, &xlate_rdev_ro_ops)
#define XLATE_REGION_DEV_RW_INIT(window_count_, window_arr_, parent_sz_) \
XLATE_REGION_DEV_INIT(window_arr_, parent_sz_, &xlate_rdev_rw_ops)
/* Helper to dynamically initialize xlate region device. */
void xlate_region_device_ro_init(struct xlate_region_device *xdev,
size_t window_count, const struct xlate_window *window_arr,
size_t parent_size);
void xlate_region_device_rw_init(struct xlate_region_device *xdev,
size_t window_count, const struct xlate_window *window_arr,
size_t parent_size);
void xlate_window_init(struct xlate_window *window, const struct region_device *access_dev,
size_t sub_region_offset, size_t sub_region_size);
/* This type can be used for incoherent access where the read and write
* operations are backed by separate drivers. An example is x86 systems
* with memory mapped media for reading but use a spi flash driver for
* writing. One needs to ensure using this object is appropriate in context. */
struct incoherent_rdev {
struct region_device rdev;
const struct region_device *read;
const struct region_device *write;
};
/* Initialize an incoherent_rdev based on the region as well as the read and
* write rdevs. The read and write rdevs should match in size to the passed
* in region. If not the initialization will fail returning NULL. Otherwise
* the function will return a pointer to the containing region_device to
* be used for region operations. Therefore, the lifetime of the returned
* pointer matches the lifetime of the incoherent_rdev object. Likewise,
* the lifetime of the read and write rdev need to match the lifetime of
* the incoherent_rdev object. */
const struct region_device *incoherent_rdev_init(struct incoherent_rdev *irdev,
const struct region *r,
const struct region_device *read,
const struct region_device *write);
#endif /* _REGION_H_ */