soc: qcom: smem: adjust for U-Boot
Port over the smem code to U-Boot and remove socinfo. Signed-off-by: Caleb Connolly <caleb.connolly@linaro.org> Reviewed-by: Simon Glass <sjg@chromium.org>
This commit is contained in:
committed by
Simon Glass
parent
6408b9a629
commit
efc20ac01f
@@ -4,6 +4,8 @@
|
||||
* Copyright (c) 2012-2013, The Linux Foundation. All rights reserved.
|
||||
*/
|
||||
|
||||
#define pr_fmt(fmt) "smem: " fmt
|
||||
|
||||
#include <dm/device.h>
|
||||
#include <dm/device_compat.h>
|
||||
#include <dm/ofnode.h>
|
||||
@@ -12,7 +14,6 @@
|
||||
#include <linux/ioport.h>
|
||||
#include <linux/sizes.h>
|
||||
#include <soc/qcom/smem.h>
|
||||
#include <soc/qcom/socinfo.h>
|
||||
|
||||
/*
|
||||
* The Qualcomm shared memory system is a allocate only heap structure that
|
||||
@@ -261,29 +262,21 @@ struct smem_region {
|
||||
|
||||
/**
|
||||
* struct qcom_smem - device data for the smem device
|
||||
* @dev: device pointer
|
||||
* @hwlock: reference to a hwspinlock
|
||||
* @ptable: virtual base of partition table
|
||||
* @global_partition: describes for global partition when in use
|
||||
* @partitions: list of partitions of current processor/host
|
||||
* @item_count: max accepted item number
|
||||
* @socinfo: platform device pointer
|
||||
* @num_regions: number of @regions
|
||||
* @regions: list of the memory regions defining the shared memory
|
||||
*/
|
||||
struct qcom_smem {
|
||||
struct device *dev;
|
||||
|
||||
struct hwspinlock *hwlock;
|
||||
|
||||
u32 item_count;
|
||||
struct platform_device *socinfo;
|
||||
struct smem_ptable *ptable;
|
||||
struct smem_partition global_partition;
|
||||
struct smem_partition partitions[SMEM_HOST_COUNT];
|
||||
|
||||
unsigned num_regions;
|
||||
struct smem_region regions[] __counted_by(num_regions);
|
||||
struct smem_region regions[];
|
||||
};
|
||||
|
||||
static void *
|
||||
@@ -352,36 +345,7 @@ static void *cached_entry_to_item(struct smem_private_entry *e)
|
||||
}
|
||||
|
||||
/* Pointer to the one and only smem handle */
|
||||
static struct qcom_smem *__smem;
|
||||
|
||||
/* Timeout (ms) for the trylock of remote spinlocks */
|
||||
#define HWSPINLOCK_TIMEOUT 1000
|
||||
|
||||
/* The qcom hwspinlock id is always plus one from the smem host id */
|
||||
#define SMEM_HOST_ID_TO_HWSPINLOCK_ID(__x) ((__x) + 1)
|
||||
|
||||
/**
|
||||
* qcom_smem_bust_hwspin_lock_by_host() - bust the smem hwspinlock for a host
|
||||
* @host: remote processor id
|
||||
*
|
||||
* Busts the hwspin_lock for the given smem host id. This helper is intended
|
||||
* for remoteproc drivers that manage remoteprocs with an equivalent smem
|
||||
* driver instance in the remote firmware. Drivers can force a release of the
|
||||
* smem hwspin_lock if the rproc unexpectedly goes into a bad state.
|
||||
*
|
||||
* Context: Process context.
|
||||
*
|
||||
* Returns: 0 on success, otherwise negative errno.
|
||||
*/
|
||||
int qcom_smem_bust_hwspin_lock_by_host(unsigned int host)
|
||||
{
|
||||
/* This function is for remote procs, so ignore SMEM_HOST_APPS */
|
||||
if (host == SMEM_HOST_APPS || host >= SMEM_HOST_COUNT)
|
||||
return -EINVAL;
|
||||
|
||||
return hwspin_lock_bust(__smem->hwlock, SMEM_HOST_ID_TO_HWSPINLOCK_ID(host));
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(qcom_smem_bust_hwspin_lock_by_host);
|
||||
static struct qcom_smem *__smem __section(".data") = NULL;
|
||||
|
||||
/**
|
||||
* qcom_smem_is_available() - Check if SMEM is available
|
||||
@@ -430,7 +394,7 @@ static int qcom_smem_alloc_private(struct qcom_smem *smem,
|
||||
/* Check that we don't grow into the cached region */
|
||||
alloc_size = sizeof(*hdr) + ALIGN(size, 8);
|
||||
if ((void *)hdr + alloc_size > cached) {
|
||||
dev_err(smem->dev, "Out of memory\n");
|
||||
log_err("Out of memory\n");
|
||||
return -ENOSPC;
|
||||
}
|
||||
|
||||
@@ -450,7 +414,7 @@ static int qcom_smem_alloc_private(struct qcom_smem *smem,
|
||||
|
||||
return 0;
|
||||
bad_canary:
|
||||
dev_err(smem->dev, "Found invalid canary in hosts %hu:%hu partition\n",
|
||||
log_err("Found invalid canary in hosts %hu:%hu partition\n",
|
||||
le16_to_cpu(phdr->host0), le16_to_cpu(phdr->host1));
|
||||
|
||||
return -EINVAL;
|
||||
@@ -501,27 +465,19 @@ static int qcom_smem_alloc_global(struct qcom_smem *smem,
|
||||
int qcom_smem_alloc(unsigned host, unsigned item, size_t size)
|
||||
{
|
||||
struct smem_partition *part;
|
||||
unsigned long flags;
|
||||
int ret;
|
||||
|
||||
if (!__smem)
|
||||
return -EPROBE_DEFER;
|
||||
|
||||
if (item < SMEM_ITEM_LAST_FIXED) {
|
||||
dev_err(__smem->dev,
|
||||
"Rejecting allocation of static entry %d\n", item);
|
||||
log_err("Rejecting allocation of static entry %d\n", item);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (WARN_ON(item >= __smem->item_count))
|
||||
return -EINVAL;
|
||||
|
||||
ret = hwspin_lock_timeout_irqsave(__smem->hwlock,
|
||||
HWSPINLOCK_TIMEOUT,
|
||||
&flags);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
if (host < SMEM_HOST_COUNT && __smem->partitions[host].virt_base) {
|
||||
part = &__smem->partitions[host];
|
||||
ret = qcom_smem_alloc_private(__smem, part, item, size);
|
||||
@@ -532,8 +488,6 @@ int qcom_smem_alloc(unsigned host, unsigned item, size_t size)
|
||||
ret = qcom_smem_alloc_global(__smem, item, size);
|
||||
}
|
||||
|
||||
hwspin_unlock_irqrestore(__smem->hwlock, &flags);
|
||||
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(qcom_smem_alloc);
|
||||
@@ -661,7 +615,7 @@ static void *qcom_smem_get_private(struct qcom_smem *smem,
|
||||
return ERR_PTR(-ENOENT);
|
||||
|
||||
invalid_canary:
|
||||
dev_err(smem->dev, "Found invalid canary in hosts %hu:%hu partition\n",
|
||||
log_err("Found invalid canary in hosts %hu:%hu partition\n",
|
||||
le16_to_cpu(phdr->host0), le16_to_cpu(phdr->host1));
|
||||
|
||||
return ERR_PTR(-EINVAL);
|
||||
@@ -797,61 +751,6 @@ phys_addr_t qcom_smem_virt_to_phys(void *p)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(qcom_smem_virt_to_phys);
|
||||
|
||||
/**
|
||||
* qcom_smem_get_soc_id() - return the SoC ID
|
||||
* @id: On success, we return the SoC ID here.
|
||||
*
|
||||
* Look up SoC ID from HW/SW build ID and return it.
|
||||
*
|
||||
* Return: 0 on success, negative errno on failure.
|
||||
*/
|
||||
int qcom_smem_get_soc_id(u32 *id)
|
||||
{
|
||||
struct socinfo *info;
|
||||
|
||||
info = qcom_smem_get(QCOM_SMEM_HOST_ANY, SMEM_HW_SW_BUILD_ID, NULL);
|
||||
if (IS_ERR(info))
|
||||
return PTR_ERR(info);
|
||||
|
||||
*id = __le32_to_cpu(info->id);
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(qcom_smem_get_soc_id);
|
||||
|
||||
/**
|
||||
* qcom_smem_get_feature_code() - return the feature code
|
||||
* @code: On success, return the feature code here.
|
||||
*
|
||||
* Look up the feature code identifier from SMEM and return it.
|
||||
*
|
||||
* Return: 0 on success, negative errno on failure.
|
||||
*/
|
||||
int qcom_smem_get_feature_code(u32 *code)
|
||||
{
|
||||
struct socinfo *info;
|
||||
u32 raw_code;
|
||||
|
||||
info = qcom_smem_get(QCOM_SMEM_HOST_ANY, SMEM_HW_SW_BUILD_ID, NULL);
|
||||
if (IS_ERR(info))
|
||||
return PTR_ERR(info);
|
||||
|
||||
/* This only makes sense for socinfo >= 16 */
|
||||
if (__le32_to_cpu(info->fmt) < SOCINFO_VERSION(0, 16))
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
raw_code = __le32_to_cpu(info->feature_code);
|
||||
|
||||
/* Ensure the value makes sense */
|
||||
if (raw_code > SOCINFO_FC_INT_MAX)
|
||||
raw_code = SOCINFO_FC_UNKNOWN;
|
||||
|
||||
*code = raw_code;
|
||||
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(qcom_smem_get_feature_code);
|
||||
|
||||
static int qcom_smem_get_sbl_version(struct qcom_smem *smem)
|
||||
{
|
||||
struct smem_header *header;
|
||||
@@ -874,8 +773,7 @@ static struct smem_ptable *qcom_smem_get_ptable(struct qcom_smem *smem)
|
||||
|
||||
version = le32_to_cpu(ptable->version);
|
||||
if (version != 1) {
|
||||
dev_err(smem->dev,
|
||||
"Unsupported partition header version %d\n", version);
|
||||
log_err("Unsupported partition header version %d\n", version);
|
||||
return ERR_PTR(-EINVAL);
|
||||
}
|
||||
return ptable;
|
||||
@@ -907,40 +805,40 @@ qcom_smem_partition_header(struct qcom_smem *smem,
|
||||
struct smem_ptable_entry *entry, u16 host0, u16 host1)
|
||||
{
|
||||
struct smem_partition_header *header;
|
||||
u32 phys_addr;
|
||||
u64 phys_addr;
|
||||
u32 size;
|
||||
|
||||
phys_addr = smem->regions[0].aux_base + le32_to_cpu(entry->offset);
|
||||
header = devm_ioremap_wc(smem->dev, phys_addr, le32_to_cpu(entry->size));
|
||||
header = (void *)phys_addr; // devm_ioremap_wc()
|
||||
|
||||
if (!header)
|
||||
return NULL;
|
||||
|
||||
if (memcmp(header->magic, SMEM_PART_MAGIC, sizeof(header->magic))) {
|
||||
dev_err(smem->dev, "bad partition magic %4ph\n", header->magic);
|
||||
log_err("bad partition magic %4ph\n", header->magic);
|
||||
return NULL;
|
||||
}
|
||||
|
||||
if (host0 != le16_to_cpu(header->host0)) {
|
||||
dev_err(smem->dev, "bad host0 (%hu != %hu)\n",
|
||||
host0, le16_to_cpu(header->host0));
|
||||
log_err("bad host0 (%hu != %hu)\n",
|
||||
host0, le16_to_cpu(header->host0));
|
||||
return NULL;
|
||||
}
|
||||
if (host1 != le16_to_cpu(header->host1)) {
|
||||
dev_err(smem->dev, "bad host1 (%hu != %hu)\n",
|
||||
host1, le16_to_cpu(header->host1));
|
||||
log_err("bad host1 (%hu != %hu)\n",
|
||||
host1, le16_to_cpu(header->host1));
|
||||
return NULL;
|
||||
}
|
||||
|
||||
size = le32_to_cpu(header->size);
|
||||
if (size != le32_to_cpu(entry->size)) {
|
||||
dev_err(smem->dev, "bad partition size (%u != %u)\n",
|
||||
log_err("bad partition size (%u != %u)\n",
|
||||
size, le32_to_cpu(entry->size));
|
||||
return NULL;
|
||||
}
|
||||
|
||||
if (le32_to_cpu(header->offset_free_uncached) > size) {
|
||||
dev_err(smem->dev, "bad partition free uncached (%u > %u)\n",
|
||||
log_err("bad partition free uncached (%u > %u)\n",
|
||||
le32_to_cpu(header->offset_free_uncached), size);
|
||||
return NULL;
|
||||
}
|
||||
@@ -957,7 +855,7 @@ static int qcom_smem_set_global_partition(struct qcom_smem *smem)
|
||||
int i;
|
||||
|
||||
if (smem->global_partition.virt_base) {
|
||||
dev_err(smem->dev, "Already found the global partition\n");
|
||||
log_err("Already found the global partition\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@@ -982,7 +880,7 @@ static int qcom_smem_set_global_partition(struct qcom_smem *smem)
|
||||
}
|
||||
|
||||
if (!found) {
|
||||
dev_err(smem->dev, "Missing entry for global partition\n");
|
||||
log_err("Missing entry for global partition\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@@ -1031,12 +929,12 @@ qcom_smem_enumerate_partitions(struct qcom_smem *smem, u16 local_host)
|
||||
continue;
|
||||
|
||||
if (remote_host >= SMEM_HOST_COUNT) {
|
||||
dev_err(smem->dev, "bad host %u\n", remote_host);
|
||||
log_err("bad host %u\n", remote_host);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (smem->partitions[remote_host].virt_base) {
|
||||
dev_err(smem->dev, "duplicate host %u\n", remote_host);
|
||||
log_err("duplicate host %u\n", remote_host);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
@@ -1059,10 +957,10 @@ static int qcom_smem_map_toc(struct qcom_smem *smem, struct smem_region *region)
|
||||
u32 ptable_start;
|
||||
|
||||
/* map starting 4K for smem header */
|
||||
region->virt_base = devm_ioremap_wc(smem->dev, region->aux_base, SZ_4K);
|
||||
region->virt_base = (void *)region->aux_base;
|
||||
ptable_start = region->aux_base + region->size - SZ_4K;
|
||||
/* map last 4k for toc */
|
||||
smem->ptable = devm_ioremap_wc(smem->dev, ptable_start, SZ_4K);
|
||||
smem->ptable = (struct smem_ptable *)(u64)ptable_start;
|
||||
|
||||
if (!region->virt_base || !smem->ptable)
|
||||
return -ENOMEM;
|
||||
@@ -1072,12 +970,12 @@ static int qcom_smem_map_toc(struct qcom_smem *smem, struct smem_region *region)
|
||||
|
||||
static int qcom_smem_map_global(struct qcom_smem *smem, u32 size)
|
||||
{
|
||||
u32 phys_addr;
|
||||
u64 phys_addr;
|
||||
|
||||
phys_addr = smem->regions[0].aux_base;
|
||||
|
||||
smem->regions[0].size = size;
|
||||
smem->regions[0].virt_base = devm_ioremap_wc(smem->dev, phys_addr, size);
|
||||
smem->regions[0].virt_base = (void *)phys_addr;
|
||||
|
||||
if (!smem->regions[0].virt_base)
|
||||
return -ENOMEM;
|
||||
@@ -1085,39 +983,14 @@ static int qcom_smem_map_global(struct qcom_smem *smem, u32 size)
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int qcom_smem_resolve_mem(struct qcom_smem *smem, const char *name,
|
||||
struct smem_region *region)
|
||||
{
|
||||
struct device *dev = smem->dev;
|
||||
struct device_node *np;
|
||||
struct resource r;
|
||||
int ret;
|
||||
|
||||
np = of_parse_phandle(dev->of_node, name, 0);
|
||||
if (!np) {
|
||||
dev_err(dev, "No %s specified\n", name);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
ret = of_address_to_resource(np, 0, &r);
|
||||
of_node_put(np);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
region->aux_base = r.start;
|
||||
region->size = resource_size(&r);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int qcom_smem_probe(struct platform_device *pdev)
|
||||
int qcom_smem_init(void)
|
||||
{
|
||||
struct smem_header *header;
|
||||
struct reserved_mem *rmem;
|
||||
struct qcom_smem *smem;
|
||||
unsigned long flags;
|
||||
int num_regions;
|
||||
int hwlock_id;
|
||||
fdt_size_t reg_size = 0;
|
||||
u32 phandle;
|
||||
ofnode node, mem_node;
|
||||
u32 version;
|
||||
u32 size;
|
||||
int ret;
|
||||
@@ -1127,83 +1000,56 @@ static int qcom_smem_probe(struct platform_device *pdev)
|
||||
return 0;
|
||||
|
||||
num_regions = 1;
|
||||
if (of_property_present(pdev->dev.of_node, "qcom,rpm-msg-ram"))
|
||||
num_regions++;
|
||||
|
||||
smem = devm_kzalloc(&pdev->dev, struct_size(smem, regions, num_regions),
|
||||
node = ofnode_by_compatible(ofnode_root(), "qcom,smem");
|
||||
if (!ofnode_valid(node))
|
||||
return -ENODEV;
|
||||
|
||||
if (ofnode_has_property(node, "memory-region")) {
|
||||
ofnode_read_u32(node, "memory-region", &phandle);
|
||||
mem_node = ofnode_get_by_phandle(phandle);
|
||||
} else {
|
||||
mem_node = node;
|
||||
}
|
||||
|
||||
smem = kzalloc(sizeof(struct smem_region) * num_regions +
|
||||
sizeof(struct qcom_smem),
|
||||
GFP_KERNEL);
|
||||
if (!smem)
|
||||
if (!smem) {
|
||||
log_err("Failed to allocate memory for smem\n");
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
smem->dev = &pdev->dev;
|
||||
smem->num_regions = num_regions;
|
||||
|
||||
rmem = of_reserved_mem_lookup(pdev->dev.of_node);
|
||||
if (rmem) {
|
||||
smem->regions[0].aux_base = rmem->base;
|
||||
smem->regions[0].size = rmem->size;
|
||||
} else {
|
||||
/*
|
||||
* Fall back to the memory-region reference, if we're not a
|
||||
* reserved-memory node.
|
||||
*/
|
||||
ret = qcom_smem_resolve_mem(smem, "memory-region", &smem->regions[0]);
|
||||
if (ret)
|
||||
return ret;
|
||||
smem->regions[0].aux_base = ofnode_get_addr(mem_node);
|
||||
reg_size = ofnode_get_size(mem_node);
|
||||
if (smem->regions[0].aux_base == FDT_ADDR_T_NONE) {
|
||||
log_err("Failed to get base address\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
if (num_regions > 1) {
|
||||
ret = qcom_smem_resolve_mem(smem, "qcom,rpm-msg-ram", &smem->regions[1]);
|
||||
if (ret)
|
||||
return ret;
|
||||
}
|
||||
|
||||
smem->regions[0].size = reg_size;
|
||||
|
||||
ret = qcom_smem_map_toc(smem, &smem->regions[0]);
|
||||
if (ret)
|
||||
if (ret) {
|
||||
log_err("Failed to map toc\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
for (i = 1; i < num_regions; i++) {
|
||||
smem->regions[i].virt_base = devm_ioremap_wc(&pdev->dev,
|
||||
smem->regions[i].aux_base,
|
||||
smem->regions[i].size);
|
||||
if (!smem->regions[i].virt_base) {
|
||||
dev_err(&pdev->dev, "failed to remap %pa\n", &smem->regions[i].aux_base);
|
||||
return -ENOMEM;
|
||||
}
|
||||
smem->regions[i].virt_base = (void *)smem->regions[i].aux_base;
|
||||
}
|
||||
|
||||
header = smem->regions[0].virt_base;
|
||||
if (le32_to_cpu(header->initialized) != 1 ||
|
||||
le32_to_cpu(header->reserved)) {
|
||||
dev_err(&pdev->dev, "SMEM is not initialized by SBL\n");
|
||||
log_err("SMEM is not initialized by SBL\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
hwlock_id = of_hwspin_lock_get_id(pdev->dev.of_node, 0);
|
||||
if (hwlock_id < 0) {
|
||||
if (hwlock_id != -EPROBE_DEFER)
|
||||
dev_err(&pdev->dev, "failed to retrieve hwlock\n");
|
||||
return hwlock_id;
|
||||
}
|
||||
|
||||
smem->hwlock = hwspin_lock_request_specific(hwlock_id);
|
||||
if (!smem->hwlock)
|
||||
return -ENXIO;
|
||||
|
||||
ret = hwspin_lock_timeout_irqsave(smem->hwlock, HWSPINLOCK_TIMEOUT, &flags);
|
||||
if (ret)
|
||||
return ret;
|
||||
size = readl_relaxed(&header->available) + readl_relaxed(&header->free_offset);
|
||||
hwspin_unlock_irqrestore(smem->hwlock, &flags);
|
||||
|
||||
version = qcom_smem_get_sbl_version(smem);
|
||||
/*
|
||||
* smem header mapping is required only in heap version scheme, so unmap
|
||||
* it here. It will be remapped in qcom_smem_map_global() when whole
|
||||
* partition is mapped again.
|
||||
*/
|
||||
devm_iounmap(smem->dev, smem->regions[0].virt_base);
|
||||
switch (version >> 16) {
|
||||
case SMEM_GLOBAL_PART_VERSION:
|
||||
ret = qcom_smem_set_global_partition(smem);
|
||||
@@ -1216,62 +1062,18 @@ static int qcom_smem_probe(struct platform_device *pdev)
|
||||
smem->item_count = SMEM_ITEM_COUNT;
|
||||
break;
|
||||
default:
|
||||
dev_err(&pdev->dev, "Unsupported SMEM version 0x%x\n", version);
|
||||
log_err("Unsupported SMEM version 0x%x\n", version);
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
BUILD_BUG_ON(SMEM_HOST_APPS >= SMEM_HOST_COUNT);
|
||||
ret = qcom_smem_enumerate_partitions(smem, SMEM_HOST_APPS);
|
||||
if (ret < 0 && ret != -ENOENT)
|
||||
if (ret < 0 && ret != -ENOENT) {
|
||||
log_err("Failed to enumerate partitions\n");
|
||||
return ret;
|
||||
}
|
||||
|
||||
__smem = smem;
|
||||
|
||||
smem->socinfo = platform_device_register_data(&pdev->dev, "qcom-socinfo",
|
||||
PLATFORM_DEVID_NONE, NULL,
|
||||
0);
|
||||
if (IS_ERR(smem->socinfo))
|
||||
dev_dbg(&pdev->dev, "failed to register socinfo device\n");
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void qcom_smem_remove(struct platform_device *pdev)
|
||||
{
|
||||
platform_device_unregister(__smem->socinfo);
|
||||
|
||||
hwspin_lock_free(__smem->hwlock);
|
||||
__smem = NULL;
|
||||
}
|
||||
|
||||
static const struct of_device_id qcom_smem_of_match[] = {
|
||||
{ .compatible = "qcom,smem" },
|
||||
{}
|
||||
};
|
||||
MODULE_DEVICE_TABLE(of, qcom_smem_of_match);
|
||||
|
||||
static struct platform_driver qcom_smem_driver = {
|
||||
.probe = qcom_smem_probe,
|
||||
.remove_new = qcom_smem_remove,
|
||||
.driver = {
|
||||
.name = "qcom-smem",
|
||||
.of_match_table = qcom_smem_of_match,
|
||||
.suppress_bind_attrs = true,
|
||||
},
|
||||
};
|
||||
|
||||
static int __init qcom_smem_init(void)
|
||||
{
|
||||
return platform_driver_register(&qcom_smem_driver);
|
||||
}
|
||||
arch_initcall(qcom_smem_init);
|
||||
|
||||
static void __exit qcom_smem_exit(void)
|
||||
{
|
||||
platform_driver_unregister(&qcom_smem_driver);
|
||||
}
|
||||
module_exit(qcom_smem_exit)
|
||||
|
||||
MODULE_AUTHOR("Bjorn Andersson <bjorn.andersson@sonymobile.com>");
|
||||
MODULE_DESCRIPTION("Qualcomm Shared Memory Manager");
|
||||
MODULE_LICENSE("GPL v2");
|
||||
|
||||
@@ -4,17 +4,12 @@
|
||||
|
||||
#define QCOM_SMEM_HOST_ANY -1
|
||||
|
||||
int qcom_smem_init(void);
|
||||
|
||||
bool qcom_smem_is_available(void);
|
||||
int qcom_smem_alloc(unsigned host, unsigned item, size_t size);
|
||||
void *qcom_smem_get(unsigned host, unsigned item, size_t *size);
|
||||
|
||||
int qcom_smem_get_free_space(unsigned host);
|
||||
|
||||
phys_addr_t qcom_smem_virt_to_phys(void *p);
|
||||
|
||||
int qcom_smem_get_soc_id(u32 *id);
|
||||
int qcom_smem_get_feature_code(u32 *code);
|
||||
|
||||
int qcom_smem_bust_hwspin_lock_by_host(unsigned int host);
|
||||
|
||||
#endif
|
||||
|
||||
Reference in New Issue
Block a user