drivers: dma: remove qmsi DMA driver

No users of this driver after dropping quark platforms.

Signed-off-by: Anas Nashif <anas.nashif@intel.com>
This commit is contained in:
Anas Nashif 2019-09-17 12:51:28 -04:00 committed by Maureen Helm
commit 5cde4ea245
5 changed files with 0 additions and 348 deletions

View file

@ -2,7 +2,6 @@
zephyr_library()
zephyr_library_sources_ifdef(CONFIG_DMA_QMSI dma_qmsi.c)
zephyr_library_sources_ifdef(CONFIG_DMA_SAM_XDMAC dma_sam_xdmac.c)
zephyr_library_sources_ifdef(CONFIG_DMA_STM32F4X dma_stm32f4x.c)
zephyr_library_sources_ifdef(CONFIG_DMA_CAVS dma_cavs.c)

View file

@ -53,8 +53,6 @@ module = DMA
module-str = dma
source "subsys/logging/Kconfig.template.log_config"
source "drivers/dma/Kconfig.qmsi"
source "drivers/dma/Kconfig.stm32f4x"
source "drivers/dma/Kconfig.sam_xdmac"

View file

@ -1,13 +0,0 @@
# Kconfig - DMA configuration options
#
#
# Copyright (c) 2016 Intel Corporation
#
# SPDX-License-Identifier: Apache-2.0
#
menuconfig DMA_QMSI
bool "Enable QMSI DMA driver"
depends on QMSI
help
QMSI DMA driver.

View file

@ -1,331 +0,0 @@
/*
* Copyright (c) 2016 Intel Corporation.
*
* SPDX-License-Identifier: Apache-2.0
*/
#include <errno.h>
#include <stdio.h>
#include <kernel.h>
#include <soc.h>
#include <device.h>
#include <init.h>
#include <drivers/dma.h>
#include "qm_dma.h"
#include "qm_isr.h"
#include "clk.h"
#define CYCLE_NOP \
__asm__ __volatile__ ("nop"); \
__asm__ __volatile__ ("nop"); \
__asm__ __volatile__ ("nop"); \
__asm__ __volatile__ ("nop")
struct dma_qmsi_config_info {
qm_dma_t instance; /* Controller instance. */
};
struct dma_qmsi_context {
u32_t index;
struct device *dev;
};
struct dma_qmsi_driver_data {
void (*transfer[QM_DMA_CHANNEL_NUM])(struct device *dev, void *data);
void (*error[QM_DMA_CHANNEL_NUM])(struct device *dev, void *data);
void *callback_data[QM_DMA_CHANNEL_NUM];
#ifdef CONFIG_DEVICE_POWER_MANAGEMENT
u32_t device_power_state;
qm_dma_context_t saved_ctx;
#endif
void (*dma_user_callback[QM_DMA_CHANNEL_NUM])(void *arg,
u32_t channel_id,
int error_code);
};
static struct dma_qmsi_context dma_context[QM_DMA_CHANNEL_NUM];
static void dma_qmsi_config(struct device *dev);
static void dma_drv_callback(void *callback_context, u32_t len,
int error_code)
{
struct dma_qmsi_context *context = callback_context;
struct dma_qmsi_driver_data *data;
u32_t channel;
channel = context->index;
data = context->dev->driver_data;
data->dma_user_callback[channel](data->callback_data[channel],
channel, error_code);
}
static int width_index(u32_t num_bytes, u32_t *index)
{
switch (num_bytes) {
case 1:
*index = QM_DMA_TRANS_WIDTH_8;
break;
case 2:
*index = QM_DMA_TRANS_WIDTH_16;
break;
case 4:
*index = QM_DMA_TRANS_WIDTH_32;
break;
case 8:
*index = QM_DMA_TRANS_WIDTH_64;
break;
case 16:
*index = QM_DMA_TRANS_WIDTH_128;
break;
case 32:
*index = QM_DMA_TRANS_WIDTH_256;
break;
default:
return -ENOTSUP;
}
return 0;
}
static int bst_index(u32_t num_units, u32_t *index)
{
switch (num_units) {
case 1:
*index = QM_DMA_BURST_TRANS_LENGTH_1;
break;
case 4:
*index = QM_DMA_BURST_TRANS_LENGTH_4;
break;
case 8:
*index = QM_DMA_BURST_TRANS_LENGTH_8;
break;
case 16:
*index = QM_DMA_BURST_TRANS_LENGTH_16;
break;
case 32:
*index = QM_DMA_BURST_TRANS_LENGTH_32;
break;
case 64:
*index = QM_DMA_BURST_TRANS_LENGTH_64;
break;
case 128:
*index = QM_DMA_BURST_TRANS_LENGTH_128;
break;
case 256:
*index = QM_DMA_BURST_TRANS_LENGTH_256;
break;
default:
return -ENOTSUP;
}
return 0;
}
static int dma_qmsi_chan_config(struct device *dev, u32_t channel,
struct dma_config *config)
{
const struct dma_qmsi_config_info *info = dev->config->config_info;
struct dma_qmsi_driver_data *data = dev->driver_data;
qm_dma_transfer_t qmsi_transfer_cfg = { 0 };
qm_dma_channel_config_t qmsi_cfg = { 0 };
u32_t temp = 0U;
int ret = 0;
if (config->block_count != 1U) {
return -ENOTSUP;
}
qmsi_cfg.handshake_interface = (qm_dma_handshake_interface_t)
config->dma_slot;
qmsi_cfg.channel_direction = (qm_dma_channel_direction_t)
config->channel_direction;
ret = width_index(config->source_data_size, &temp);
if (ret != 0) {
return ret;
}
qmsi_cfg.source_transfer_width = (qm_dma_transfer_width_t) temp;
ret = width_index(config->dest_data_size, &temp);
if (ret != 0) {
return ret;
}
qmsi_cfg.destination_transfer_width = (qm_dma_transfer_width_t) temp;
ret = bst_index(config->dest_burst_length, &temp);
if (ret != 0) {
return ret;
}
qmsi_cfg.destination_burst_length = (qm_dma_burst_length_t) temp;
ret = bst_index(config->source_burst_length, &temp);
if (ret != 0) {
return ret;
}
qmsi_cfg.source_burst_length = (qm_dma_burst_length_t) temp;
/* TODO: add support for using other DMA transfer types. */
qmsi_cfg.transfer_type = QM_DMA_TYPE_SINGLE;
data->callback_data[channel] = config->callback_arg;
data->dma_user_callback[channel] = config->dma_callback;
dma_context[channel].index = channel;
dma_context[channel].dev = dev;
qmsi_cfg.callback_context = &dma_context[channel];
qmsi_cfg.client_callback = dma_drv_callback;
ret = qm_dma_channel_set_config(info->instance, channel, &qmsi_cfg);
if (ret != 0) {
return ret;
}
qmsi_transfer_cfg.block_size = config->head_block->block_size;
qmsi_transfer_cfg.source_address = (u32_t *)
config->head_block->source_address;
qmsi_transfer_cfg.destination_address = (u32_t *)
config->head_block->dest_address;
return qm_dma_transfer_set_config(info->instance, channel,
&qmsi_transfer_cfg);
}
static int dma_qmsi_start(struct device *dev, u32_t channel)
{
int ret;
const struct dma_qmsi_config_info *info = dev->config->config_info;
ret = qm_dma_transfer_start(info->instance, channel);
CYCLE_NOP;
return ret;
}
static int dma_qmsi_stop(struct device *dev, u32_t channel)
{
const struct dma_qmsi_config_info *info = dev->config->config_info;
return qm_dma_transfer_terminate(info->instance, channel);
}
static const struct dma_driver_api dma_funcs = {
.config = dma_qmsi_chan_config,
.start = dma_qmsi_start,
.stop = dma_qmsi_stop
};
#ifdef CONFIG_DEVICE_POWER_MANAGEMENT
static void dma_qmsi_set_power_state(struct device *dev, u32_t power_state)
{
struct dma_qmsi_driver_data *ctx = dev->driver_data;
ctx->device_power_state = power_state;
}
static u32_t dma_qmsi_get_power_state(struct device *dev)
{
struct dma_qmsi_driver_data *ctx = dev->driver_data;
return ctx->device_power_state;
}
#else
#define dma_qmsi_set_power_state(...)
#endif
int dma_qmsi_init(struct device *dev)
{
const struct dma_qmsi_config_info *info = dev->config->config_info;
dma_qmsi_config(dev);
qm_dma_init(info->instance);
dma_qmsi_set_power_state(dev, DEVICE_PM_ACTIVE_STATE);
return 0;
}
static const struct dma_qmsi_config_info dma_qmsi_config_data = {
.instance = QM_DMA_0,
};
static struct dma_qmsi_driver_data dma_qmsi_dev_data;
#ifdef CONFIG_DEVICE_POWER_MANAGEMENT
static int dma_suspend_device(struct device *dev)
{
const struct dma_qmsi_config_info *info = dev->config->config_info;
struct dma_qmsi_driver_data *ctx = dev->driver_data;
qm_dma_save_context(info->instance, &ctx->saved_ctx);
dma_qmsi_set_power_state(dev, DEVICE_PM_SUSPEND_STATE);
return 0;
}
static int dma_resume_device(struct device *dev)
{
const struct dma_qmsi_config_info *info = dev->config->config_info;
struct dma_qmsi_driver_data *ctx = dev->driver_data;
qm_dma_restore_context(info->instance, &ctx->saved_ctx);
dma_qmsi_set_power_state(dev, DEVICE_PM_ACTIVE_STATE);
return 0;
}
static int dma_qmsi_device_ctrl(struct device *dev, u32_t ctrl_command,
void *context, device_pm_cb cb, void *arg)
{
int ret = 0;
if (ctrl_command == DEVICE_PM_SET_POWER_STATE) {
if (*((u32_t *)context) == DEVICE_PM_SUSPEND_STATE) {
ret = dma_suspend_device(dev);
} else if (*((u32_t *)context) == DEVICE_PM_ACTIVE_STATE) {
ret = dma_resume_device(dev);
}
} else if (ctrl_command == DEVICE_PM_GET_POWER_STATE) {
*((u32_t *)context) = dma_qmsi_get_power_state(dev);
}
if (cb) {
cb(dev, ret, context, arg);
}
return ret;
}
#endif
DEVICE_DEFINE(dma_qmsi, CONFIG_DMA_0_NAME, &dma_qmsi_init, dma_qmsi_device_ctrl,
&dma_qmsi_dev_data, &dma_qmsi_config_data, POST_KERNEL,
CONFIG_KERNEL_INIT_PRIORITY_DEVICE, (void *)&dma_funcs);
static void dma_qmsi_config(struct device *dev)
{
ARG_UNUSED(dev);
IRQ_CONNECT(IRQ_GET_NUMBER(QM_IRQ_DMA_0_INT_0), CONFIG_DMA_0_IRQ_PRI,
qm_dma_0_isr_0, DEVICE_GET(dma_qmsi), 0);
irq_enable(IRQ_GET_NUMBER(QM_IRQ_DMA_0_INT_0));
QM_IR_UNMASK_INTERRUPTS(QM_INTERRUPT_ROUTER->dma_0_int_0_mask);
IRQ_CONNECT(IRQ_GET_NUMBER(QM_IRQ_DMA_0_INT_1), CONFIG_DMA_0_IRQ_PRI,
qm_dma_0_isr_1, DEVICE_GET(dma_qmsi), 0);
irq_enable(IRQ_GET_NUMBER(QM_IRQ_DMA_0_INT_1));
QM_IR_UNMASK_INTERRUPTS(QM_INTERRUPT_ROUTER->dma_0_int_1_mask);
IRQ_CONNECT(IRQ_GET_NUMBER(QM_IRQ_DMA_0_ERROR_INT),
CONFIG_DMA_0_IRQ_PRI, qm_dma_0_error_isr,
DEVICE_GET(dma_qmsi), 0);
irq_enable(IRQ_GET_NUMBER(QM_IRQ_DMA_0_ERROR_INT));
#if (QM_LAKEMONT)
QM_INTERRUPT_ROUTER->dma_0_error_int_mask &= ~QM_IR_DMA_ERROR_HOST_MASK;
#elif (QM_SENSOR)
QM_INTERRUPT_ROUTER->dma_0_error_int_mask &= ~QM_IR_DMA_ERROR_SS_MASK;
#endif
}

View file

@ -1,5 +1,4 @@
CONFIG_DMA=y
CONFIG_DMA_QMSI=y
CONFIG_ZTEST=y
CONFIG_CONSOLE_HANDLER=y
CONFIG_SHELL=y