From cc432b761c884a0bd8e9d83db0a4e26109fc08b1 Mon Sep 17 00:00:00 2001 From: chen <15335560115@163.com> Date: 星期五, 08 十一月 2024 15:35:38 +0800 Subject: [PATCH] 安邦手环GPS删除部分无用数据和修改4G波特率9600出厂测试固件 --- keil/include/drivers/mk_spi.c | 923 +++++++++++++++++++++++++++++++++++++++++++++++++++++++++ 1 files changed, 923 insertions(+), 0 deletions(-) diff --git a/keil/include/drivers/mk_spi.c b/keil/include/drivers/mk_spi.c new file mode 100644 index 0000000..431f7d0 --- /dev/null +++ b/keil/include/drivers/mk_spi.c @@ -0,0 +1,923 @@ +/* + * Copyright (c) 2019-2023 Beijing Hanwei Innovation Technology Ltd. Co. and + * its subsidiaries and affiliates (collectly called MKSEMI). + * + * All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions are met: + * + * 1. Redistributions of source code must retain the above copyright notice, + * this list of conditions and the following disclaimer. + * + * 2. Redistributions in binary form, except as embedded into an MKSEMI + * integrated circuit in a product or a software update for such product, + * must reproduce the above copyright notice, this list of conditions and + * the following disclaimer in the documentation and/or other materials + * provided with the distribution. + * + * 3. Neither the name of MKSEMI nor the names of its contributors may be used + * to endorse or promote products derived from this software without + * specific prior written permission. + * + * 4. This software, with or without modification, must only be used with a + * MKSEMI integrated circuit. + * + * 5. Any software provided in binary form under this license must not be + * reverse engineered, decompiled, modified and/or disassembled. + * + * THIS SOFTWARE IS PROVIDED BY MKSEMI "AS IS" AND ANY EXPRESS OR IMPLIED + * WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF + * MERCHANTABILITY, NONINFRINGEMENT, AND FITNESS FOR A PARTICULAR PURPOSE ARE + * DISCLAIMED. IN NO EVENT SHALL MKSEMI OR CONTRIBUTORS BE LIABLE FOR ANY + * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES + * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; + * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND + * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF + * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "mk_spi.h" +#include "mk_trace.h" +#include "mk_clock.h" +#include "mk_reset.h" +#include "string.h" + +static struct SPI_HANDLE_T spi_handle[SPI_MAX_NUM] = { + { + .base = SPI0, + .irq = SPI0_IRQn, + .dma_rx_ch = DMA_CH0, + .dma_tx_ch = DMA_CH1, + .tx_dummy = 0, + }, + { + .base = SPI1, + .irq = SPI1_IRQn, + .dma_rx_ch = DMA_CH4, + .dma_tx_ch = DMA_CH5, + .tx_dummy = 0, + }, +}; + +#if SPI_DMA_MODE_EN +static void spi_dma_callback(void *ch, uint32_t err_code); +static void spi_dma_abort_callback(void *ch, uint32_t err_code); +#endif + +static uint8_t spi_get_frame_size(enum SPI_DEV_T id) +{ + uint8_t bits, cnt; + + bits = GET_BIT_FIELD(spi_handle[id].base->CTRL0, SPI_CTRL0_DSS_MSK, SPI_CTRL0_DSS_POS); + if (bits < 8) + { + cnt = 1; + } + else + { + cnt = 2; + } + + return cnt; +} + +static int spi_state_set(enum SPI_DEV_T id, enum SPI_STATE_T state) +{ + int ret = DRV_OK; + uint32_t lock = int_lock(); + + // update state + switch (spi_handle[id].state) + { + case SPI_STATE_READY: + spi_handle[id].state = state; + break; + case SPI_STATE_BUSY_RX: + if (state == SPI_STATE_BUSY_TX) + { + spi_handle[id].state = SPI_STATE_BUSY_TX_RX; + } + else + { + ret = DRV_BUSY; + } + break; + case SPI_STATE_BUSY_TX: + if (state == SPI_STATE_BUSY_RX) + { + spi_handle[id].state = SPI_STATE_BUSY_TX_RX; + } + else + { + ret = DRV_BUSY; + } + break; + case SPI_STATE_BUSY_TX_RX: + ret = DRV_BUSY; + break; + case SPI_STATE_RESET: + case SPI_STATE_TIMEOUT: + case SPI_STATE_ERROR: + ret = DRV_ERROR; + break; + } + int_unlock(lock); + + return ret; +} + +static void spi_state_clear(enum SPI_DEV_T id, enum SPI_STATE_T state) +{ + uint32_t lock = int_lock(); + // update state + spi_handle[id].state &= ~state; + if (spi_handle[id].state == 0) + { + spi_handle[id].state = SPI_STATE_READY; + } + int_unlock(lock); +} + +enum SPI_STATE_T spi_state_get(enum SPI_DEV_T id) +{ + return spi_handle[id].state; +} + +bool spi_is_busy(enum SPI_DEV_T id) +{ + return ((spi_handle[id].state != SPI_STATE_RESET) && (spi_handle[id].base->CTRL1 & SPI_CTRL1_SSE_MSK) && + (spi_handle[id].base->STATUS0 & SPI_STATUS0_BSY_MSK)); +} + +int spi_open(enum SPI_DEV_T id, struct SPI_CFG_T *config) +{ + if ((id >= SPI_MAX_NUM) && (config == NULL)) + { + return DRV_ERROR; + } + else if (id == SPI_ID0) + { + // enable SPI0 clock + clock_enable(CLOCK_SPI0); + reset_module(RESET_MODULE_SPI0); + } + else if (id == SPI_ID1) + { + // enable SPI1 clock + clock_enable(CLOCK_SPI1); + reset_module(RESET_MODULE_SPI1); + } + + uint32_t val; + uint32_t cpsdvsr, scr; + uint32_t spi_clk = clock_get_frequency(CLOCK_AHB_CLK); + + ASSERT(config->bit_rate <= spi_clk / (SPI_CPSDVSR_MIN * (1 + SPI_SCR_MIN)), "SPI rate too large: %d", config->bit_rate); + ASSERT(config->bit_rate >= spi_clk / (SPI_CPSDVSR_MAX * (1 + SPI_SCR_MAX)), "SPI rate too small: %d", config->bit_rate); + ASSERT(config->data_bits <= SPI_DATA_BITS_MAX && config->data_bits >= SPI_DATA_BITS_MIN, "Invalid SPI data bits: %d", config->data_bits); + + // reset device handle + spi_handle[id].tx_buff = NULL; + spi_handle[id].tx_callback = NULL; + spi_handle[id].tx_count = 0; + spi_handle[id].tx_size = 0; + spi_handle[id].rx_buff = NULL; + spi_handle[id].rx_callback = NULL; + spi_handle[id].rx_count = 0; + spi_handle[id].rx_size = 0; + spi_handle[id].slave = config->slave; + spi_handle[id].dma_rx = config->dma_rx; + spi_handle[id].dma_tx = config->dma_tx; + spi_handle[id].int_rx = config->int_rx; + spi_handle[id].int_tx = config->int_tx; + + // caculate divisor + val = (spi_clk + config->bit_rate - 1) / config->bit_rate; + cpsdvsr = (val + SPI_SCR_MAX) / (SPI_SCR_MAX + 1); + if (cpsdvsr < SPI_CPSDVSR_MIN) + { + cpsdvsr = SPI_CPSDVSR_MIN; + } + else + { + if (cpsdvsr & 0x1) + { + cpsdvsr += 1; + } + if (cpsdvsr > SPI_CPSDVSR_MAX) + { + cpsdvsr = SPI_CPSDVSR_MAX; + } + } + scr = (val + cpsdvsr - 1) / cpsdvsr; + if (scr > SPI_SCR_MIN) + { + scr -= 1; + } + if (scr > SPI_SCR_MAX) + { + scr = SPI_SCR_MAX; + } + + val = SPI_CTRL0_SCR(scr) | (config->clk_phase ? SPI_CTRL0_SPH_MSK : 0) | (config->clk_polarity ? SPI_CTRL0_SPO_MSK : 0) | + (config->ti_mode ? SPI_CTRL0_FRF(1) : 0) | SPI_CTRL0_DSS(config->data_bits - 1); + spi_handle[id].base->CTRL0 = val; + + // enable SPI + val = (config->slave ? SPI_CTRL1_MS_MSK : 0) | SPI_CTRL1_SSE_MSK; + spi_handle[id].base->CTRL1 = val; + + spi_handle[id].base->PRESCALER = SPI_PRESCALER(cpsdvsr); + +#if SPI_INT_MODE_EN + if (spi_handle[id].int_rx || spi_handle[id].int_tx) + { + NVIC_SetPriority(spi_handle[id].irq, IRQ_PRIORITY_NORMAL); + NVIC_ClearPendingIRQ(spi_handle[id].irq); + NVIC_EnableIRQ(spi_handle[id].irq); + } +#endif + + spi_handle[id].state = SPI_STATE_READY; + return DRV_OK; +} + +int spi_close(enum SPI_DEV_T id) +{ + if (id >= SPI_MAX_NUM) + { + return DRV_ERROR; + } + +#if SPI_INT_MODE_EN + if (spi_handle[id].int_rx || spi_handle[id].int_tx) + { + NVIC_DisableIRQ(spi_handle[id].irq); + NVIC_ClearPendingIRQ(spi_handle[id].irq); + } +#endif + + // disable SPI + spi_handle[id].base->CTRL1 &= ~SPI_CTRL1_SSE_MSK; + + if (id == SPI_ID0) + { + // disable SPI0 clock + clock_disable(CLOCK_SPI0); + } + else if (id == SPI_ID1) + { + // disable SPI1 clock + clock_disable(CLOCK_SPI1); + } + + spi_handle[id].state = SPI_STATE_RESET; + return DRV_OK; +} + +#if defined(__GNUC__) +#pragma GCC diagnostic push +#pragma GCC diagnostic ignored "-Wcast-qual" +#endif + +/* + * TX + */ +int spi_send(enum SPI_DEV_T id, uint8_t *tx_buf, uint32_t len, drv_callback_t callback) +{ + uint8_t frame_size = spi_get_frame_size(id); + if ((tx_buf == 0) || (len == 0) || (len % frame_size)) + { + LOG_INFO(TRACE_MODULE_DRIVER, "Invalid SPI send parameters\r\n"); + return DRV_ERROR; + } + + int ret = spi_state_set(id, SPI_STATE_BUSY_TX); + if (ret != DRV_OK) + { + return ret; + } + + spi_handle[id].tx_buff = tx_buf; + spi_handle[id].tx_count = 0; + spi_handle[id].tx_size = len; + spi_handle[id].tx_callback = callback; + + uint32_t frame_data = 0U; + + if (spi_handle[id].dma_tx) + { +#if SPI_DMA_MODE_EN + struct DMA_CH_CFG_T spi_tx_dma_cfg = { + .fifo_th = (frame_size == 2 ? DMA_FIFO_TH_2 : DMA_FIFO_TH_1), + .src_burst_size = DMA_SRC_BURST_SIZE_1, + .src_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .dst_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .src_addr_ctrl = DMA_ADDR_INC, + .dst_addr_ctrl = DMA_ADDR_FIXED, + .src_req_sel = DMA_REQ_MEM, + .dst_req_sel = (id == SPI_ID1 ? DMA_REQ_SPI1_TX : DMA_REQ_SPI0_TX), + }; + + spi_handle[id].base->DMA_EN = SPI_DMA_EN_TX_MSK; + dma_open(spi_handle[id].dma_tx_ch, &spi_tx_dma_cfg); + dma_transfer(spi_handle[id].dma_tx_ch, tx_buf, (uint8_t *)&spi_handle[id].base->DATA, len / frame_size, spi_dma_callback); +#endif + } + else if (spi_handle[id].int_tx) + { +#if SPI_INT_MODE_EN + // write data to FIFO + while ((spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK) && (spi_handle[id].tx_count < spi_handle[id].tx_size)) + { + frame_data = 0; + memcpy(&frame_data, &spi_handle[id].tx_buff[spi_handle[id].tx_count], frame_size); + spi_handle[id].base->DATA = frame_data; + spi_handle[id].tx_count += frame_size; + } + // enable interrupts + spi_handle[id].base->INTR_EN |= SPI_INTR_EN_TX_MSK; +#endif + } + else + { +#if SPI_POLL_MODE_EN + // polling + while (spi_handle[id].tx_count < spi_handle[id].tx_size) + { + if ((spi_handle[id].base->CTRL1 & SPI_CTRL1_SSE_MSK) == 0) + { + ret = DRV_ERROR; + break; + } + + if (spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK) + { + frame_data = 0; + memcpy(&frame_data, &spi_handle[id].tx_buff[spi_handle[id].tx_count], frame_size); + spi_handle[id].base->DATA = frame_data; + spi_handle[id].tx_count += frame_size; + } + } + + while (spi_is_busy(id)) + { + } + + spi_state_clear(id, SPI_STATE_BUSY_TX); + + if (spi_handle[id].tx_callback) + { + spi_handle[id].tx_callback(&id, 0); + } +#endif + } + + return ret; +} + +/* + * RX + * Master RX DMA mode need to enable TX DMA to send dummy data + */ +int spi_receive(enum SPI_DEV_T id, uint8_t *rx_buf, uint32_t len, drv_callback_t callback) +{ + uint8_t frame_size = spi_get_frame_size(id); + if ((rx_buf == 0) || (len == 0) || (len % frame_size)) + { + LOG_INFO(TRACE_MODULE_DRIVER, "Invalid SPI recv parameters\r\n"); + return DRV_ERROR; + } + + int ret = spi_state_set(id, SPI_STATE_BUSY_RX); + if (ret != DRV_OK) + { + return ret; + } + + spi_handle[id].rx_buff = rx_buf; + spi_handle[id].rx_count = 0; + spi_handle[id].rx_size = len; + spi_handle[id].rx_callback = callback; + + uint32_t frame_data = 0U; + uint32_t dummy_cnt = 0; + + // clear FIFO + while (spi_handle[id].base->STATUS0 & SPI_STATUS0_RNE_MSK) + { + spi_handle[id].base->DATA; + } + spi_handle[id].base->INTR_CLR = ~0UL; + + if (spi_handle[id].dma_rx) + { +#if SPI_DMA_MODE_EN + struct DMA_CH_CFG_T spi_rx_dma_cfg = { + .fifo_th = (frame_size == 2 ? DMA_FIFO_TH_2 : DMA_FIFO_TH_1), + .src_burst_size = DMA_SRC_BURST_SIZE_1, + .src_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .dst_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .src_addr_ctrl = DMA_ADDR_FIXED, + .dst_addr_ctrl = DMA_ADDR_INC, + .src_req_sel = (id == SPI_ID1 ? DMA_REQ_SPI1_RX : DMA_REQ_SPI0_RX), + .dst_req_sel = DMA_REQ_MEM, + }; + + spi_handle[id].base->DMA_EN = (spi_handle[id].slave == 0) ? (SPI_DMA_EN_RX_MSK | SPI_DMA_EN_TX_MSK) : SPI_DMA_EN_RX_MSK; + dma_open(spi_handle[id].dma_rx_ch, &spi_rx_dma_cfg); + dma_transfer(spi_handle[id].dma_rx_ch, (uint8_t *)&spi_handle[id].base->DATA, rx_buf, len / frame_size, spi_dma_callback); + + if (spi_handle[id].slave == false) + { + struct DMA_CH_CFG_T spi_tx_dma_cfg = { + .fifo_th = (frame_size == 2 ? DMA_FIFO_TH_2 : DMA_FIFO_TH_1), + .src_burst_size = DMA_SRC_BURST_SIZE_1, + .src_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .dst_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .src_addr_ctrl = DMA_ADDR_FIXED, + .dst_addr_ctrl = DMA_ADDR_FIXED, + .src_req_sel = DMA_REQ_MEM, + .dst_req_sel = (id == SPI_ID1 ? DMA_REQ_SPI1_TX : DMA_REQ_SPI0_TX), + }; + + dma_open(spi_handle[id].dma_tx_ch, &spi_tx_dma_cfg); + dma_transfer(spi_handle[id].dma_tx_ch, &spi_handle[id].tx_dummy, (uint8_t *)&spi_handle[id].base->DATA, len / frame_size, NULL); + } +#endif + } + else if (spi_handle[id].int_rx) + { +#if SPI_INT_MODE_EN + // enable interrupts + spi_handle[id].base->INTR_EN |= SPI_INTR_EN_RX_MSK | SPI_INTR_EN_RT_MSK | SPI_INTR_EN_ROR_MSK; + + if ((spi_handle[id].slave == 0) && (((spi_handle[id].state & SPI_STATE_BUSY_TX) == 0) && (spi_handle[id].base->STATUS0 & SPI_STATUS0_TFE_MSK))) + { + // send dummy data as many as possible + dummy_cnt = spi_handle[id].rx_size - spi_handle[id].rx_count; + while ((dummy_cnt >= frame_size) && (spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK)) + { + spi_handle[id].base->DATA = spi_handle[id].tx_dummy; + dummy_cnt -= frame_size; + } + } +#endif + } + else + { +#if SPI_POLL_MODE_EN + // polling + while (spi_handle[id].rx_count < spi_handle[id].rx_size) + { + if ((spi_handle[id].base->CTRL1 & SPI_CTRL1_SSE_MSK) == 0) + { + ret = DRV_ERROR; + break; + } + + if (spi_handle[id].base->STATUS0 & SPI_STATUS0_RNE_MSK) + { + frame_data = spi_handle[id].base->DATA; + memcpy(&spi_handle[id].rx_buff[spi_handle[id].rx_count], &frame_data, frame_size); + spi_handle[id].rx_count += frame_size; + } + + if ((spi_handle[id].slave == 0) && (((spi_handle[id].state & SPI_STATE_BUSY_TX) == 0) && (spi_handle[id].base->STATUS0 & SPI_STATUS0_TFE_MSK))) + { + if (dummy_cnt < spi_handle[id].rx_size) + { + // send dummy data + spi_handle[id].base->DATA = spi_handle[id].tx_dummy; + dummy_cnt++; + } + } + } + + // update state + spi_state_clear(id, SPI_STATE_BUSY_RX); + + if (spi_handle[id].rx_callback) + { + spi_handle[id].rx_callback(&id, 0); + } +#endif + } + + return ret; +} + +/* + * Full duplex + * limitation: int_tx == int_rx, dma_tx == dma_rx + */ +int spi_transfer(enum SPI_DEV_T id, uint8_t *tx_buf, uint8_t *rx_buf, uint32_t len, drv_callback_t callback) +{ + uint8_t frame_size = spi_get_frame_size(id); + if ((rx_buf == 0) || (len == 0) || (len % frame_size)) + { + LOG_INFO(TRACE_MODULE_DRIVER, "Invalid SPI transfer parameters\r\n"); + return DRV_ERROR; + } + if ((spi_handle[id].dma_rx != spi_handle[id].dma_tx) || (spi_handle[id].int_rx != spi_handle[id].int_tx)) + { + LOG_INFO(TRACE_MODULE_DRIVER, "Invalid SPI configuration\r\n"); + return DRV_ERROR; + } + + int ret = spi_state_set(id, SPI_STATE_BUSY_TX_RX); + if (ret != DRV_OK) + { + return ret; + } + + spi_handle[id].tx_buff = tx_buf; + spi_handle[id].tx_count = 0; + spi_handle[id].tx_size = len; + spi_handle[id].tx_callback = NULL; + + spi_handle[id].rx_buff = rx_buf; + spi_handle[id].rx_count = 0; + spi_handle[id].rx_size = len; + spi_handle[id].rx_callback = callback; + + // clear FIFO + while (spi_handle[id].base->STATUS0 & SPI_STATUS0_RNE_MSK) + { + spi_handle[id].base->DATA; + } + spi_handle[id].base->INTR_CLR = ~0UL; + + uint32_t frame_data = 0U; + + if (spi_handle[id].dma_tx) + { +#if SPI_DMA_MODE_EN + struct DMA_CH_CFG_T spi_rx_dma_cfg = { + .fifo_th = (frame_size == 2 ? DMA_FIFO_TH_2 : DMA_FIFO_TH_1), + .src_burst_size = DMA_SRC_BURST_SIZE_1, + .src_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .dst_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .src_addr_ctrl = DMA_ADDR_FIXED, + .dst_addr_ctrl = DMA_ADDR_INC, + .src_req_sel = (id == SPI_ID1 ? DMA_REQ_SPI1_RX : DMA_REQ_SPI0_RX), + .dst_req_sel = DMA_REQ_MEM, + }; + + spi_handle[id].base->DMA_EN = SPI_DMA_EN_TX_MSK | SPI_DMA_EN_RX_MSK; + dma_open(spi_handle[id].dma_rx_ch, &spi_rx_dma_cfg); + dma_transfer(spi_handle[id].dma_rx_ch, (uint8_t *)&spi_handle[id].base->DATA, rx_buf, len / frame_size, spi_dma_callback); + + struct DMA_CH_CFG_T spi_tx_dma_cfg = { + .fifo_th = (frame_size == 2 ? DMA_FIFO_TH_2 : DMA_FIFO_TH_1), + .src_burst_size = DMA_SRC_BURST_SIZE_1, + .src_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .dst_width = (frame_size == 2 ? DMA_WIDTH_2B : DMA_WIDTH_1B), + .src_addr_ctrl = DMA_ADDR_INC, + .dst_addr_ctrl = DMA_ADDR_FIXED, + .src_req_sel = DMA_REQ_MEM, + .dst_req_sel = (id == SPI_ID1 ? DMA_REQ_SPI1_TX : DMA_REQ_SPI0_TX), + }; + + dma_open(spi_handle[id].dma_tx_ch, &spi_tx_dma_cfg); + dma_transfer(spi_handle[id].dma_tx_ch, tx_buf, (uint8_t *)&spi_handle[id].base->DATA, len / frame_size, spi_dma_callback); +#endif + } + else if (spi_handle[id].int_tx) + { +#if SPI_INT_MODE_EN + // write data to FIFO + while ((spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK) && (spi_handle[id].tx_count < spi_handle[id].tx_size)) + { + frame_data = 0; + memcpy(&frame_data, &spi_handle[id].tx_buff[spi_handle[id].tx_count], frame_size); + spi_handle[id].base->DATA = frame_data; + spi_handle[id].tx_count += frame_size; + } + // enable interrupts + spi_handle[id].base->INTR_EN |= SPI_INTR_EN_TX_MSK | SPI_INTR_EN_RX_MSK | SPI_INTR_EN_RT_MSK | SPI_INTR_EN_ROR_MSK; +#endif + } + else + { +#if SPI_POLL_MODE_EN + // polling + while (spi_handle[id].rx_count < spi_handle[id].rx_size) + { + if ((spi_handle[id].base->CTRL1 & SPI_CTRL1_SSE_MSK) == 0) + { + ret = DRV_ERROR; + break; + } + + if ((spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK) && (spi_handle[id].tx_count < spi_handle[id].tx_size)) + { + frame_data = 0; + memcpy(&frame_data, &spi_handle[id].tx_buff[spi_handle[id].tx_count], frame_size); + spi_handle[id].base->DATA = frame_data; + spi_handle[id].tx_count += frame_size; + } + + if (spi_handle[id].base->STATUS0 & SPI_STATUS0_RNE_MSK) + { + frame_data = spi_handle[id].base->DATA; + memcpy(&spi_handle[id].rx_buff[spi_handle[id].rx_count], &frame_data, frame_size); + spi_handle[id].rx_count += frame_size; + } + } + + // update state + spi_state_clear(id, SPI_STATE_BUSY_TX_RX); + + if (spi_handle[id].rx_callback) + { + spi_handle[id].rx_callback(&id, 0); + } +#endif + } + + return ret; +} + +int spi_abort_dma(enum SPI_DEV_T id, uint32_t abort_direction, drv_callback_t abort_tx_callback, drv_callback_t abort_rx_callback) +{ +#if SPI_DMA_MODE_EN + /* + * disable the SPI DMA rx request if enabled + * if (REG_IS_BIT_SET(spi_handle[id].base->DMA_EN, SPI_DMA_EN_RX_MSK)) + * if (spi_handle[id].state & SPI_STATE_BUSY_RX) + */ + if (SPI_DMA_ABORT_RX & abort_direction) + { + spi_handle[id].rx_abort_callback = abort_rx_callback; + spi_handle[id].base->DMA_EN &= ~SPI_DMA_EN_RX_MSK; + // dma_abort(spi_handle[id].dma_rx_ch, spi_dma_abort_callback); + dma_force_abort(spi_handle[id].dma_rx_ch, spi_dma_abort_callback); + } + + /* + * disable the SPI DMA tx request if enabled + * if (REG_IS_BIT_SET(spi_handle[id].base->DMA_EN, SPI_DMA_EN_TX_MSK)) + * if (spi_handle[id].state & SPI_STATE_BUSY_TX) + */ + if (SPI_DMA_ABORT_TX & abort_direction) + { + spi_handle[id].tx_abort_callback = abort_tx_callback; + spi_handle[id].base->DMA_EN &= ~SPI_DMA_EN_TX_MSK; + // dma_abort(spi_handle[id].dma_tx_ch, spi_dma_abort_callback); + dma_force_abort(spi_handle[id].dma_tx_ch, spi_dma_abort_callback); + } +#endif + + return DRV_OK; +} + +#if defined(__GNUC__) +#pragma GCC diagnostic pop +#endif + +#if SPI_DMA_MODE_EN +static void spi_dma_abort_callback(void *ch, uint32_t err_code) +{ + uint8_t ch_num = *(uint8_t *)ch; + drv_callback_t usr_callback = NULL; + enum SPI_DEV_T id; + + if ((ch_num == spi_handle[SPI_ID0].dma_tx_ch) || (ch_num == spi_handle[SPI_ID1].dma_tx_ch)) + { + id = (ch_num == spi_handle[SPI_ID0].dma_tx_ch ? SPI_ID0 : SPI_ID1); + + // TX dma abort + usr_callback = spi_handle[id].tx_abort_callback; + + // update state + spi_handle[id].state = SPI_STATE_READY; + + spi_handle[id].tx_buff = NULL; + spi_handle[id].tx_abort_callback = NULL; + spi_handle[id].tx_callback = NULL; + spi_handle[id].tx_count = 0; + spi_handle[id].tx_size = 0; + } + else if ((ch_num == spi_handle[SPI_ID0].dma_rx_ch) || (ch_num == spi_handle[SPI_ID1].dma_rx_ch)) + { + id = (ch_num == spi_handle[SPI_ID0].dma_rx_ch ? SPI_ID0 : SPI_ID1); + + // RX dma abort + usr_callback = spi_handle[id].rx_abort_callback; + + spi_handle[id].state = SPI_STATE_READY; + + spi_handle[id].rx_buff = NULL; + spi_handle[id].rx_abort_callback = NULL; + spi_handle[id].rx_callback = NULL; + spi_handle[id].rx_count = 0; + spi_handle[id].rx_size = 0; + } + else + { + ASSERT(0, "Unexpected dma channel\r\n"); + } + + if (usr_callback) + { + usr_callback(&id, err_code); + } +} + +static void spi_dma_callback(void *ch, uint32_t err_code) +{ + uint8_t ch_num = *(uint8_t *)ch; + drv_callback_t usr_callback = NULL; + enum SPI_DEV_T id; + + if ((ch_num == spi_handle[SPI_ID0].dma_tx_ch) || (ch_num == spi_handle[SPI_ID1].dma_tx_ch)) + { + id = (ch_num == spi_handle[SPI_ID0].dma_tx_ch ? SPI_ID0 : SPI_ID1); + + if (err_code == DMA_INT_TYPE_DONE) + { + spi_handle[id].base->DMA_EN &= ~SPI_DMA_EN_TX_MSK; + + // TX done + usr_callback = spi_handle[id].tx_callback; + + // update state + spi_state_clear(id, SPI_STATE_BUSY_TX); + } + else + { + // update state + if ((spi_handle[id].state == SPI_STATE_BUSY_TX_RX) || (spi_handle[id].state == SPI_STATE_BUSY_TX)) + { + spi_handle[id].state = SPI_STATE_ERROR; + } + else + { + ASSERT(0, "Unexpected spi state\r\n"); + } + } + spi_handle[id].tx_buff = NULL; + spi_handle[id].tx_callback = NULL; + spi_handle[id].tx_count = 0; + spi_handle[id].tx_size = 0; + } + else if ((ch_num == spi_handle[SPI_ID0].dma_rx_ch) || (ch_num == spi_handle[SPI_ID1].dma_rx_ch)) + { + id = (ch_num == spi_handle[SPI_ID0].dma_rx_ch ? SPI_ID0 : SPI_ID1); + + if (err_code == DMA_INT_TYPE_DONE) + { + spi_handle[id].base->DMA_EN &= ~SPI_DMA_EN_RX_MSK; + // RX done + usr_callback = spi_handle[id].rx_callback; + + // update state + spi_state_clear(id, SPI_STATE_BUSY_RX); + } + else + { + // update state + if ((spi_handle[id].state == SPI_STATE_BUSY_TX_RX) || (spi_handle[id].state == SPI_STATE_BUSY_RX)) + { + spi_handle[id].state = SPI_STATE_ERROR; + } + else + { + ASSERT(0, "Unexpected spi state\r\n"); + } + } + spi_handle[id].rx_buff = NULL; + spi_handle[id].rx_callback = NULL; + spi_handle[id].rx_count = 0; + spi_handle[id].rx_size = 0; + } + else + { + ASSERT(0, "Unexpected dma channel\r\n"); + } + + if (usr_callback) + { + usr_callback(&id, err_code); + } +} +#endif + +#if SPI_INT_MODE_EN +static void spi_irq_handler(enum SPI_DEV_T id) +{ + drv_callback_t usr_callback = NULL; + // what caused the interrupt? + uint32_t int_stat = spi_handle[id].base->INTR_STATUS; + + uint32_t frame_data = 0U; + uint8_t frame_size = spi_get_frame_size(id); + + if (int_stat & (SPI_INTR_STATUS_RX_MSK | SPI_INTR_STATUS_ROR_MSK | SPI_INTR_STATUS_RT_MSK)) + { + if (int_stat & SPI_INTR_STATUS_ROR_MSK) + { + spi_handle[id].base->INTR_CLR = SPI_INTR_CLR_ROR_MSK; + LOG_INFO(TRACE_MODULE_DRIVER, "SPI receive overrun\r\n"); + } + else if (int_stat & SPI_INTR_STATUS_RT_MSK) + { + // rx timeout interrupt happens to fetch less than 4 bytes data in FIFO + spi_handle[id].base->INTR_CLR = SPI_INTR_CLR_RT_MSK; + } + + // received data - read FIFO + if (spi_handle[id].state & SPI_STATE_BUSY_RX) + { + while ((spi_handle[id].base->STATUS0 & SPI_STATUS0_RNE_MSK) && (spi_handle[id].rx_count < spi_handle[id].rx_size)) + { + frame_data = spi_handle[id].base->DATA; + memcpy(&spi_handle[id].rx_buff[spi_handle[id].rx_count], &frame_data, frame_size); + spi_handle[id].rx_count += frame_size; + } + + if (spi_handle[id].rx_count == spi_handle[id].rx_size) + { + // RX done - disable interrupts + spi_handle[id].base->INTR_EN &= ~(SPI_INTR_EN_RX_MSK | SPI_INTR_EN_RT_MSK | SPI_INTR_EN_ROR_MSK); + usr_callback = spi_handle[id].rx_callback; + + // update state + spi_state_clear(id, SPI_STATE_BUSY_RX); + + spi_handle[id].rx_buff = NULL; + spi_handle[id].rx_callback = NULL; + spi_handle[id].rx_count = 0; + spi_handle[id].rx_size = 0; + } + else if ((spi_handle[id].slave == 0) && (((spi_handle[id].state & SPI_STATE_BUSY_TX) == 0) && (spi_handle[id].base->STATUS0 & SPI_STATUS0_TFE_MSK))) + { + // send dummy data as many as possible + uint32_t dummy_cnt = spi_handle[id].rx_size - spi_handle[id].rx_count; + while ((dummy_cnt >= frame_size) && (spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK)) + { + spi_handle[id].base->DATA = spi_handle[id].tx_dummy; + dummy_cnt -= frame_size; + } + } + } + } + else if (int_stat & SPI_INTR_STATUS_TX_MSK) + { + if (spi_handle[id].state & SPI_STATE_BUSY_TX) + { + if (spi_handle[id].tx_count == spi_handle[id].tx_size) + { + // TX done - disable interrupt + spi_handle[id].base->INTR_EN &= ~SPI_INTR_EN_TX_MSK; + usr_callback = spi_handle[id].tx_callback; + + // update state + spi_state_clear(id, SPI_STATE_BUSY_TX); + + spi_handle[id].tx_buff = NULL; + spi_handle[id].tx_callback = NULL; + spi_handle[id].tx_count = 0; + spi_handle[id].tx_size = 0; + } + else + { + // TX continue - write FIFO + while ((spi_handle[id].base->STATUS0 & SPI_STATUS0_TNF_MSK) && (spi_handle[id].tx_count < spi_handle[id].tx_size)) + { + frame_data = 0; + memcpy(&frame_data, &spi_handle[id].tx_buff[spi_handle[id].tx_count], frame_size); + spi_handle[id].base->DATA = frame_data; + spi_handle[id].tx_count += frame_size; + } + } + } + } + else + { + // it will reach here since FIFO was read empty by last interrupt + } + + if (usr_callback) + { + usr_callback(&id, int_stat); + } +} +#endif + +void SPI0_IRQHandler(void) +{ +#if SPI_INT_MODE_EN + spi_irq_handler(SPI_ID0); +#endif +} + +void SPI1_IRQHandler(void) +{ +#if SPI_INT_MODE_EN + spi_irq_handler(SPI_ID1); +#endif +} -- Gitblit v1.9.3