[rtems-libbsd commit] at91_mci: Get rid of bounce buffer
Sebastian Huber
sebh at rtems.org
Fri Apr 27 11:30:36 UTC 2018
Module: rtems-libbsd
Branch: master
Commit: d45899bce8b0e0b813295b17c1d0da7312d7a1fc
Changeset: http://git.rtems.org/rtems-libbsd/commit/?id=d45899bce8b0e0b813295b17c1d0da7312d7a1fc
Author: Sebastian Huber <sebastian.huber at embedded-brains.de>
Date: Fri Apr 27 08:39:22 2018 +0200
at91_mci: Get rid of bounce buffer
---
freebsd/sys/arm/at91/at91_mci.c | 120 +++++++++++++++++++---------------------
1 file changed, 57 insertions(+), 63 deletions(-)
diff --git a/freebsd/sys/arm/at91/at91_mci.c b/freebsd/sys/arm/at91/at91_mci.c
index 82041da..557efa8 100644
--- a/freebsd/sys/arm/at91/at91_mci.c
+++ b/freebsd/sys/arm/at91/at91_mci.c
@@ -140,15 +140,14 @@ static sXdmad *pXdmad = &XDMAD_Instance;
* than 16K gets turned into a dcache_wbinv_all(). That needlessly flushes the
* entire data cache, impacting overall system performance.
*/
-#define BBCOUNT 2
#ifndef __rtems__
-#define BBSIZE (16*1024)
-#define MAX_BLOCKS ((BBSIZE*BBCOUNT)/512)
-#else /* __rtems__ */
+#define BBCOUNT 2
#define BBSIZE (32*1024)
#define MAX_BLOCKS ((BBSIZE)/512)
/* FIXME: It would be better to split the DMA up in that case like in the
* original driver. But that would need some rework. */
+#else /* __rtems__ */
+#define MAX_BLOCKS 32
#endif /* __rtems__ */
#ifndef __rtems__
@@ -177,17 +176,20 @@ struct at91_mci_softc {
#ifdef __rtems__
RTEMS_INTERRUPT_LOCK_MEMBER(sc_lock)
#endif /* __rtems__ */
+#ifndef __rtems__
bus_dma_tag_t dmatag;
+#endif /* __rtems__ */
struct mmc_host host;
int bus_busy;
struct mmc_request *req;
struct mmc_command *curcmd;
+#ifndef __rtems__
bus_dmamap_t bbuf_map[BBCOUNT];
char * bbuf_vaddr[BBCOUNT]; /* bounce bufs in KVA space */
uint32_t bbuf_len[BBCOUNT]; /* len currently queued for bounce buf */
uint32_t bbuf_curidx; /* which bbuf is the active DMA buffer */
uint32_t xfer_offset; /* offset so far into caller's buf */
-#ifdef __rtems__
+#else /* __rtems__ */
uint32_t xdma_tx_channel;
uint32_t xdma_rx_channel;
uint8_t xdma_tx_perid;
@@ -207,10 +209,10 @@ static void at91_mci_intr(void *);
static int at91_mci_activate(device_t dev);
static void at91_mci_deactivate(device_t dev);
static int at91_mci_is_mci1rev2xx(void);
-#ifdef __rtems__
+#ifndef __rtems__
static void at91_mci_read_done(struct at91_mci_softc *sc, uint32_t sr);
-static void at91_mci_write_done(struct at91_mci_softc *sc, uint32_t sr);
#endif /* __rtems__ */
+static void at91_mci_write_done(struct at91_mci_softc *sc, uint32_t sr);
#ifndef __rtems__
#define AT91_MCI_LOCK(_sc) mtx_lock(&(_sc)->sc_mtx)
@@ -251,6 +253,7 @@ WR4(struct at91_mci_softc *sc, bus_size_t off, uint32_t val)
bus_write_4(sc->mem_res, off, val);
}
+#ifndef __rtems__
static void
at91_bswap_buf(struct at91_mci_softc *sc, void * dptr, void * sptr, uint32_t memsize)
{
@@ -295,6 +298,7 @@ at91_mci_getaddr(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
return;
*(bus_addr_t *)arg = segs[0].ds_addr;
}
+#endif /* __rtems__ */
static void
at91_mci_pdc_disable(struct at91_mci_softc *sc)
@@ -430,7 +434,11 @@ at91_mci_attach(device_t dev)
struct sysctl_ctx_list *sctx;
struct sysctl_oid *soid;
device_t child;
+#ifndef __rtems__
int err, i;
+#else /* __rtems__ */
+ int err;
+#endif /* __rtems__ */
#ifdef __rtems__
#ifdef LIBBSP_ARM_ATSAM_BSP_H
@@ -515,6 +523,7 @@ at91_mci_attach(device_t dev)
at91_mci_fini(dev);
at91_mci_init(dev);
+#ifndef __rtems__
/*
* Allocate DMA tags and maps and bounce buffers.
*
@@ -543,7 +552,6 @@ at91_mci_attach(device_t dev)
/*
* Activate the interrupt
*/
-#ifndef __rtems__
err = bus_setup_intr(dev, sc->irq_res, INTR_TYPE_MISC | INTR_MPSAFE,
NULL, at91_mci_intr, sc, &sc->intrhand);
#else /* __rtems__ */
@@ -604,6 +612,9 @@ at91_mci_attach(device_t dev)
sc->host.caps |= MMC_CAP_4_BIT_DATA;
child = device_add_child(dev, "mmc", 0);
+#ifdef __rtems__
+ (void)child;
+#endif /* __rtems__ */
device_set_ivars(dev, &sc->host);
err = bus_generic_attach(dev);
out:
@@ -615,14 +626,18 @@ out:
static int
at91_mci_detach(device_t dev)
{
+#ifndef __rtems__
struct at91_mci_softc *sc = device_get_softc(dev);
+#endif /* __rtems__ */
at91_mci_fini(dev);
at91_mci_deactivate(dev);
+#ifndef __rtems__
bus_dmamem_free(sc->dmatag, sc->bbuf_vaddr[0], sc->bbuf_map[0]);
bus_dmamem_free(sc->dmatag, sc->bbuf_vaddr[1], sc->bbuf_map[1]);
bus_dma_tag_destroy(sc->dmatag);
+#endif /* __rtems__ */
return (EBUSY); /* XXX */
}
@@ -759,7 +774,7 @@ static LinkedListDescriporView1 dma_desc[MAX_BLOCKS];
static void
at91_mci_setup_xdma(struct at91_mci_softc *sc, bool read, uint32_t block_size,
- uint32_t number_blocks, bus_addr_t paddr, uint32_t len)
+ uint32_t block_count, void *data, uint32_t len)
{
sXdmadCfg *xdma_cfg;
uint32_t xdma_channel;
@@ -782,17 +797,17 @@ at91_mci_setup_xdma(struct at91_mci_softc *sc, bool read, uint32_t block_size,
xdma_channel = sc->xdma_tx_channel;
}
- for (i = 0; i < number_blocks; ++i) {
+ for (i = 0; i < block_count; ++i) {
if (read) {
dma_desc[i].mbr_sa = sa_rdr;
- dma_desc[i].mbr_da = ((uint32_t)paddr) + i * block_size;
+ dma_desc[i].mbr_da = ((uint32_t)data) + i * block_size;
} else {
- dma_desc[i].mbr_sa = ((uint32_t)paddr) + i * block_size;
+ dma_desc[i].mbr_sa = ((uint32_t)data) + i * block_size;
dma_desc[i].mbr_da = da_tdr;
}
dma_desc[i].mbr_ubc = XDMA_UBC_NVIEW_NDV1 |
XDMA_UBC_NDEN_UPDATED | (block_size/4);
- if (i == number_blocks - 1) {
+ if (i == block_count - 1) {
dma_desc[i].mbr_ubc |= XDMA_UBC_NDE_FETCH_DIS;
dma_desc[i].mbr_nda = 0;
} else {
@@ -808,9 +823,9 @@ at91_mci_setup_xdma(struct at91_mci_softc *sc, bool read, uint32_t block_size,
/* FIXME: Is that correct? */
if (read) {
- rtems_cache_invalidate_multiple_data_lines(paddr, len);
+ rtems_cache_invalidate_multiple_data_lines(data, len);
} else {
- rtems_cache_flush_multiple_data_lines(paddr, len);
+ rtems_cache_flush_multiple_data_lines(data, len);
}
rtems_cache_flush_multiple_data_lines(dma_desc, sizeof(dma_desc));
@@ -819,17 +834,16 @@ at91_mci_setup_xdma(struct at91_mci_softc *sc, bool read, uint32_t block_size,
panic("Could not start XDMA: %d.", rc);
}
-
#endif /* __rtems__ */
static void
at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
{
uint32_t cmdr, mr;
+ struct mmc_data *data;
#ifdef __rtems__
- uint32_t number_blocks;
+ uint32_t block_count;
uint32_t block_size;
#endif /* __rtems__ */
- struct mmc_data *data;
sc->curcmd = cmd;
data = cmd->data;
@@ -927,16 +941,6 @@ at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
mr = RD4(sc,MCI_MR) & ~MCI_MR_BLKLEN;
mr |= min(data->len, 512) << 16;
WR4(sc, MCI_MR, mr | MCI_MR_PDCMODE|MCI_MR_PDCPADV);
-#else /* __rtems__ */
- mr = RD4(sc,MCI_MR);
- WR4(sc, MCI_MR, mr | MCI_MR_PDCPADV);
-
- WR4(sc, MCI_DMA, MCI_DMA_DMAEN | MCI_DMA_CHKSIZE_1);
-
- block_size = min(data->len, 512);
- number_blocks = data->len / block_size;
- WR4(sc, MCI_BLKR, block_size << 16 | number_blocks);
-#endif /* __rtems__ */
/*
* Set up DMA.
@@ -970,29 +974,33 @@ at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
*/
sc->xfer_offset = 0;
sc->bbuf_curidx = 0;
+#else /* __rtems__ */
+ mr = RD4(sc,MCI_MR);
+ WR4(sc, MCI_MR, mr | MCI_MR_PDCPADV);
+
+ WR4(sc, MCI_DMA, MCI_DMA_DMAEN | MCI_DMA_CHKSIZE_1);
+
+ block_size = min(data->len, 512);
+ block_count = data->len / block_size;
+ WR4(sc, MCI_BLKR, (block_size << 16) | block_count);
+#endif /* __rtems__ */
if (data->flags & (MMC_DATA_READ | MMC_DATA_WRITE)) {
+#ifndef __rtems__
uint32_t len;
uint32_t remaining = data->len;
bus_addr_t paddr;
int err;
-#ifndef __rtems__
if (remaining > (BBCOUNT*BBSIZE))
-#else /* __rtems__ */
- if (remaining > (BBSIZE))
-#endif /* __rtems__ */
panic("IO read size exceeds MAXDATA\n");
+#endif /* __rtems__ */
if (data->flags & MMC_DATA_READ) {
#ifndef __rtems__
if (remaining > 2048) // XXX
len = remaining / 2;
else
-#else
- /* FIXME: This reduces performance. Set up DMA in two
- * parts instead like done on AT91. */
-#endif /* __rtems__ */
len = remaining;
err = bus_dmamap_load(sc->dmatag, sc->bbuf_map[0],
sc->bbuf_vaddr[0], len, at91_mci_getaddr,
@@ -1001,7 +1009,6 @@ at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
panic("IO read dmamap_load failed\n");
bus_dmamap_sync(sc->dmatag, sc->bbuf_map[0],
BUS_DMASYNC_PREREAD);
-#ifndef __rtems__
WR4(sc, PDC_RPR, paddr);
WR4(sc, PDC_RCR, len / 4);
sc->bbuf_len[0] = len;
@@ -1024,16 +1031,11 @@ at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
}
WR4(sc, PDC_PTCR, PDC_PTCR_RXTEN);
#else /* __rtems__ */
- at91_mci_setup_xdma(sc, true, block_size,
- number_blocks, paddr, len);
-
- sc->bbuf_len[0] = len;
- remaining -= len;
- sc->bbuf_len[1] = 0;
- if (remaining != 0)
- panic("Still rx-data left. This should never happen.");
+ at91_mci_setup_xdma(sc, true, block_size, block_count,
+ data->data, data->len);
#endif /* __rtems__ */
} else {
+#ifndef __rtems__
len = min(BBSIZE, remaining);
at91_bswap_buf(sc, sc->bbuf_vaddr[0], data->data, len);
err = bus_dmamap_load(sc->dmatag, sc->bbuf_map[0],
@@ -1043,7 +1045,6 @@ at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
panic("IO write dmamap_load failed\n");
bus_dmamap_sync(sc->dmatag, sc->bbuf_map[0],
BUS_DMASYNC_PREWRITE);
-#ifndef __rtems__
/*
* Erratum workaround: PDC transfer length on a write
* must not be smaller than 12 bytes (3 words); only
@@ -1073,15 +1074,8 @@ at91_mci_start_cmd(struct at91_mci_softc *sc, struct mmc_command *cmd)
}
/* do not enable PDC xfer until CMDRDY asserted */
#else /* __rtems__ */
- at91_mci_setup_xdma(sc, false, block_size,
- number_blocks, paddr, len);
-
- sc->bbuf_len[0] = len;
- remaining -= len;
- sc->bbuf_len[1] = 0;
- if (remaining != 0)
- panic("Still tx-data left. This should never happen.");
-
+ at91_mci_setup_xdma(sc, false, block_size, block_count,
+ data->data, data->len);
#endif /* __rtems__ */
}
data->xfer_len = 0; /* XXX what's this? appears to be unused. */
@@ -1190,6 +1184,7 @@ at91_mci_release_host(device_t brdev, device_t reqdev)
return (0);
}
+#ifndef __rtems__
static void
at91_mci_read_done(struct at91_mci_softc *sc, uint32_t sr)
{
@@ -1207,10 +1202,8 @@ at91_mci_read_done(struct at91_mci_softc *sc, uint32_t sr)
* operation, otherwise we wait for another ENDRX for the next bufer.
*/
-#ifndef __rtems__
bus_dmamap_sync(sc->dmatag, sc->bbuf_map[curidx], BUS_DMASYNC_POSTREAD);
bus_dmamap_unload(sc->dmatag, sc->bbuf_map[curidx]);
-#endif /* __rtems__ */
at91_bswap_buf(sc, dataptr + sc->xfer_offset, sc->bbuf_vaddr[curidx], len);
@@ -1235,13 +1228,10 @@ at91_mci_read_done(struct at91_mci_softc *sc, uint32_t sr)
at91_mci_next_operation(sc);
} else {
WR4(sc, PDC_RNCR, 0);
-#ifndef __rtems__
WR4(sc, MCI_IER, MCI_SR_ERROR | MCI_SR_ENDRX);
-#else /* __rtems__ */
- WR4(sc, MCI_IER, MCI_SR_ERROR | MCI_SR_XFRDONE);
-#endif /* __rtems__ */
}
}
+#endif /* __rtems__ */
static void
at91_mci_write_done(struct at91_mci_softc *sc, uint32_t sr)
@@ -1259,9 +1249,11 @@ at91_mci_write_done(struct at91_mci_softc *sc, uint32_t sr)
WR4(sc, PDC_PTCR, PDC_PTCR_RXTDIS | PDC_PTCR_TXTDIS);
+#ifndef __rtems__
bus_dmamap_sync(sc->dmatag, sc->bbuf_map[sc->bbuf_curidx],
BUS_DMASYNC_POSTWRITE);
bus_dmamap_unload(sc->dmatag, sc->bbuf_map[sc->bbuf_curidx]);
+#endif /* __rtems__ */
if ((cmd->data->flags & MMC_DATA_MULTI) || (sr & MCI_SR_NOTBUSY)) {
cmd->error = MMC_ERR_NONE;
@@ -1532,9 +1524,11 @@ at91_mci_intr(void *arg)
}
#else /* __rtems__ */
if (isr & MCI_SR_XFRDONE) {
- struct mmc_command *cmd = sc->curcmd;
if (cmd->data->flags & MMC_DATA_READ) {
- at91_mci_read_done(sc, sr);
+ WR4(sc, PDC_PTCR, PDC_PTCR_RXTDIS |
+ PDC_PTCR_TXTDIS);
+ cmd->error = MMC_ERR_NONE;
+ at91_mci_next_operation(sc);
} else {
if (sr & MCI_SR_BLKE)
isr |= MCI_SR_BLKE;
More information about the vc
mailing list