Blame view
drivers/dma/altera-msgdma.c
24.9 KB
2874c5fd2 treewide: Replace... |
1 |
// SPDX-License-Identifier: GPL-2.0-or-later |
a85c6f1b2 dmaengine: Add dr... |
2 3 4 5 6 7 8 |
/* * DMA driver for Altera mSGDMA IP core * * Copyright (C) 2017 Stefan Roese <sr@denx.de> * * Based on drivers/dma/xilinx/zynqmp_dma.c, which is: * Copyright (C) 2016 Xilinx, Inc. All rights reserved. |
a85c6f1b2 dmaengine: Add dr... |
9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 |
*/ #include <linux/bitops.h> #include <linux/delay.h> #include <linux/dma-mapping.h> #include <linux/dmapool.h> #include <linux/init.h> #include <linux/interrupt.h> #include <linux/io.h> #include <linux/iopoll.h> #include <linux/module.h> #include <linux/platform_device.h> #include <linux/slab.h> #include "dmaengine.h" #define MSGDMA_MAX_TRANS_LEN U32_MAX #define MSGDMA_DESC_NUM 1024 /** * struct msgdma_extended_desc - implements an extended descriptor * @read_addr_lo: data buffer source address low bits * @write_addr_lo: data buffer destination address low bits * @len: the number of bytes to transfer per descriptor * @burst_seq_num: bit 31:24 write burst * bit 23:16 read burst * bit 15:00 sequence number * @stride: bit 31:16 write stride * bit 15:00 read stride * @read_addr_hi: data buffer source address high bits * @write_addr_hi: data buffer destination address high bits * @control: characteristics of the transfer */ struct msgdma_extended_desc { u32 read_addr_lo; u32 write_addr_lo; u32 len; u32 burst_seq_num; u32 stride; u32 read_addr_hi; u32 write_addr_hi; u32 control; }; /* mSGDMA descriptor control field bit definitions */ #define MSGDMA_DESC_CTL_SET_CH(x) ((x) & 0xff) #define MSGDMA_DESC_CTL_GEN_SOP BIT(8) #define MSGDMA_DESC_CTL_GEN_EOP BIT(9) #define MSGDMA_DESC_CTL_PARK_READS BIT(10) #define MSGDMA_DESC_CTL_PARK_WRITES BIT(11) #define MSGDMA_DESC_CTL_END_ON_EOP BIT(12) #define MSGDMA_DESC_CTL_END_ON_LEN BIT(13) #define MSGDMA_DESC_CTL_TR_COMP_IRQ BIT(14) #define MSGDMA_DESC_CTL_EARLY_IRQ BIT(15) #define MSGDMA_DESC_CTL_TR_ERR_IRQ GENMASK(23, 16) #define MSGDMA_DESC_CTL_EARLY_DONE BIT(24) /* * Writing "1" the "go" bit commits the entire descriptor into the * descriptor FIFO(s) */ #define MSGDMA_DESC_CTL_GO BIT(31) /* Tx buffer control flags */ #define MSGDMA_DESC_CTL_TX_FIRST (MSGDMA_DESC_CTL_GEN_SOP | \ MSGDMA_DESC_CTL_TR_ERR_IRQ | \ MSGDMA_DESC_CTL_GO) #define MSGDMA_DESC_CTL_TX_MIDDLE (MSGDMA_DESC_CTL_TR_ERR_IRQ | \ MSGDMA_DESC_CTL_GO) #define MSGDMA_DESC_CTL_TX_LAST (MSGDMA_DESC_CTL_GEN_EOP | \ MSGDMA_DESC_CTL_TR_COMP_IRQ | \ MSGDMA_DESC_CTL_TR_ERR_IRQ | \ MSGDMA_DESC_CTL_GO) #define MSGDMA_DESC_CTL_TX_SINGLE (MSGDMA_DESC_CTL_GEN_SOP | \ MSGDMA_DESC_CTL_GEN_EOP | \ MSGDMA_DESC_CTL_TR_COMP_IRQ | \ MSGDMA_DESC_CTL_TR_ERR_IRQ | \ MSGDMA_DESC_CTL_GO) #define MSGDMA_DESC_CTL_RX_SINGLE (MSGDMA_DESC_CTL_END_ON_EOP | \ MSGDMA_DESC_CTL_END_ON_LEN | \ MSGDMA_DESC_CTL_TR_COMP_IRQ | \ MSGDMA_DESC_CTL_EARLY_IRQ | \ MSGDMA_DESC_CTL_TR_ERR_IRQ | \ MSGDMA_DESC_CTL_GO) /* mSGDMA extended descriptor stride definitions */ #define MSGDMA_DESC_STRIDE_RD 0x00000001 #define MSGDMA_DESC_STRIDE_WR 0x00010000 #define MSGDMA_DESC_STRIDE_RW 0x00010001 |
6084fc2ec dmaengine: altera... |
102 103 104 105 106 107 108 109 |
/* mSGDMA dispatcher control and status register map */ #define MSGDMA_CSR_STATUS 0x00 /* Read / Clear */ #define MSGDMA_CSR_CONTROL 0x04 /* Read / Write */ #define MSGDMA_CSR_RW_FILL_LEVEL 0x08 /* 31:16 - write fill level */ /* 15:00 - read fill level */ #define MSGDMA_CSR_RESP_FILL_LEVEL 0x0c /* response FIFO fill level */ #define MSGDMA_CSR_RW_SEQ_NUM 0x10 /* 31:16 - write seq number */ /* 15:00 - read seq number */ |
a85c6f1b2 dmaengine: Add dr... |
110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 |
/* mSGDMA CSR status register bit definitions */ #define MSGDMA_CSR_STAT_BUSY BIT(0) #define MSGDMA_CSR_STAT_DESC_BUF_EMPTY BIT(1) #define MSGDMA_CSR_STAT_DESC_BUF_FULL BIT(2) #define MSGDMA_CSR_STAT_RESP_BUF_EMPTY BIT(3) #define MSGDMA_CSR_STAT_RESP_BUF_FULL BIT(4) #define MSGDMA_CSR_STAT_STOPPED BIT(5) #define MSGDMA_CSR_STAT_RESETTING BIT(6) #define MSGDMA_CSR_STAT_STOPPED_ON_ERR BIT(7) #define MSGDMA_CSR_STAT_STOPPED_ON_EARLY BIT(8) #define MSGDMA_CSR_STAT_IRQ BIT(9) #define MSGDMA_CSR_STAT_MASK GENMASK(9, 0) #define MSGDMA_CSR_STAT_MASK_WITHOUT_IRQ GENMASK(8, 0) #define DESC_EMPTY (MSGDMA_CSR_STAT_DESC_BUF_EMPTY | \ MSGDMA_CSR_STAT_RESP_BUF_EMPTY) /* mSGDMA CSR control register bit definitions */ #define MSGDMA_CSR_CTL_STOP BIT(0) #define MSGDMA_CSR_CTL_RESET BIT(1) #define MSGDMA_CSR_CTL_STOP_ON_ERR BIT(2) #define MSGDMA_CSR_CTL_STOP_ON_EARLY BIT(3) #define MSGDMA_CSR_CTL_GLOBAL_INTR BIT(4) #define MSGDMA_CSR_CTL_STOP_DESCS BIT(5) /* mSGDMA CSR fill level bits */ #define MSGDMA_CSR_WR_FILL_LEVEL_GET(v) (((v) & 0xffff0000) >> 16) #define MSGDMA_CSR_RD_FILL_LEVEL_GET(v) ((v) & 0x0000ffff) #define MSGDMA_CSR_RESP_FILL_LEVEL_GET(v) ((v) & 0x0000ffff) #define MSGDMA_CSR_SEQ_NUM_GET(v) (((v) & 0xffff0000) >> 16) /* mSGDMA response register map */ |
6084fc2ec dmaengine: altera... |
144 145 |
#define MSGDMA_RESP_BYTES_TRANSFERRED 0x00 #define MSGDMA_RESP_STATUS 0x04 |
a85c6f1b2 dmaengine: Add dr... |
146 147 148 149 150 151 152 153 154 |
/* mSGDMA response register bit definitions */ #define MSGDMA_RESP_EARLY_TERM BIT(8) #define MSGDMA_RESP_ERR_MASK 0xff /** * struct msgdma_sw_desc - implements a sw descriptor * @async_tx: support for the async_tx api * @hw_desc: assosiated HW descriptor |
f15f720b8 dmaengine: altera... |
155 156 |
* @node: node to move from the free list to the tx list * @tx_list: transmit list node |
a85c6f1b2 dmaengine: Add dr... |
157 158 159 160 161 162 163 |
*/ struct msgdma_sw_desc { struct dma_async_tx_descriptor async_tx; struct msgdma_extended_desc hw_desc; struct list_head node; struct list_head tx_list; }; |
f15f720b8 dmaengine: altera... |
164 |
/* |
a85c6f1b2 dmaengine: Add dr... |
165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 |
* struct msgdma_device - DMA device structure */ struct msgdma_device { spinlock_t lock; struct device *dev; struct tasklet_struct irq_tasklet; struct list_head pending_list; struct list_head free_list; struct list_head active_list; struct list_head done_list; u32 desc_free_cnt; bool idle; struct dma_device dmadev; struct dma_chan dmachan; dma_addr_t hw_desq; struct msgdma_sw_desc *sw_desq; unsigned int npendings; struct dma_slave_config slave_cfg; int irq; /* mSGDMA controller */ |
6084fc2ec dmaengine: altera... |
189 |
void __iomem *csr; |
a85c6f1b2 dmaengine: Add dr... |
190 191 |
/* mSGDMA descriptors */ |
6084fc2ec dmaengine: altera... |
192 |
void __iomem *desc; |
a85c6f1b2 dmaengine: Add dr... |
193 194 |
/* mSGDMA response */ |
6084fc2ec dmaengine: altera... |
195 |
void __iomem *resp; |
a85c6f1b2 dmaengine: Add dr... |
196 197 198 199 200 201 202 203 204 205 206 207 208 209 |
}; #define to_mdev(chan) container_of(chan, struct msgdma_device, dmachan) #define tx_to_desc(tx) container_of(tx, struct msgdma_sw_desc, async_tx) /** * msgdma_get_descriptor - Get the sw descriptor from the pool * @mdev: Pointer to the Altera mSGDMA device structure * * Return: The sw descriptor */ static struct msgdma_sw_desc *msgdma_get_descriptor(struct msgdma_device *mdev) { struct msgdma_sw_desc *desc; |
edf10919e dmaengine: altera... |
210 |
unsigned long flags; |
a85c6f1b2 dmaengine: Add dr... |
211 |
|
edf10919e dmaengine: altera... |
212 |
spin_lock_irqsave(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
213 214 |
desc = list_first_entry(&mdev->free_list, struct msgdma_sw_desc, node); list_del(&desc->node); |
edf10919e dmaengine: altera... |
215 |
spin_unlock_irqrestore(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 248 249 250 251 252 253 254 255 256 257 258 259 |
INIT_LIST_HEAD(&desc->tx_list); return desc; } /** * msgdma_free_descriptor - Issue pending transactions * @mdev: Pointer to the Altera mSGDMA device structure * @desc: Transaction descriptor pointer */ static void msgdma_free_descriptor(struct msgdma_device *mdev, struct msgdma_sw_desc *desc) { struct msgdma_sw_desc *child, *next; mdev->desc_free_cnt++; list_add_tail(&desc->node, &mdev->free_list); list_for_each_entry_safe(child, next, &desc->tx_list, node) { mdev->desc_free_cnt++; list_move_tail(&child->node, &mdev->free_list); } } /** * msgdma_free_desc_list - Free descriptors list * @mdev: Pointer to the Altera mSGDMA device structure * @list: List to parse and delete the descriptor */ static void msgdma_free_desc_list(struct msgdma_device *mdev, struct list_head *list) { struct msgdma_sw_desc *desc, *next; list_for_each_entry_safe(desc, next, list, node) msgdma_free_descriptor(mdev, desc); } /** * msgdma_desc_config - Configure the descriptor * @desc: Hw descriptor pointer * @dst: Destination buffer address * @src: Source buffer address * @len: Transfer length |
f15f720b8 dmaengine: altera... |
260 |
* @stride: Read/write stride value to set |
a85c6f1b2 dmaengine: Add dr... |
261 262 263 264 265 266 267 268 269 270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 291 292 293 294 295 296 297 298 299 300 301 302 303 304 305 |
*/ static void msgdma_desc_config(struct msgdma_extended_desc *desc, dma_addr_t dst, dma_addr_t src, size_t len, u32 stride) { /* Set lower 32bits of src & dst addresses in the descriptor */ desc->read_addr_lo = lower_32_bits(src); desc->write_addr_lo = lower_32_bits(dst); /* Set upper 32bits of src & dst addresses in the descriptor */ desc->read_addr_hi = upper_32_bits(src); desc->write_addr_hi = upper_32_bits(dst); desc->len = len; desc->stride = stride; desc->burst_seq_num = 0; /* 0 will result in max burst length */ /* * Don't set interrupt on xfer end yet, this will be done later * for the "last" descriptor */ desc->control = MSGDMA_DESC_CTL_TR_ERR_IRQ | MSGDMA_DESC_CTL_GO | MSGDMA_DESC_CTL_END_ON_LEN; } /** * msgdma_desc_config_eod - Mark the descriptor as end descriptor * @desc: Hw descriptor pointer */ static void msgdma_desc_config_eod(struct msgdma_extended_desc *desc) { desc->control |= MSGDMA_DESC_CTL_TR_COMP_IRQ; } /** * msgdma_tx_submit - Submit DMA transaction * @tx: Async transaction descriptor pointer * * Return: cookie value */ static dma_cookie_t msgdma_tx_submit(struct dma_async_tx_descriptor *tx) { struct msgdma_device *mdev = to_mdev(tx->chan); struct msgdma_sw_desc *new; dma_cookie_t cookie; |
edf10919e dmaengine: altera... |
306 |
unsigned long flags; |
a85c6f1b2 dmaengine: Add dr... |
307 308 |
new = tx_to_desc(tx); |
edf10919e dmaengine: altera... |
309 |
spin_lock_irqsave(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
310 311 312 |
cookie = dma_cookie_assign(tx); list_add_tail(&new->node, &mdev->pending_list); |
edf10919e dmaengine: altera... |
313 |
spin_unlock_irqrestore(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
314 315 316 317 318 319 320 321 322 323 324 325 326 327 328 329 330 331 332 333 334 335 336 |
return cookie; } /** * msgdma_prep_memcpy - prepare descriptors for memcpy transaction * @dchan: DMA channel * @dma_dst: Destination buffer address * @dma_src: Source buffer address * @len: Transfer length * @flags: transfer ack flags * * Return: Async transaction descriptor on success and NULL on failure */ static struct dma_async_tx_descriptor * msgdma_prep_memcpy(struct dma_chan *dchan, dma_addr_t dma_dst, dma_addr_t dma_src, size_t len, ulong flags) { struct msgdma_device *mdev = to_mdev(dchan); struct msgdma_sw_desc *new, *first = NULL; struct msgdma_extended_desc *desc; size_t copy; u32 desc_cnt; |
edf10919e dmaengine: altera... |
337 |
unsigned long irqflags; |
a85c6f1b2 dmaengine: Add dr... |
338 339 |
desc_cnt = DIV_ROUND_UP(len, MSGDMA_MAX_TRANS_LEN); |
edf10919e dmaengine: altera... |
340 |
spin_lock_irqsave(&mdev->lock, irqflags); |
a85c6f1b2 dmaengine: Add dr... |
341 |
if (desc_cnt > mdev->desc_free_cnt) { |
c5709d376 dmaengine: altera... |
342 |
spin_unlock_irqrestore(&mdev->lock, irqflags); |
a85c6f1b2 dmaengine: Add dr... |
343 344 345 346 347 |
dev_dbg(mdev->dev, "mdev %p descs are not available ", mdev); return NULL; } mdev->desc_free_cnt -= desc_cnt; |
edf10919e dmaengine: altera... |
348 |
spin_unlock_irqrestore(&mdev->lock, irqflags); |
a85c6f1b2 dmaengine: Add dr... |
349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 |
do { /* Allocate and populate the descriptor */ new = msgdma_get_descriptor(mdev); copy = min_t(size_t, len, MSGDMA_MAX_TRANS_LEN); desc = &new->hw_desc; msgdma_desc_config(desc, dma_dst, dma_src, copy, MSGDMA_DESC_STRIDE_RW); len -= copy; dma_src += copy; dma_dst += copy; if (!first) first = new; else list_add_tail(&new->node, &first->tx_list); } while (len); msgdma_desc_config_eod(desc); async_tx_ack(&first->async_tx); first->async_tx.flags = flags; return &first->async_tx; } /** |
a85c6f1b2 dmaengine: Add dr... |
375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 |
* msgdma_prep_slave_sg - prepare descriptors for a slave sg transaction * * @dchan: DMA channel * @sgl: Destination scatter list * @sg_len: Number of entries in destination scatter list * @dir: DMA transfer direction * @flags: transfer ack flags * @context: transfer context (unused) */ static struct dma_async_tx_descriptor * msgdma_prep_slave_sg(struct dma_chan *dchan, struct scatterlist *sgl, unsigned int sg_len, enum dma_transfer_direction dir, unsigned long flags, void *context) { struct msgdma_device *mdev = to_mdev(dchan); struct dma_slave_config *cfg = &mdev->slave_cfg; struct msgdma_sw_desc *new, *first = NULL; void *desc = NULL; size_t len, avail; dma_addr_t dma_dst, dma_src; u32 desc_cnt = 0, i; struct scatterlist *sg; u32 stride; |
edf10919e dmaengine: altera... |
399 |
unsigned long irqflags; |
a85c6f1b2 dmaengine: Add dr... |
400 401 402 |
for_each_sg(sgl, sg, sg_len, i) desc_cnt += DIV_ROUND_UP(sg_dma_len(sg), MSGDMA_MAX_TRANS_LEN); |
edf10919e dmaengine: altera... |
403 |
spin_lock_irqsave(&mdev->lock, irqflags); |
a85c6f1b2 dmaengine: Add dr... |
404 |
if (desc_cnt > mdev->desc_free_cnt) { |
c5709d376 dmaengine: altera... |
405 |
spin_unlock_irqrestore(&mdev->lock, irqflags); |
a85c6f1b2 dmaengine: Add dr... |
406 407 408 409 410 |
dev_dbg(mdev->dev, "mdev %p descs are not available ", mdev); return NULL; } mdev->desc_free_cnt -= desc_cnt; |
edf10919e dmaengine: altera... |
411 |
spin_unlock_irqrestore(&mdev->lock, irqflags); |
a85c6f1b2 dmaengine: Add dr... |
412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 434 435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 |
avail = sg_dma_len(sgl); /* Run until we are out of scatterlist entries */ while (true) { /* Allocate and populate the descriptor */ new = msgdma_get_descriptor(mdev); desc = &new->hw_desc; len = min_t(size_t, avail, MSGDMA_MAX_TRANS_LEN); if (dir == DMA_MEM_TO_DEV) { dma_src = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; dma_dst = cfg->dst_addr; stride = MSGDMA_DESC_STRIDE_RD; } else { dma_src = cfg->src_addr; dma_dst = sg_dma_address(sgl) + sg_dma_len(sgl) - avail; stride = MSGDMA_DESC_STRIDE_WR; } msgdma_desc_config(desc, dma_dst, dma_src, len, stride); avail -= len; if (!first) first = new; else list_add_tail(&new->node, &first->tx_list); /* Fetch the next scatterlist entry */ if (avail == 0) { if (sg_len == 0) break; sgl = sg_next(sgl); if (sgl == NULL) break; sg_len--; avail = sg_dma_len(sgl); } } msgdma_desc_config_eod(desc); first->async_tx.flags = flags; return &first->async_tx; } static int msgdma_dma_config(struct dma_chan *dchan, struct dma_slave_config *config) { struct msgdma_device *mdev = to_mdev(dchan); memcpy(&mdev->slave_cfg, config, sizeof(*config)); return 0; } static void msgdma_reset(struct msgdma_device *mdev) { u32 val; int ret; /* Reset mSGDMA */ |
6084fc2ec dmaengine: altera... |
474 475 |
iowrite32(MSGDMA_CSR_STAT_MASK, mdev->csr + MSGDMA_CSR_STATUS); iowrite32(MSGDMA_CSR_CTL_RESET, mdev->csr + MSGDMA_CSR_CONTROL); |
a85c6f1b2 dmaengine: Add dr... |
476 |
|
6084fc2ec dmaengine: altera... |
477 |
ret = readl_poll_timeout(mdev->csr + MSGDMA_CSR_STATUS, val, |
a85c6f1b2 dmaengine: Add dr... |
478 479 480 481 482 483 484 |
(val & MSGDMA_CSR_STAT_RESETTING) == 0, 1, 10000); if (ret) dev_err(mdev->dev, "DMA channel did not reset "); /* Clear all status bits */ |
6084fc2ec dmaengine: altera... |
485 |
iowrite32(MSGDMA_CSR_STAT_MASK, mdev->csr + MSGDMA_CSR_STATUS); |
a85c6f1b2 dmaengine: Add dr... |
486 487 488 |
/* Enable the DMA controller including interrupts */ iowrite32(MSGDMA_CSR_CTL_STOP_ON_ERR | MSGDMA_CSR_CTL_STOP_ON_EARLY | |
6084fc2ec dmaengine: altera... |
489 |
MSGDMA_CSR_CTL_GLOBAL_INTR, mdev->csr + MSGDMA_CSR_CONTROL); |
a85c6f1b2 dmaengine: Add dr... |
490 491 492 493 494 495 496 |
mdev->idle = true; }; static void msgdma_copy_one(struct msgdma_device *mdev, struct msgdma_sw_desc *desc) { |
6084fc2ec dmaengine: altera... |
497 |
void __iomem *hw_desc = mdev->desc; |
a85c6f1b2 dmaengine: Add dr... |
498 499 500 501 502 |
/* * Check if the DESC FIFO it not full. If its full, we need to wait * for at least one entry to become free again */ |
6084fc2ec dmaengine: altera... |
503 504 |
while (ioread32(mdev->csr + MSGDMA_CSR_STATUS) & MSGDMA_CSR_STAT_DESC_BUF_FULL) |
a85c6f1b2 dmaengine: Add dr... |
505 506 507 508 509 510 511 512 513 514 515 |
mdelay(1); /* * The descriptor needs to get copied into the descriptor FIFO * of the DMA controller. The descriptor will get flushed to the * FIFO, once the last word (control word) is written. Since we * are not 100% sure that memcpy() writes all word in the "correct" * oder (address from low to high) on all architectures, we make * sure this control word is written last by single coding it and * adding some write-barriers here. */ |
6084fc2ec dmaengine: altera... |
516 517 |
memcpy((void __force *)hw_desc, &desc->hw_desc, sizeof(desc->hw_desc) - sizeof(u32)); |
a85c6f1b2 dmaengine: Add dr... |
518 519 520 521 |
/* Write control word last to flush this descriptor into the FIFO */ mdev->idle = false; wmb(); |
6084fc2ec dmaengine: altera... |
522 523 |
iowrite32(desc->hw_desc.control, hw_desc + offsetof(struct msgdma_extended_desc, control)); |
a85c6f1b2 dmaengine: Add dr... |
524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562 563 564 565 566 567 568 569 |
wmb(); } /** * msgdma_copy_desc_to_fifo - copy descriptor(s) into controller FIFO * @mdev: Pointer to the Altera mSGDMA device structure * @desc: Transaction descriptor pointer */ static void msgdma_copy_desc_to_fifo(struct msgdma_device *mdev, struct msgdma_sw_desc *desc) { struct msgdma_sw_desc *sdesc, *next; msgdma_copy_one(mdev, desc); list_for_each_entry_safe(sdesc, next, &desc->tx_list, node) msgdma_copy_one(mdev, sdesc); } /** * msgdma_start_transfer - Initiate the new transfer * @mdev: Pointer to the Altera mSGDMA device structure */ static void msgdma_start_transfer(struct msgdma_device *mdev) { struct msgdma_sw_desc *desc; if (!mdev->idle) return; desc = list_first_entry_or_null(&mdev->pending_list, struct msgdma_sw_desc, node); if (!desc) return; list_splice_tail_init(&mdev->pending_list, &mdev->active_list); msgdma_copy_desc_to_fifo(mdev, desc); } /** * msgdma_issue_pending - Issue pending transactions * @chan: DMA channel pointer */ static void msgdma_issue_pending(struct dma_chan *chan) { struct msgdma_device *mdev = to_mdev(chan); |
edf10919e dmaengine: altera... |
570 |
unsigned long flags; |
a85c6f1b2 dmaengine: Add dr... |
571 |
|
edf10919e dmaengine: altera... |
572 |
spin_lock_irqsave(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
573 |
msgdma_start_transfer(mdev); |
edf10919e dmaengine: altera... |
574 |
spin_unlock_irqrestore(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
575 576 577 578 579 580 581 582 583 584 585 586 587 588 589 590 591 592 593 594 595 596 597 598 599 600 601 602 603 604 605 606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637 638 |
} /** * msgdma_chan_desc_cleanup - Cleanup the completed descriptors * @mdev: Pointer to the Altera mSGDMA device structure */ static void msgdma_chan_desc_cleanup(struct msgdma_device *mdev) { struct msgdma_sw_desc *desc, *next; list_for_each_entry_safe(desc, next, &mdev->done_list, node) { dma_async_tx_callback callback; void *callback_param; list_del(&desc->node); callback = desc->async_tx.callback; callback_param = desc->async_tx.callback_param; if (callback) { spin_unlock(&mdev->lock); callback(callback_param); spin_lock(&mdev->lock); } /* Run any dependencies, then free the descriptor */ msgdma_free_descriptor(mdev, desc); } } /** * msgdma_complete_descriptor - Mark the active descriptor as complete * @mdev: Pointer to the Altera mSGDMA device structure */ static void msgdma_complete_descriptor(struct msgdma_device *mdev) { struct msgdma_sw_desc *desc; desc = list_first_entry_or_null(&mdev->active_list, struct msgdma_sw_desc, node); if (!desc) return; list_del(&desc->node); dma_cookie_complete(&desc->async_tx); list_add_tail(&desc->node, &mdev->done_list); } /** * msgdma_free_descriptors - Free channel descriptors * @mdev: Pointer to the Altera mSGDMA device structure */ static void msgdma_free_descriptors(struct msgdma_device *mdev) { msgdma_free_desc_list(mdev, &mdev->active_list); msgdma_free_desc_list(mdev, &mdev->pending_list); msgdma_free_desc_list(mdev, &mdev->done_list); } /** * msgdma_free_chan_resources - Free channel resources * @dchan: DMA channel pointer */ static void msgdma_free_chan_resources(struct dma_chan *dchan) { struct msgdma_device *mdev = to_mdev(dchan); |
edf10919e dmaengine: altera... |
639 |
unsigned long flags; |
a85c6f1b2 dmaengine: Add dr... |
640 |
|
edf10919e dmaengine: altera... |
641 |
spin_lock_irqsave(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
642 |
msgdma_free_descriptors(mdev); |
edf10919e dmaengine: altera... |
643 |
spin_unlock_irqrestore(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
644 645 646 647 648 649 650 651 652 653 654 655 656 657 658 659 660 661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 |
kfree(mdev->sw_desq); } /** * msgdma_alloc_chan_resources - Allocate channel resources * @dchan: DMA channel * * Return: Number of descriptors on success and failure value on error */ static int msgdma_alloc_chan_resources(struct dma_chan *dchan) { struct msgdma_device *mdev = to_mdev(dchan); struct msgdma_sw_desc *desc; int i; mdev->sw_desq = kcalloc(MSGDMA_DESC_NUM, sizeof(*desc), GFP_NOWAIT); if (!mdev->sw_desq) return -ENOMEM; mdev->idle = true; mdev->desc_free_cnt = MSGDMA_DESC_NUM; INIT_LIST_HEAD(&mdev->free_list); for (i = 0; i < MSGDMA_DESC_NUM; i++) { desc = mdev->sw_desq + i; dma_async_tx_descriptor_init(&desc->async_tx, &mdev->dmachan); desc->async_tx.tx_submit = msgdma_tx_submit; list_add_tail(&desc->node, &mdev->free_list); } return MSGDMA_DESC_NUM; } /** * msgdma_tasklet - Schedule completion tasklet |
62e13a583 dmaengine: altera... |
680 |
* @t: Pointer to the Altera sSGDMA channel structure |
a85c6f1b2 dmaengine: Add dr... |
681 |
*/ |
6752e40d6 dmaengine: altera... |
682 |
static void msgdma_tasklet(struct tasklet_struct *t) |
a85c6f1b2 dmaengine: Add dr... |
683 |
{ |
6752e40d6 dmaengine: altera... |
684 |
struct msgdma_device *mdev = from_tasklet(mdev, t, irq_tasklet); |
a85c6f1b2 dmaengine: Add dr... |
685 |
u32 count; |
6084fc2ec dmaengine: altera... |
686 687 |
u32 __maybe_unused size; u32 __maybe_unused status; |
edf10919e dmaengine: altera... |
688 |
unsigned long flags; |
a85c6f1b2 dmaengine: Add dr... |
689 |
|
edf10919e dmaengine: altera... |
690 |
spin_lock_irqsave(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
691 692 |
/* Read number of responses that are available */ |
6084fc2ec dmaengine: altera... |
693 |
count = ioread32(mdev->csr + MSGDMA_CSR_RESP_FILL_LEVEL); |
a85c6f1b2 dmaengine: Add dr... |
694 695 696 697 698 699 700 701 702 703 704 |
dev_dbg(mdev->dev, "%s (%d): response count=%d ", __func__, __LINE__, count); while (count--) { /* * Read both longwords to purge this response from the FIFO * On Avalon-MM implementations, size and status do not * have any real values, like transferred bytes or error * bits. So we need to just drop these values. */ |
6084fc2ec dmaengine: altera... |
705 |
size = ioread32(mdev->resp + MSGDMA_RESP_BYTES_TRANSFERRED); |
d9ec46416 dmaengine: altera... |
706 |
status = ioread32(mdev->resp + MSGDMA_RESP_STATUS); |
a85c6f1b2 dmaengine: Add dr... |
707 708 709 710 |
msgdma_complete_descriptor(mdev); msgdma_chan_desc_cleanup(mdev); } |
edf10919e dmaengine: altera... |
711 |
spin_unlock_irqrestore(&mdev->lock, flags); |
a85c6f1b2 dmaengine: Add dr... |
712 713 714 715 716 717 718 719 720 721 722 723 724 |
} /** * msgdma_irq_handler - Altera mSGDMA Interrupt handler * @irq: IRQ number * @data: Pointer to the Altera mSGDMA device structure * * Return: IRQ_HANDLED/IRQ_NONE */ static irqreturn_t msgdma_irq_handler(int irq, void *data) { struct msgdma_device *mdev = data; u32 status; |
6084fc2ec dmaengine: altera... |
725 |
status = ioread32(mdev->csr + MSGDMA_CSR_STATUS); |
a85c6f1b2 dmaengine: Add dr... |
726 727 728 729 730 731 732 733 734 735 736 |
if ((status & MSGDMA_CSR_STAT_BUSY) == 0) { /* Start next transfer if the DMA controller is idle */ spin_lock(&mdev->lock); mdev->idle = true; msgdma_start_transfer(mdev); spin_unlock(&mdev->lock); } tasklet_schedule(&mdev->irq_tasklet); /* Clear interrupt in mSGDMA controller */ |
6084fc2ec dmaengine: altera... |
737 |
iowrite32(MSGDMA_CSR_STAT_IRQ, mdev->csr + MSGDMA_CSR_STATUS); |
a85c6f1b2 dmaengine: Add dr... |
738 739 740 741 742 743 744 745 746 747 748 749 750 751 752 753 754 755 756 757 758 759 760 761 762 763 764 765 766 767 768 769 770 771 772 773 774 775 |
return IRQ_HANDLED; } /** * msgdma_chan_remove - Channel remove function * @mdev: Pointer to the Altera mSGDMA device structure */ static void msgdma_dev_remove(struct msgdma_device *mdev) { if (!mdev) return; devm_free_irq(mdev->dev, mdev->irq, mdev); tasklet_kill(&mdev->irq_tasklet); list_del(&mdev->dmachan.device_node); } static int request_and_map(struct platform_device *pdev, const char *name, struct resource **res, void __iomem **ptr) { struct resource *region; struct device *device = &pdev->dev; *res = platform_get_resource_byname(pdev, IORESOURCE_MEM, name); if (*res == NULL) { dev_err(device, "resource %s not defined ", name); return -ENODEV; } region = devm_request_mem_region(device, (*res)->start, resource_size(*res), dev_name(device)); if (region == NULL) { dev_err(device, "unable to request %s ", name); return -EBUSY; } |
4bdc0d676 remove ioremap_no... |
776 |
*ptr = devm_ioremap(device, region->start, |
a85c6f1b2 dmaengine: Add dr... |
777 778 |
resource_size(region)); if (*ptr == NULL) { |
4bdc0d676 remove ioremap_no... |
779 |
dev_err(device, "ioremap of %s failed!", name); |
a85c6f1b2 dmaengine: Add dr... |
780 781 782 783 784 785 786 787 788 789 790 791 792 793 794 795 796 797 798 799 800 801 802 803 804 805 |
return -ENOMEM; } return 0; } /** * msgdma_probe - Driver probe function * @pdev: Pointer to the platform_device structure * * Return: '0' on success and failure value on error */ static int msgdma_probe(struct platform_device *pdev) { struct msgdma_device *mdev; struct dma_device *dma_dev; struct resource *dma_res; int ret; mdev = devm_kzalloc(&pdev->dev, sizeof(*mdev), GFP_NOWAIT); if (!mdev) return -ENOMEM; mdev->dev = &pdev->dev; /* Map CSR space */ |
6084fc2ec dmaengine: altera... |
806 |
ret = request_and_map(pdev, "csr", &dma_res, &mdev->csr); |
a85c6f1b2 dmaengine: Add dr... |
807 808 809 810 |
if (ret) return ret; /* Map (extended) descriptor space */ |
6084fc2ec dmaengine: altera... |
811 |
ret = request_and_map(pdev, "desc", &dma_res, &mdev->desc); |
a85c6f1b2 dmaengine: Add dr... |
812 813 814 815 |
if (ret) return ret; /* Map response space */ |
6084fc2ec dmaengine: altera... |
816 |
ret = request_and_map(pdev, "resp", &dma_res, &mdev->resp); |
a85c6f1b2 dmaengine: Add dr... |
817 818 819 820 821 822 823 824 825 826 827 828 829 830 |
if (ret) return ret; platform_set_drvdata(pdev, mdev); /* Get interrupt nr from platform data */ mdev->irq = platform_get_irq(pdev, 0); if (mdev->irq < 0) return -ENXIO; ret = devm_request_irq(&pdev->dev, mdev->irq, msgdma_irq_handler, 0, dev_name(&pdev->dev), mdev); if (ret) return ret; |
6752e40d6 dmaengine: altera... |
831 |
tasklet_setup(&mdev->irq_tasklet, msgdma_tasklet); |
a85c6f1b2 dmaengine: Add dr... |
832 833 834 835 836 837 838 839 840 841 842 843 844 845 846 |
dma_cookie_init(&mdev->dmachan); spin_lock_init(&mdev->lock); INIT_LIST_HEAD(&mdev->active_list); INIT_LIST_HEAD(&mdev->pending_list); INIT_LIST_HEAD(&mdev->done_list); INIT_LIST_HEAD(&mdev->free_list); dma_dev = &mdev->dmadev; /* Set DMA capabilities */ dma_cap_zero(dma_dev->cap_mask); dma_cap_set(DMA_MEMCPY, dma_dev->cap_mask); |
a85c6f1b2 dmaengine: Add dr... |
847 848 849 850 851 852 853 854 855 856 857 858 859 860 861 862 863 864 |
dma_cap_set(DMA_SLAVE, dma_dev->cap_mask); dma_dev->src_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); dma_dev->dst_addr_widths = BIT(DMA_SLAVE_BUSWIDTH_4_BYTES); dma_dev->directions = BIT(DMA_MEM_TO_DEV) | BIT(DMA_DEV_TO_MEM) | BIT(DMA_MEM_TO_MEM); dma_dev->residue_granularity = DMA_RESIDUE_GRANULARITY_DESCRIPTOR; /* Init DMA link list */ INIT_LIST_HEAD(&dma_dev->channels); /* Set base routines */ dma_dev->device_tx_status = dma_cookie_status; dma_dev->device_issue_pending = msgdma_issue_pending; dma_dev->dev = &pdev->dev; dma_dev->copy_align = DMAENGINE_ALIGN_4_BYTES; dma_dev->device_prep_dma_memcpy = msgdma_prep_memcpy; |
a85c6f1b2 dmaengine: Add dr... |
865 866 867 868 869 870 871 872 873 874 875 876 877 878 879 880 881 882 883 884 885 886 887 888 889 890 891 892 893 894 895 896 897 898 899 900 901 902 903 904 905 906 907 908 909 910 911 912 913 914 915 916 917 918 919 920 921 922 923 924 925 926 927 928 929 930 931 932 |
dma_dev->device_prep_slave_sg = msgdma_prep_slave_sg; dma_dev->device_config = msgdma_dma_config; dma_dev->device_alloc_chan_resources = msgdma_alloc_chan_resources; dma_dev->device_free_chan_resources = msgdma_free_chan_resources; mdev->dmachan.device = dma_dev; list_add_tail(&mdev->dmachan.device_node, &dma_dev->channels); /* Set DMA mask to 64 bits */ ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(64)); if (ret) { dev_warn(&pdev->dev, "unable to set coherent mask to 64"); ret = dma_set_mask_and_coherent(&pdev->dev, DMA_BIT_MASK(32)); if (ret) goto fail; } msgdma_reset(mdev); ret = dma_async_device_register(dma_dev); if (ret) goto fail; dev_notice(&pdev->dev, "Altera mSGDMA driver probe success "); return 0; fail: msgdma_dev_remove(mdev); return ret; } /** * msgdma_dma_remove - Driver remove function * @pdev: Pointer to the platform_device structure * * Return: Always '0' */ static int msgdma_remove(struct platform_device *pdev) { struct msgdma_device *mdev = platform_get_drvdata(pdev); dma_async_device_unregister(&mdev->dmadev); msgdma_dev_remove(mdev); dev_notice(&pdev->dev, "Altera mSGDMA driver removed "); return 0; } static struct platform_driver msgdma_driver = { .driver = { .name = "altera-msgdma", }, .probe = msgdma_probe, .remove = msgdma_remove, }; module_platform_driver(msgdma_driver); MODULE_ALIAS("platform:altera-msgdma"); MODULE_DESCRIPTION("Altera mSGDMA driver"); MODULE_AUTHOR("Stefan Roese <sr@denx.de>"); MODULE_LICENSE("GPL"); |