spi-rockchip.c revision b839b785182497da67374db216b28213ee7bf1b4
1/* 2 * Copyright (c) 2014, Fuzhou Rockchip Electronics Co., Ltd 3 * Author: Addy Ke <addy.ke@rock-chips.com> 4 * 5 * This program is free software; you can redistribute it and/or modify it 6 * under the terms and conditions of the GNU General Public License, 7 * version 2, as published by the Free Software Foundation. 8 * 9 * This program is distributed in the hope it will be useful, but WITHOUT 10 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or 11 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for 12 * more details. 13 * 14 */ 15 16#include <linux/init.h> 17#include <linux/module.h> 18#include <linux/clk.h> 19#include <linux/err.h> 20#include <linux/delay.h> 21#include <linux/interrupt.h> 22#include <linux/platform_device.h> 23#include <linux/slab.h> 24#include <linux/spi/spi.h> 25#include <linux/scatterlist.h> 26#include <linux/of.h> 27#include <linux/pm_runtime.h> 28#include <linux/io.h> 29#include <linux/scatterlist.h> 30#include <linux/dmaengine.h> 31 32#define DRIVER_NAME "rockchip-spi" 33 34/* SPI register offsets */ 35#define ROCKCHIP_SPI_CTRLR0 0x0000 36#define ROCKCHIP_SPI_CTRLR1 0x0004 37#define ROCKCHIP_SPI_SSIENR 0x0008 38#define ROCKCHIP_SPI_SER 0x000c 39#define ROCKCHIP_SPI_BAUDR 0x0010 40#define ROCKCHIP_SPI_TXFTLR 0x0014 41#define ROCKCHIP_SPI_RXFTLR 0x0018 42#define ROCKCHIP_SPI_TXFLR 0x001c 43#define ROCKCHIP_SPI_RXFLR 0x0020 44#define ROCKCHIP_SPI_SR 0x0024 45#define ROCKCHIP_SPI_IPR 0x0028 46#define ROCKCHIP_SPI_IMR 0x002c 47#define ROCKCHIP_SPI_ISR 0x0030 48#define ROCKCHIP_SPI_RISR 0x0034 49#define ROCKCHIP_SPI_ICR 0x0038 50#define ROCKCHIP_SPI_DMACR 0x003c 51#define ROCKCHIP_SPI_DMATDLR 0x0040 52#define ROCKCHIP_SPI_DMARDLR 0x0044 53#define ROCKCHIP_SPI_TXDR 0x0400 54#define ROCKCHIP_SPI_RXDR 0x0800 55 56/* Bit fields in CTRLR0 */ 57#define CR0_DFS_OFFSET 0 58 59#define CR0_CFS_OFFSET 2 60 61#define CR0_SCPH_OFFSET 6 62 63#define CR0_SCPOL_OFFSET 7 64 65#define CR0_CSM_OFFSET 8 66#define CR0_CSM_KEEP 0x0 67/* ss_n be high for half sclk_out cycles */ 68#define CR0_CSM_HALF 0X1 69/* ss_n be high for one sclk_out cycle */ 70#define CR0_CSM_ONE 0x2 71 72/* ss_n to sclk_out delay */ 73#define CR0_SSD_OFFSET 10 74/* 75 * The period between ss_n active and 76 * sclk_out active is half sclk_out cycles 77 */ 78#define CR0_SSD_HALF 0x0 79/* 80 * The period between ss_n active and 81 * sclk_out active is one sclk_out cycle 82 */ 83#define CR0_SSD_ONE 0x1 84 85#define CR0_EM_OFFSET 11 86#define CR0_EM_LITTLE 0x0 87#define CR0_EM_BIG 0x1 88 89#define CR0_FBM_OFFSET 12 90#define CR0_FBM_MSB 0x0 91#define CR0_FBM_LSB 0x1 92 93#define CR0_BHT_OFFSET 13 94#define CR0_BHT_16BIT 0x0 95#define CR0_BHT_8BIT 0x1 96 97#define CR0_RSD_OFFSET 14 98 99#define CR0_FRF_OFFSET 16 100#define CR0_FRF_SPI 0x0 101#define CR0_FRF_SSP 0x1 102#define CR0_FRF_MICROWIRE 0x2 103 104#define CR0_XFM_OFFSET 18 105#define CR0_XFM_MASK (0x03 << SPI_XFM_OFFSET) 106#define CR0_XFM_TR 0x0 107#define CR0_XFM_TO 0x1 108#define CR0_XFM_RO 0x2 109 110#define CR0_OPM_OFFSET 20 111#define CR0_OPM_MASTER 0x0 112#define CR0_OPM_SLAVE 0x1 113 114#define CR0_MTM_OFFSET 0x21 115 116/* Bit fields in SER, 2bit */ 117#define SER_MASK 0x3 118 119/* Bit fields in SR, 5bit */ 120#define SR_MASK 0x1f 121#define SR_BUSY (1 << 0) 122#define SR_TF_FULL (1 << 1) 123#define SR_TF_EMPTY (1 << 2) 124#define SR_RF_EMPTY (1 << 3) 125#define SR_RF_FULL (1 << 4) 126 127/* Bit fields in ISR, IMR, ISR, RISR, 5bit */ 128#define INT_MASK 0x1f 129#define INT_TF_EMPTY (1 << 0) 130#define INT_TF_OVERFLOW (1 << 1) 131#define INT_RF_UNDERFLOW (1 << 2) 132#define INT_RF_OVERFLOW (1 << 3) 133#define INT_RF_FULL (1 << 4) 134 135/* Bit fields in ICR, 4bit */ 136#define ICR_MASK 0x0f 137#define ICR_ALL (1 << 0) 138#define ICR_RF_UNDERFLOW (1 << 1) 139#define ICR_RF_OVERFLOW (1 << 2) 140#define ICR_TF_OVERFLOW (1 << 3) 141 142/* Bit fields in DMACR */ 143#define RF_DMA_EN (1 << 0) 144#define TF_DMA_EN (1 << 1) 145 146#define RXBUSY (1 << 0) 147#define TXBUSY (1 << 1) 148 149enum rockchip_ssi_type { 150 SSI_MOTO_SPI = 0, 151 SSI_TI_SSP, 152 SSI_NS_MICROWIRE, 153}; 154 155struct rockchip_spi_dma_data { 156 struct dma_chan *ch; 157 enum dma_transfer_direction direction; 158 dma_addr_t addr; 159}; 160 161struct rockchip_spi { 162 struct device *dev; 163 struct spi_master *master; 164 165 struct clk *spiclk; 166 struct clk *apb_pclk; 167 168 void __iomem *regs; 169 /*depth of the FIFO buffer */ 170 u32 fifo_len; 171 /* max bus freq supported */ 172 u32 max_freq; 173 /* supported slave numbers */ 174 enum rockchip_ssi_type type; 175 176 u16 mode; 177 u8 tmode; 178 u8 bpw; 179 u8 n_bytes; 180 unsigned len; 181 u32 speed; 182 183 const void *tx; 184 const void *tx_end; 185 void *rx; 186 void *rx_end; 187 188 u32 state; 189 /* protect state */ 190 spinlock_t lock; 191 192 struct completion xfer_completion; 193 194 u32 use_dma; 195 struct sg_table tx_sg; 196 struct sg_table rx_sg; 197 struct rockchip_spi_dma_data dma_rx; 198 struct rockchip_spi_dma_data dma_tx; 199}; 200 201static inline void spi_enable_chip(struct rockchip_spi *rs, int enable) 202{ 203 writel_relaxed((enable ? 1 : 0), rs->regs + ROCKCHIP_SPI_SSIENR); 204} 205 206static inline void spi_set_clk(struct rockchip_spi *rs, u16 div) 207{ 208 writel_relaxed(div, rs->regs + ROCKCHIP_SPI_BAUDR); 209} 210 211static inline void flush_fifo(struct rockchip_spi *rs) 212{ 213 while (readl_relaxed(rs->regs + ROCKCHIP_SPI_RXFLR)) 214 readl_relaxed(rs->regs + ROCKCHIP_SPI_RXDR); 215} 216 217static inline void wait_for_idle(struct rockchip_spi *rs) 218{ 219 unsigned long timeout = jiffies + msecs_to_jiffies(5); 220 221 do { 222 if (!(readl_relaxed(rs->regs + ROCKCHIP_SPI_SR) & SR_BUSY)) 223 return; 224 } while (time_before(jiffies, timeout)); 225 226 dev_warn(rs->dev, "spi controller is in busy state!\n"); 227} 228 229static u32 get_fifo_len(struct rockchip_spi *rs) 230{ 231 u32 fifo; 232 233 for (fifo = 2; fifo < 32; fifo++) { 234 writel_relaxed(fifo, rs->regs + ROCKCHIP_SPI_TXFTLR); 235 if (fifo != readl_relaxed(rs->regs + ROCKCHIP_SPI_TXFTLR)) 236 break; 237 } 238 239 writel_relaxed(0, rs->regs + ROCKCHIP_SPI_TXFTLR); 240 241 return (fifo == 31) ? 0 : fifo; 242} 243 244static inline u32 tx_max(struct rockchip_spi *rs) 245{ 246 u32 tx_left, tx_room; 247 248 tx_left = (rs->tx_end - rs->tx) / rs->n_bytes; 249 tx_room = rs->fifo_len - readl_relaxed(rs->regs + ROCKCHIP_SPI_TXFLR); 250 251 return min(tx_left, tx_room); 252} 253 254static inline u32 rx_max(struct rockchip_spi *rs) 255{ 256 u32 rx_left = (rs->rx_end - rs->rx) / rs->n_bytes; 257 u32 rx_room = (u32)readl_relaxed(rs->regs + ROCKCHIP_SPI_RXFLR); 258 259 return min(rx_left, rx_room); 260} 261 262static void rockchip_spi_set_cs(struct spi_device *spi, bool enable) 263{ 264 u32 ser; 265 struct rockchip_spi *rs = spi_master_get_devdata(spi->master); 266 267 ser = readl_relaxed(rs->regs + ROCKCHIP_SPI_SER) & SER_MASK; 268 269 /* 270 * drivers/spi/spi.c: 271 * static void spi_set_cs(struct spi_device *spi, bool enable) 272 * { 273 * if (spi->mode & SPI_CS_HIGH) 274 * enable = !enable; 275 * 276 * if (spi->cs_gpio >= 0) 277 * gpio_set_value(spi->cs_gpio, !enable); 278 * else if (spi->master->set_cs) 279 * spi->master->set_cs(spi, !enable); 280 * } 281 * 282 * Note: enable(rockchip_spi_set_cs) = !enable(spi_set_cs) 283 */ 284 if (!enable) 285 ser |= 1 << spi->chip_select; 286 else 287 ser &= ~(1 << spi->chip_select); 288 289 writel_relaxed(ser, rs->regs + ROCKCHIP_SPI_SER); 290} 291 292static int rockchip_spi_prepare_message(struct spi_master *master, 293 struct spi_message *msg) 294{ 295 struct rockchip_spi *rs = spi_master_get_devdata(master); 296 struct spi_device *spi = msg->spi; 297 298 rs->mode = spi->mode; 299 300 return 0; 301} 302 303static int rockchip_spi_unprepare_message(struct spi_master *master, 304 struct spi_message *msg) 305{ 306 unsigned long flags; 307 struct rockchip_spi *rs = spi_master_get_devdata(master); 308 309 spin_lock_irqsave(&rs->lock, flags); 310 311 /* 312 * For DMA mode, we need terminate DMA channel and flush 313 * fifo for the next transfer if DMA thansfer timeout. 314 * unprepare_message() was called by core if transfer complete 315 * or timeout. Maybe it is reasonable for error handling here. 316 */ 317 if (rs->use_dma) { 318 if (rs->state & RXBUSY) { 319 dmaengine_terminate_all(rs->dma_rx.ch); 320 flush_fifo(rs); 321 } 322 323 if (rs->state & TXBUSY) 324 dmaengine_terminate_all(rs->dma_tx.ch); 325 } 326 327 spin_unlock_irqrestore(&rs->lock, flags); 328 329 return 0; 330} 331 332static void rockchip_spi_pio_writer(struct rockchip_spi *rs) 333{ 334 u32 max = tx_max(rs); 335 u32 txw = 0; 336 337 while (max--) { 338 if (rs->n_bytes == 1) 339 txw = *(u8 *)(rs->tx); 340 else 341 txw = *(u16 *)(rs->tx); 342 343 writel_relaxed(txw, rs->regs + ROCKCHIP_SPI_TXDR); 344 rs->tx += rs->n_bytes; 345 } 346} 347 348static void rockchip_spi_pio_reader(struct rockchip_spi *rs) 349{ 350 u32 max = rx_max(rs); 351 u32 rxw; 352 353 while (max--) { 354 rxw = readl_relaxed(rs->regs + ROCKCHIP_SPI_RXDR); 355 if (rs->n_bytes == 1) 356 *(u8 *)(rs->rx) = (u8)rxw; 357 else 358 *(u16 *)(rs->rx) = (u16)rxw; 359 rs->rx += rs->n_bytes; 360 } 361} 362 363static int rockchip_spi_pio_transfer(struct rockchip_spi *rs) 364{ 365 int remain = 0; 366 367 do { 368 if (rs->tx) { 369 remain = rs->tx_end - rs->tx; 370 rockchip_spi_pio_writer(rs); 371 } 372 373 if (rs->rx) { 374 remain = rs->rx_end - rs->rx; 375 rockchip_spi_pio_reader(rs); 376 } 377 378 cpu_relax(); 379 } while (remain); 380 381 /* If tx, wait until the FIFO data completely. */ 382 if (rs->tx) 383 wait_for_idle(rs); 384 385 return 0; 386} 387 388static void rockchip_spi_dma_rxcb(void *data) 389{ 390 unsigned long flags; 391 struct rockchip_spi *rs = data; 392 393 spin_lock_irqsave(&rs->lock, flags); 394 395 rs->state &= ~RXBUSY; 396 if (!(rs->state & TXBUSY)) 397 spi_finalize_current_transfer(rs->master); 398 399 spin_unlock_irqrestore(&rs->lock, flags); 400} 401 402static void rockchip_spi_dma_txcb(void *data) 403{ 404 unsigned long flags; 405 struct rockchip_spi *rs = data; 406 407 /* Wait until the FIFO data completely. */ 408 wait_for_idle(rs); 409 410 spin_lock_irqsave(&rs->lock, flags); 411 412 rs->state &= ~TXBUSY; 413 if (!(rs->state & RXBUSY)) 414 spi_finalize_current_transfer(rs->master); 415 416 spin_unlock_irqrestore(&rs->lock, flags); 417} 418 419static int rockchip_spi_dma_transfer(struct rockchip_spi *rs) 420{ 421 unsigned long flags; 422 struct dma_slave_config rxconf, txconf; 423 struct dma_async_tx_descriptor *rxdesc, *txdesc; 424 425 spin_lock_irqsave(&rs->lock, flags); 426 rs->state &= ~RXBUSY; 427 rs->state &= ~TXBUSY; 428 spin_unlock_irqrestore(&rs->lock, flags); 429 430 if (rs->rx) { 431 rxconf.direction = rs->dma_rx.direction; 432 rxconf.src_addr = rs->dma_rx.addr; 433 rxconf.src_addr_width = rs->n_bytes; 434 rxconf.src_maxburst = rs->n_bytes; 435 dmaengine_slave_config(rs->dma_rx.ch, &rxconf); 436 437 rxdesc = dmaengine_prep_slave_sg( 438 rs->dma_rx.ch, 439 rs->rx_sg.sgl, rs->rx_sg.nents, 440 rs->dma_rx.direction, DMA_PREP_INTERRUPT); 441 442 rxdesc->callback = rockchip_spi_dma_rxcb; 443 rxdesc->callback_param = rs; 444 } 445 446 if (rs->tx) { 447 txconf.direction = rs->dma_tx.direction; 448 txconf.dst_addr = rs->dma_tx.addr; 449 txconf.dst_addr_width = rs->n_bytes; 450 txconf.dst_maxburst = rs->n_bytes; 451 dmaengine_slave_config(rs->dma_tx.ch, &txconf); 452 453 txdesc = dmaengine_prep_slave_sg( 454 rs->dma_tx.ch, 455 rs->tx_sg.sgl, rs->tx_sg.nents, 456 rs->dma_tx.direction, DMA_PREP_INTERRUPT); 457 458 txdesc->callback = rockchip_spi_dma_txcb; 459 txdesc->callback_param = rs; 460 } 461 462 /* rx must be started before tx due to spi instinct */ 463 if (rs->rx) { 464 spin_lock_irqsave(&rs->lock, flags); 465 rs->state |= RXBUSY; 466 spin_unlock_irqrestore(&rs->lock, flags); 467 dmaengine_submit(rxdesc); 468 dma_async_issue_pending(rs->dma_rx.ch); 469 } 470 471 if (rs->tx) { 472 spin_lock_irqsave(&rs->lock, flags); 473 rs->state |= TXBUSY; 474 spin_unlock_irqrestore(&rs->lock, flags); 475 dmaengine_submit(txdesc); 476 dma_async_issue_pending(rs->dma_tx.ch); 477 } 478 479 return 1; 480} 481 482static void rockchip_spi_config(struct rockchip_spi *rs) 483{ 484 u32 div = 0; 485 u32 dmacr = 0; 486 487 u32 cr0 = (CR0_BHT_8BIT << CR0_BHT_OFFSET) 488 | (CR0_SSD_ONE << CR0_SSD_OFFSET); 489 490 cr0 |= (rs->n_bytes << CR0_DFS_OFFSET); 491 cr0 |= ((rs->mode & 0x3) << CR0_SCPH_OFFSET); 492 cr0 |= (rs->tmode << CR0_XFM_OFFSET); 493 cr0 |= (rs->type << CR0_FRF_OFFSET); 494 495 if (rs->use_dma) { 496 if (rs->tx) 497 dmacr |= TF_DMA_EN; 498 if (rs->rx) 499 dmacr |= RF_DMA_EN; 500 } 501 502 /* div doesn't support odd number */ 503 div = rs->max_freq / rs->speed; 504 div = (div + 1) & 0xfffe; 505 506 spi_enable_chip(rs, 0); 507 508 writel_relaxed(cr0, rs->regs + ROCKCHIP_SPI_CTRLR0); 509 510 writel_relaxed(rs->len - 1, rs->regs + ROCKCHIP_SPI_CTRLR1); 511 writel_relaxed(rs->fifo_len / 2 - 1, rs->regs + ROCKCHIP_SPI_TXFTLR); 512 writel_relaxed(rs->fifo_len / 2 - 1, rs->regs + ROCKCHIP_SPI_RXFTLR); 513 514 writel_relaxed(0, rs->regs + ROCKCHIP_SPI_DMATDLR); 515 writel_relaxed(0, rs->regs + ROCKCHIP_SPI_DMARDLR); 516 writel_relaxed(dmacr, rs->regs + ROCKCHIP_SPI_DMACR); 517 518 spi_set_clk(rs, div); 519 520 dev_dbg(rs->dev, "cr0 0x%x, div %d\n", cr0, div); 521 522 spi_enable_chip(rs, 1); 523} 524 525static int rockchip_spi_transfer_one( 526 struct spi_master *master, 527 struct spi_device *spi, 528 struct spi_transfer *xfer) 529{ 530 int ret = 0; 531 struct rockchip_spi *rs = spi_master_get_devdata(master); 532 533 WARN_ON((readl_relaxed(rs->regs + ROCKCHIP_SPI_SR) & SR_BUSY)); 534 535 if (!xfer->tx_buf && !xfer->rx_buf) { 536 dev_err(rs->dev, "No buffer for transfer\n"); 537 return -EINVAL; 538 } 539 540 rs->speed = xfer->speed_hz; 541 rs->bpw = xfer->bits_per_word; 542 rs->n_bytes = rs->bpw >> 3; 543 544 rs->tx = xfer->tx_buf; 545 rs->tx_end = rs->tx + xfer->len; 546 rs->rx = xfer->rx_buf; 547 rs->rx_end = rs->rx + xfer->len; 548 rs->len = xfer->len; 549 550 rs->tx_sg = xfer->tx_sg; 551 rs->rx_sg = xfer->rx_sg; 552 553 if (rs->tx && rs->rx) 554 rs->tmode = CR0_XFM_TR; 555 else if (rs->tx) 556 rs->tmode = CR0_XFM_TO; 557 else if (rs->rx) 558 rs->tmode = CR0_XFM_RO; 559 560 if (master->can_dma && master->can_dma(master, spi, xfer)) 561 rs->use_dma = 1; 562 else 563 rs->use_dma = 0; 564 565 rockchip_spi_config(rs); 566 567 if (rs->use_dma) 568 ret = rockchip_spi_dma_transfer(rs); 569 else 570 ret = rockchip_spi_pio_transfer(rs); 571 572 return ret; 573} 574 575static bool rockchip_spi_can_dma(struct spi_master *master, 576 struct spi_device *spi, 577 struct spi_transfer *xfer) 578{ 579 struct rockchip_spi *rs = spi_master_get_devdata(master); 580 581 return (xfer->len > rs->fifo_len); 582} 583 584static int rockchip_spi_probe(struct platform_device *pdev) 585{ 586 int ret = 0; 587 struct rockchip_spi *rs; 588 struct spi_master *master; 589 struct resource *mem; 590 591 master = spi_alloc_master(&pdev->dev, sizeof(struct rockchip_spi)); 592 if (!master) 593 return -ENOMEM; 594 595 platform_set_drvdata(pdev, master); 596 597 rs = spi_master_get_devdata(master); 598 memset(rs, 0, sizeof(struct rockchip_spi)); 599 600 /* Get basic io resource and map it */ 601 mem = platform_get_resource(pdev, IORESOURCE_MEM, 0); 602 rs->regs = devm_ioremap_resource(&pdev->dev, mem); 603 if (IS_ERR(rs->regs)) { 604 dev_err(&pdev->dev, "Failed to map SPI region\n"); 605 ret = PTR_ERR(rs->regs); 606 goto err_ioremap_resource; 607 } 608 609 rs->apb_pclk = devm_clk_get(&pdev->dev, "apb_pclk"); 610 if (IS_ERR(rs->apb_pclk)) { 611 dev_err(&pdev->dev, "Failed to get apb_pclk\n"); 612 ret = PTR_ERR(rs->apb_pclk); 613 goto err_ioremap_resource; 614 } 615 616 rs->spiclk = devm_clk_get(&pdev->dev, "spiclk"); 617 if (IS_ERR(rs->spiclk)) { 618 dev_err(&pdev->dev, "Failed to get spi_pclk\n"); 619 ret = PTR_ERR(rs->spiclk); 620 goto err_ioremap_resource; 621 } 622 623 ret = clk_prepare_enable(rs->apb_pclk); 624 if (ret) { 625 dev_err(&pdev->dev, "Failed to enable apb_pclk\n"); 626 goto err_ioremap_resource; 627 } 628 629 ret = clk_prepare_enable(rs->spiclk); 630 if (ret) { 631 dev_err(&pdev->dev, "Failed to enable spi_clk\n"); 632 goto err_spiclk_enable; 633 } 634 635 spi_enable_chip(rs, 0); 636 637 rs->type = SSI_MOTO_SPI; 638 rs->master = master; 639 rs->dev = &pdev->dev; 640 rs->max_freq = clk_get_rate(rs->spiclk); 641 642 rs->fifo_len = get_fifo_len(rs); 643 if (!rs->fifo_len) { 644 dev_err(&pdev->dev, "Failed to get fifo length\n"); 645 goto err_get_fifo_len; 646 } 647 648 spin_lock_init(&rs->lock); 649 650 pm_runtime_set_active(&pdev->dev); 651 pm_runtime_enable(&pdev->dev); 652 653 master->auto_runtime_pm = true; 654 master->bus_num = pdev->id; 655 master->mode_bits = SPI_CPOL | SPI_CPHA | SPI_LOOP; 656 master->num_chipselect = 2; 657 master->dev.of_node = pdev->dev.of_node; 658 master->bits_per_word_mask = SPI_BPW_MASK(16) | SPI_BPW_MASK(8); 659 660 master->set_cs = rockchip_spi_set_cs; 661 master->prepare_message = rockchip_spi_prepare_message; 662 master->unprepare_message = rockchip_spi_unprepare_message; 663 master->transfer_one = rockchip_spi_transfer_one; 664 665 rs->dma_tx.ch = dma_request_slave_channel(rs->dev, "tx"); 666 if (!rs->dma_tx.ch) 667 dev_warn(rs->dev, "Failed to request TX DMA channel\n"); 668 669 rs->dma_rx.ch = dma_request_slave_channel(rs->dev, "rx"); 670 if (!rs->dma_rx.ch) { 671 if (rs->dma_tx.ch) { 672 dma_release_channel(rs->dma_tx.ch); 673 rs->dma_tx.ch = NULL; 674 } 675 dev_warn(rs->dev, "Failed to request RX DMA channel\n"); 676 } 677 678 if (rs->dma_tx.ch && rs->dma_rx.ch) { 679 rs->dma_tx.addr = (dma_addr_t)(mem->start + ROCKCHIP_SPI_TXDR); 680 rs->dma_rx.addr = (dma_addr_t)(mem->start + ROCKCHIP_SPI_RXDR); 681 rs->dma_tx.direction = DMA_MEM_TO_DEV; 682 rs->dma_tx.direction = DMA_DEV_TO_MEM; 683 684 master->can_dma = rockchip_spi_can_dma; 685 master->dma_tx = rs->dma_tx.ch; 686 master->dma_rx = rs->dma_rx.ch; 687 } 688 689 ret = devm_spi_register_master(&pdev->dev, master); 690 if (ret) { 691 dev_err(&pdev->dev, "Failed to register master\n"); 692 goto err_register_master; 693 } 694 695 return 0; 696 697err_register_master: 698 if (rs->dma_tx.ch) 699 dma_release_channel(rs->dma_tx.ch); 700 if (rs->dma_rx.ch) 701 dma_release_channel(rs->dma_rx.ch); 702err_get_fifo_len: 703 clk_disable_unprepare(rs->spiclk); 704err_spiclk_enable: 705 clk_disable_unprepare(rs->apb_pclk); 706err_ioremap_resource: 707 spi_master_put(master); 708 709 return ret; 710} 711 712static int rockchip_spi_remove(struct platform_device *pdev) 713{ 714 struct spi_master *master = spi_master_get(platform_get_drvdata(pdev)); 715 struct rockchip_spi *rs = spi_master_get_devdata(master); 716 717 pm_runtime_disable(&pdev->dev); 718 719 clk_disable_unprepare(rs->spiclk); 720 clk_disable_unprepare(rs->apb_pclk); 721 722 if (rs->dma_tx.ch) 723 dma_release_channel(rs->dma_tx.ch); 724 if (rs->dma_rx.ch) 725 dma_release_channel(rs->dma_rx.ch); 726 727 spi_master_put(master); 728 729 return 0; 730} 731 732#ifdef CONFIG_PM_SLEEP 733static int rockchip_spi_suspend(struct device *dev) 734{ 735 int ret = 0; 736 struct spi_master *master = dev_get_drvdata(dev); 737 struct rockchip_spi *rs = spi_master_get_devdata(master); 738 739 ret = spi_master_suspend(rs->master); 740 if (ret) 741 return ret; 742 743 if (!pm_runtime_suspended(dev)) { 744 clk_disable_unprepare(rs->spiclk); 745 clk_disable_unprepare(rs->apb_pclk); 746 } 747 748 return ret; 749} 750 751static int rockchip_spi_resume(struct device *dev) 752{ 753 int ret = 0; 754 struct spi_master *master = dev_get_drvdata(dev); 755 struct rockchip_spi *rs = spi_master_get_devdata(master); 756 757 if (!pm_runtime_suspended(dev)) { 758 ret = clk_prepare_enable(rs->apb_pclk); 759 if (ret < 0) 760 return ret; 761 762 ret = clk_prepare_enable(rs->spiclk); 763 if (ret < 0) { 764 clk_disable_unprepare(rs->apb_pclk); 765 return ret; 766 } 767 } 768 769 ret = spi_master_resume(rs->master); 770 if (ret < 0) { 771 clk_disable_unprepare(rs->spiclk); 772 clk_disable_unprepare(rs->apb_pclk); 773 } 774 775 return ret; 776} 777#endif /* CONFIG_PM_SLEEP */ 778 779#ifdef CONFIG_PM_RUNTIME 780static int rockchip_spi_runtime_suspend(struct device *dev) 781{ 782 struct spi_master *master = dev_get_drvdata(dev); 783 struct rockchip_spi *rs = spi_master_get_devdata(master); 784 785 clk_disable_unprepare(rs->spiclk); 786 clk_disable_unprepare(rs->apb_pclk); 787 788 return 0; 789} 790 791static int rockchip_spi_runtime_resume(struct device *dev) 792{ 793 int ret; 794 struct spi_master *master = dev_get_drvdata(dev); 795 struct rockchip_spi *rs = spi_master_get_devdata(master); 796 797 ret = clk_prepare_enable(rs->apb_pclk); 798 if (ret) 799 return ret; 800 801 ret = clk_prepare_enable(rs->spiclk); 802 if (ret) 803 clk_disable_unprepare(rs->apb_pclk); 804 805 return ret; 806} 807#endif /* CONFIG_PM_RUNTIME */ 808 809static const struct dev_pm_ops rockchip_spi_pm = { 810 SET_SYSTEM_SLEEP_PM_OPS(rockchip_spi_suspend, rockchip_spi_resume) 811 SET_RUNTIME_PM_OPS(rockchip_spi_runtime_suspend, 812 rockchip_spi_runtime_resume, NULL) 813}; 814 815static const struct of_device_id rockchip_spi_dt_match[] = { 816 { .compatible = "rockchip,rk3066-spi", }, 817 { .compatible = "rockchip,rk3188-spi", }, 818 { .compatible = "rockchip,rk3288-spi", }, 819 { }, 820}; 821MODULE_DEVICE_TABLE(of, rockchip_spi_dt_match); 822 823static struct platform_driver rockchip_spi_driver = { 824 .driver = { 825 .name = DRIVER_NAME, 826 .owner = THIS_MODULE, 827 .pm = &rockchip_spi_pm, 828 .of_match_table = of_match_ptr(rockchip_spi_dt_match), 829 }, 830 .probe = rockchip_spi_probe, 831 .remove = rockchip_spi_remove, 832}; 833 834module_platform_driver(rockchip_spi_driver); 835 836MODULE_AUTHOR("Addy Ke <addy.ke@rock-chips.com>"); 837MODULE_DESCRIPTION("ROCKCHIP SPI Controller Driver"); 838MODULE_LICENSE("GPL v2"); 839