Lines Matching refs:pd

144 	int (*setup)(struct sh_mobile_i2c_data *pd);
185 static void iic_wr(struct sh_mobile_i2c_data *pd, int offs, unsigned char data)
188 data |= pd->icic;
190 iowrite8(data, pd->reg + offs);
193 static unsigned char iic_rd(struct sh_mobile_i2c_data *pd, int offs)
195 return ioread8(pd->reg + offs);
198 static void iic_set_clr(struct sh_mobile_i2c_data *pd, int offs,
201 iic_wr(pd, offs, (iic_rd(pd, offs) | set) & ~clr);
238 static int sh_mobile_i2c_check_timing(struct sh_mobile_i2c_data *pd)
240 u16 max_val = pd->flags & IIC_FLAG_HAS_ICIC67 ? 0x1ff : 0xff;
242 if (pd->iccl > max_val || pd->icch > max_val) {
243 dev_err(pd->dev, "timing values out of range: L/H=0x%x/0x%x\n",
244 pd->iccl, pd->icch);
249 if (pd->iccl & 0x100)
250 pd->icic |= ICIC_ICCLB8;
252 pd->icic &= ~ICIC_ICCLB8;
255 if (pd->icch & 0x100)
256 pd->icic |= ICIC_ICCHB8;
258 pd->icic &= ~ICIC_ICCHB8;
260 dev_dbg(pd->dev, "timing values: L/H=0x%x/0x%x\n", pd->iccl, pd->icch);
264 static int sh_mobile_i2c_init(struct sh_mobile_i2c_data *pd)
269 i2c_clk_khz = clk_get_rate(pd->clk) / 1000 / pd->clks_per_count;
271 if (pd->bus_speed == I2C_MAX_STANDARD_MODE_FREQ) {
275 } else if (pd->bus_speed == I2C_MAX_FAST_MODE_FREQ) {
280 dev_err(pd->dev, "unrecognized bus speed %lu Hz\n",
281 pd->bus_speed);
285 pd->iccl = sh_mobile_i2c_iccl(i2c_clk_khz, tLOW, tf);
286 pd->icch = sh_mobile_i2c_icch(i2c_clk_khz, tHIGH, tf);
288 return sh_mobile_i2c_check_timing(pd);
291 static int sh_mobile_i2c_v2_init(struct sh_mobile_i2c_data *pd)
296 clks_per_cycle = clk_get_rate(pd->clk) / pd->bus_speed;
297 pd->iccl = DIV_ROUND_UP(clks_per_cycle * 5 / 9 - 1, pd->clks_per_count);
298 pd->icch = DIV_ROUND_UP(clks_per_cycle * 4 / 9 - 5, pd->clks_per_count);
300 return sh_mobile_i2c_check_timing(pd);
303 static unsigned char i2c_op(struct sh_mobile_i2c_data *pd, enum sh_mobile_i2c_op op)
308 dev_dbg(pd->dev, "op %d\n", op);
310 spin_lock_irqsave(&pd->lock, flags);
314 iic_wr(pd, ICCR, ICCR_ICE | ICCR_TRS | ICCR_BBSY);
317 iic_wr(pd, ICIC, ICIC_WAITE | ICIC_ALE | ICIC_TACKE);
318 iic_wr(pd, ICDR, i2c_8bit_addr_from_msg(pd->msg));
321 iic_wr(pd, ICDR, pd->msg->buf[pd->pos]);
324 iic_wr(pd, ICCR, pd->send_stop ? ICCR_ICE | ICCR_TRS
328 iic_wr(pd, ICCR, ICCR_ICE | ICCR_SCP);
331 ret = iic_rd(pd, ICDR);
334 if (!pd->atomic_xfer)
335 iic_wr(pd, ICIC,
337 iic_wr(pd, ICCR, ICCR_ICE | ICCR_RACK);
340 if (!pd->atomic_xfer)
341 iic_wr(pd, ICIC,
343 ret = iic_rd(pd, ICDR);
344 iic_wr(pd, ICCR, ICCR_ICE | ICCR_RACK);
348 spin_unlock_irqrestore(&pd->lock, flags);
350 dev_dbg(pd->dev, "op %d, data out 0x%02x\n", op, ret);
354 static int sh_mobile_i2c_isr_tx(struct sh_mobile_i2c_data *pd)
356 if (pd->pos == pd->msg->len) {
357 i2c_op(pd, OP_TX_STOP);
361 if (pd->pos == -1)
362 i2c_op(pd, OP_TX_FIRST);
364 i2c_op(pd, OP_TX);
366 pd->pos++;
370 static int sh_mobile_i2c_isr_rx(struct sh_mobile_i2c_data *pd)
375 real_pos = pd->pos - 2;
377 if (pd->pos == -1) {
378 i2c_op(pd, OP_TX_FIRST);
379 } else if (pd->pos == 0) {
380 i2c_op(pd, OP_TX_TO_RX);
381 } else if (pd->pos == pd->msg->len) {
382 if (pd->stop_after_dma) {
384 i2c_op(pd, OP_RX_STOP);
385 pd->pos++;
390 i2c_op(pd, OP_RX_STOP);
392 pd->msg->buf[real_pos] = i2c_op(pd, OP_RX_STOP_DATA);
394 pd->msg->buf[real_pos] = i2c_op(pd, OP_RX);
398 pd->pos++;
399 return pd->pos == (pd->msg->len + 2);
404 struct sh_mobile_i2c_data *pd = dev_id;
408 sr = iic_rd(pd, ICSR);
409 pd->sr |= sr; /* remember state */
411 dev_dbg(pd->dev, "i2c_isr 0x%02x 0x%02x %s %d %d!\n", sr, pd->sr,
412 (pd->msg->flags & I2C_M_RD) ? "read" : "write",
413 pd->pos, pd->msg->len);
416 if (pd->dma_direction == DMA_TO_DEVICE && pd->pos == 0)
417 iic_set_clr(pd, ICIC, ICIC_TDMAE, 0);
420 iic_wr(pd, ICSR, sr & ~(ICSR_AL | ICSR_TACK));
421 else if (pd->msg->flags & I2C_M_RD)
422 wakeup = sh_mobile_i2c_isr_rx(pd);
424 wakeup = sh_mobile_i2c_isr_tx(pd);
427 if (pd->dma_direction == DMA_FROM_DEVICE && pd->pos == 1)
428 iic_set_clr(pd, ICIC, ICIC_RDMAE, 0);
431 iic_wr(pd, ICSR, sr & ~ICSR_WAIT);
434 pd->sr |= SW_DONE;
435 if (!pd->atomic_xfer)
436 wake_up(&pd->wait);
440 iic_rd(pd, ICSR);
445 static void sh_mobile_i2c_cleanup_dma(struct sh_mobile_i2c_data *pd, bool terminate)
447 struct dma_chan *chan = pd->dma_direction == DMA_FROM_DEVICE
448 ? pd->dma_rx : pd->dma_tx;
454 dma_unmap_single(chan->device->dev, sg_dma_address(&pd->sg),
455 pd->msg->len, pd->dma_direction);
457 pd->dma_direction = DMA_NONE;
462 struct sh_mobile_i2c_data *pd = data;
464 sh_mobile_i2c_cleanup_dma(pd, false);
465 pd->pos = pd->msg->len;
466 pd->stop_after_dma = true;
468 iic_set_clr(pd, ICIC, 0, ICIC_TDMAE | ICIC_RDMAE);
507 static void sh_mobile_i2c_xfer_dma(struct sh_mobile_i2c_data *pd)
509 bool read = pd->msg->flags & I2C_M_RD;
511 struct dma_chan *chan = read ? pd->dma_rx : pd->dma_tx;
518 chan = pd->dma_rx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_DEV_TO_MEM,
519 pd->res->start + ICDR);
521 chan = pd->dma_tx = sh_mobile_i2c_request_dma_chan(pd->dev, DMA_MEM_TO_DEV,
522 pd->res->start + ICDR);
528 dma_addr = dma_map_single(chan->device->dev, pd->dma_buf, pd->msg->len, dir);
530 dev_dbg(pd->dev, "dma map failed, using PIO\n");
534 sg_dma_len(&pd->sg) = pd->msg->len;
535 sg_dma_address(&pd->sg) = dma_addr;
537 pd->dma_direction = dir;
539 txdesc = dmaengine_prep_slave_sg(chan, &pd->sg, 1,
543 dev_dbg(pd->dev, "dma prep slave sg failed, using PIO\n");
544 sh_mobile_i2c_cleanup_dma(pd, false);
549 txdesc->callback_param = pd;
553 dev_dbg(pd->dev, "submitting dma failed, using PIO\n");
554 sh_mobile_i2c_cleanup_dma(pd, false);
561 static void start_ch(struct sh_mobile_i2c_data *pd, struct i2c_msg *usr_msg,
566 iic_wr(pd, ICCR, ICCR_SCP);
569 iic_wr(pd, ICCR, ICCR_ICE | ICCR_SCP);
572 iic_wr(pd, ICCL, pd->iccl & 0xff);
573 iic_wr(pd, ICCH, pd->icch & 0xff);
576 pd->msg = usr_msg;
577 pd->pos = -1;
578 pd->sr = 0;
580 if (pd->atomic_xfer)
583 pd->dma_buf = i2c_get_dma_safe_msg_buf(pd->msg, 8);
584 if (pd->dma_buf)
585 sh_mobile_i2c_xfer_dma(pd);
588 iic_wr(pd, ICIC, ICIC_DTEE | ICIC_WAITE | ICIC_ALE | ICIC_TACKE);
591 static int poll_dte(struct sh_mobile_i2c_data *pd)
596 u_int8_t val = iic_rd(pd, ICSR);
610 static int poll_busy(struct sh_mobile_i2c_data *pd)
615 u_int8_t val = iic_rd(pd, ICSR);
617 dev_dbg(pd->dev, "val 0x%02x pd->sr 0x%02x\n", val, pd->sr);
625 val |= pd->sr;
639 static int sh_mobile_xfer(struct sh_mobile_i2c_data *pd,
648 pm_runtime_get_sync(pd->dev);
652 bool do_start = pd->send_stop || !i;
654 pd->send_stop = i == num - 1 || msg->flags & I2C_M_STOP;
655 pd->stop_after_dma = false;
657 start_ch(pd, msg, do_start);
660 i2c_op(pd, OP_START);
662 if (pd->atomic_xfer) {
663 unsigned long j = jiffies + pd->adap.timeout;
667 !(pd->sr & (ICSR_TACK | SW_DONE))) {
668 unsigned char sr = iic_rd(pd, ICSR);
672 sh_mobile_i2c_isr(0, pd);
681 time_left = wait_event_timeout(pd->wait,
682 pd->sr & (ICSR_TACK | SW_DONE),
683 pd->adap.timeout);
686 i2c_put_dma_safe_msg_buf(pd->dma_buf, pd->msg,
687 pd->stop_after_dma);
691 dev_err(pd->dev, "Transfer request timed out\n");
692 if (pd->dma_direction != DMA_NONE)
693 sh_mobile_i2c_cleanup_dma(pd, true);
699 if (pd->send_stop)
700 err = poll_busy(pd);
702 err = poll_dte(pd);
708 iic_wr(pd, ICCR, ICCR_SCP);
711 pm_runtime_put_sync(pd->dev);
720 struct sh_mobile_i2c_data *pd = i2c_get_adapdata(adapter);
722 pd->atomic_xfer = false;
723 return sh_mobile_xfer(pd, msgs, num);
730 struct sh_mobile_i2c_data *pd = i2c_get_adapdata(adapter);
732 pd->atomic_xfer = true;
733 return sh_mobile_xfer(pd, msgs, num);
754 static int sh_mobile_i2c_r8a7740_workaround(struct sh_mobile_i2c_data *pd)
756 iic_set_clr(pd, ICCR, ICCR_ICE, 0);
757 iic_rd(pd, ICCR); /* dummy read */
759 iic_set_clr(pd, ICSTART, ICSTART_ICSTART, 0);
760 iic_rd(pd, ICSTART); /* dummy read */
764 iic_wr(pd, ICCR, ICCR_SCP);
765 iic_wr(pd, ICSTART, 0);
769 iic_wr(pd, ICCR, ICCR_TRS);
771 iic_wr(pd, ICCR, 0);
773 iic_wr(pd, ICCR, ICCR_TRS);
776 return sh_mobile_i2c_init(pd);
818 static void sh_mobile_i2c_release_dma(struct sh_mobile_i2c_data *pd)
820 if (!IS_ERR(pd->dma_tx)) {
821 dma_release_channel(pd->dma_tx);
822 pd->dma_tx = ERR_PTR(-EPROBE_DEFER);
825 if (!IS_ERR(pd->dma_rx)) {
826 dma_release_channel(pd->dma_rx);
827 pd->dma_rx = ERR_PTR(-EPROBE_DEFER);
831 static int sh_mobile_i2c_hook_irqs(struct platform_device *dev, struct sh_mobile_i2c_data *pd)
843 0, dev_name(&dev->dev), pd);
857 0, dev_name(&dev->dev), pd);
872 struct sh_mobile_i2c_data *pd;
878 pd = devm_kzalloc(&dev->dev, sizeof(struct sh_mobile_i2c_data), GFP_KERNEL);
879 if (!pd)
882 pd->clk = devm_clk_get(&dev->dev, NULL);
883 if (IS_ERR(pd->clk)) {
885 return PTR_ERR(pd->clk);
888 ret = sh_mobile_i2c_hook_irqs(dev, pd);
892 pd->dev = &dev->dev;
893 platform_set_drvdata(dev, pd);
895 pd->reg = devm_platform_get_and_ioremap_resource(dev, 0, &pd->res);
896 if (IS_ERR(pd->reg))
897 return PTR_ERR(pd->reg);
900 pd->bus_speed = (ret || !bus_speed) ? I2C_MAX_STANDARD_MODE_FREQ : bus_speed;
901 pd->clks_per_count = 1;
904 if (resource_size(pd->res) > 0x17)
905 pd->flags |= IIC_FLAG_HAS_ICIC67;
912 pd->clks_per_count = config->clks_per_count;
913 ret = config->setup(pd);
915 ret = sh_mobile_i2c_init(pd);
923 sg_init_table(&pd->sg, 1);
924 pd->dma_direction = DMA_NONE;
925 pd->dma_rx = pd->dma_tx = ERR_PTR(-EPROBE_DEFER);
928 adap = &pd->adap;
929 i2c_set_adapdata(adap, pd);
941 spin_lock_init(&pd->lock);
942 init_waitqueue_head(&pd->wait);
946 sh_mobile_i2c_release_dma(pd);
950 dev_info(&dev->dev, "I2C adapter %d, bus speed %lu Hz\n", adap->nr, pd->bus_speed);
957 struct sh_mobile_i2c_data *pd = platform_get_drvdata(dev);
959 i2c_del_adapter(&pd->adap);
960 sh_mobile_i2c_release_dma(pd);
966 struct sh_mobile_i2c_data *pd = dev_get_drvdata(dev);
968 i2c_mark_adapter_suspended(&pd->adap);
974 struct sh_mobile_i2c_data *pd = dev_get_drvdata(dev);
976 i2c_mark_adapter_resumed(&pd->adap);