Lines Matching defs:hwdesc

178 	struct udma_hwdesc hwdesc[];
391 return d->hwdesc[idx].cppi5_desc_paddr;
396 return d->hwdesc[idx].cppi5_desc_vaddr;
432 if (!d->hwdesc[i].cppi5_desc_vaddr)
436 d->hwdesc[i].cppi5_desc_vaddr,
437 d->hwdesc[i].cppi5_desc_paddr);
439 d->hwdesc[i].cppi5_desc_vaddr = NULL;
441 } else if (d->hwdesc[0].cppi5_desc_vaddr) {
444 dma_free_coherent(ud->dev, d->hwdesc[0].cppi5_desc_size,
445 d->hwdesc[0].cppi5_desc_vaddr,
446 d->hwdesc[0].cppi5_desc_paddr);
448 d->hwdesc[0].cppi5_desc_vaddr = NULL;
922 h_desc = d->hwdesc[d->desc_idx].cppi5_desc_vaddr;
930 struct cppi5_host_desc_t *h_desc = d->hwdesc[0].cppi5_desc_vaddr;
1901 struct udma_hwdesc *hwdesc;
1919 d = kzalloc(sizeof(*d) + sizeof(d->hwdesc[0]), GFP_NOWAIT);
1926 hwdesc = &d->hwdesc[0];
1930 hwdesc->cppi5_desc_size = uc->config.hdesc_size;
1931 hwdesc->cppi5_desc_vaddr = dma_pool_zalloc(uc->hdesc_pool,
1933 &hwdesc->cppi5_desc_paddr);
1935 hwdesc->cppi5_desc_size = cppi5_trdesc_calc_size(tr_size,
1937 hwdesc->cppi5_desc_size = ALIGN(hwdesc->cppi5_desc_size,
1939 hwdesc->cppi5_desc_vaddr = dma_alloc_coherent(uc->ud->dev,
1940 hwdesc->cppi5_desc_size,
1941 &hwdesc->cppi5_desc_paddr,
1945 if (!hwdesc->cppi5_desc_vaddr) {
1951 hwdesc->tr_req_base = hwdesc->cppi5_desc_vaddr + tr_size;
1953 hwdesc->tr_resp_base = hwdesc->tr_req_base + tr_size * tr_count;
1955 tr_desc = hwdesc->cppi5_desc_vaddr;
2048 tr_req = d->hwdesc[0].tr_req_base;
2160 d = kzalloc(struct_size(d, hwdesc, sglen), GFP_NOWAIT);
2173 struct udma_hwdesc *hwdesc = &d->hwdesc[i];
2178 hwdesc->cppi5_desc_vaddr = dma_pool_zalloc(uc->hdesc_pool,
2180 &hwdesc->cppi5_desc_paddr);
2181 if (!hwdesc->cppi5_desc_vaddr) {
2191 hwdesc->cppi5_desc_size = uc->config.hdesc_size;
2192 desc = hwdesc->cppi5_desc_vaddr;
2211 hwdesc->cppi5_desc_paddr);
2226 h_desc = d->hwdesc[0].cppi5_desc_vaddr;
2250 h_desc = d->hwdesc[0].cppi5_desc_vaddr;
2278 h_desc = d->hwdesc[0].cppi5_desc_vaddr;
2307 h_desc = d->hwdesc[0].cppi5_desc_vaddr;
2417 tr_req = d->hwdesc[0].tr_req_base;
2471 d = kzalloc(struct_size(d, hwdesc, periods), GFP_NOWAIT);
2484 struct udma_hwdesc *hwdesc = &d->hwdesc[i];
2488 hwdesc->cppi5_desc_vaddr = dma_pool_zalloc(uc->hdesc_pool,
2490 &hwdesc->cppi5_desc_paddr);
2491 if (!hwdesc->cppi5_desc_vaddr) {
2500 hwdesc->cppi5_desc_size = uc->config.hdesc_size;
2501 h_desc = hwdesc->cppi5_desc_vaddr;
2624 tr_req = d->hwdesc[0].tr_req_base;
2905 status = d->hwdesc[0].tr_resp_base->status;
3355 struct udma_hwdesc *hwdesc;
3372 hwdesc = &rx_flush->hwdescs[0];
3374 hwdesc->cppi5_desc_size = cppi5_trdesc_calc_size(tr_size, 1);
3375 hwdesc->cppi5_desc_size = ALIGN(hwdesc->cppi5_desc_size,
3378 hwdesc->cppi5_desc_vaddr = devm_kzalloc(dev, hwdesc->cppi5_desc_size,
3380 if (!hwdesc->cppi5_desc_vaddr)
3383 hwdesc->cppi5_desc_paddr = dma_map_single(dev, hwdesc->cppi5_desc_vaddr,
3384 hwdesc->cppi5_desc_size,
3386 if (dma_mapping_error(dev, hwdesc->cppi5_desc_paddr))
3390 hwdesc->tr_req_base = hwdesc->cppi5_desc_vaddr + tr_size;
3392 hwdesc->tr_resp_base = hwdesc->tr_req_base + tr_size;
3394 tr_desc = hwdesc->cppi5_desc_vaddr;
3399 tr_req = hwdesc->tr_req_base;
3408 dma_sync_single_for_device(dev, hwdesc->cppi5_desc_paddr,
3409 hwdesc->cppi5_desc_size, DMA_TO_DEVICE);
3412 hwdesc = &rx_flush->hwdescs[1];
3413 hwdesc->cppi5_desc_size = ALIGN(sizeof(struct cppi5_host_desc_t) +
3418 hwdesc->cppi5_desc_vaddr = devm_kzalloc(dev, hwdesc->cppi5_desc_size,
3420 if (!hwdesc->cppi5_desc_vaddr)
3423 hwdesc->cppi5_desc_paddr = dma_map_single(dev, hwdesc->cppi5_desc_vaddr,
3424 hwdesc->cppi5_desc_size,
3426 if (dma_mapping_error(dev, hwdesc->cppi5_desc_paddr))
3429 desc = hwdesc->cppi5_desc_vaddr;
3438 dma_sync_single_for_device(dev, hwdesc->cppi5_desc_paddr,
3439 hwdesc->cppi5_desc_size, DMA_TO_DEVICE);