Lines Matching defs:lcla_pool

578  * @lcla_pool: lcla pool settings and data.
615 struct d40_lcla_pool lcla_pool;
733 spin_lock_irqsave(&d40c->base->lcla_pool.lock, flags);
742 if (!d40c->base->lcla_pool.alloc_map[idx]) {
743 d40c->base->lcla_pool.alloc_map[idx] = d40d;
750 spin_unlock_irqrestore(&d40c->base->lcla_pool.lock, flags);
765 spin_lock_irqsave(&d40c->base->lcla_pool.lock, flags);
770 if (d40c->base->lcla_pool.alloc_map[idx] == d40d) {
771 d40c->base->lcla_pool.alloc_map[idx] = NULL;
780 spin_unlock_irqrestore(&d40c->base->lcla_pool.lock, flags);
855 struct d40_lcla_pool *pool = &chan->base->lcla_pool;
3285 base->lcla_pool.alloc_map = devm_kcalloc(dev, num_phy_chans
3287 sizeof(*base->lcla_pool.alloc_map),
3289 if (!base->lcla_pool.alloc_map)
3370 struct d40_lcla_pool *pool = &base->lcla_pool;
3387 base->lcla_pool.pages = SZ_1K * base->num_phy_chans / PAGE_SIZE;
3391 base->lcla_pool.pages);
3395 base->lcla_pool.pages);
3399 free_pages(page_list[j], base->lcla_pool.pages);
3409 free_pages(page_list[j], base->lcla_pool.pages);
3412 base->lcla_pool.base = (void *)page_list[i];
3420 __func__, base->lcla_pool.pages);
3421 base->lcla_pool.base_unaligned = kmalloc(SZ_1K *
3425 if (!base->lcla_pool.base_unaligned) {
3430 base->lcla_pool.base = PTR_ALIGN(base->lcla_pool.base_unaligned,
3443 writel(virt_to_phys(base->lcla_pool.base),
3573 base->lcla_pool.base = devm_ioremap(dev, res->start,
3575 if (!base->lcla_pool.base) {
3590 spin_lock_init(&base->lcla_pool.lock);
3656 if (base->lcla_pool.dma_addr)
3657 dma_unmap_single(base->dev, base->lcla_pool.dma_addr,
3661 if (!base->lcla_pool.base_unaligned && base->lcla_pool.base)
3662 free_pages((unsigned long)base->lcla_pool.base,
3663 base->lcla_pool.pages);
3665 kfree(base->lcla_pool.base_unaligned);