summaryrefslogtreecommitdiff
path: root/sound/core/memalloc.c
diff options
context:
space:
mode:
authorTakashi Iwai <tiwai@suse.de>2021-06-09 18:25:50 +0200
committerTakashi Iwai <tiwai@suse.de>2021-06-10 10:15:21 +0200
commita202bd1ad86d59c07f24f0a201c5ade320b51e30 (patch)
tree54a7ef6463fbc5e02033fd9c97d77b6a8c4df195 /sound/core/memalloc.c
parent37af81c5998f4b0f23fb452cffa4b8a1c00ce95b (diff)
ALSA: core: Move mmap handler into memalloc ops
This patch moves the mmap handling code into the common memalloc handler. It allows us to reduce the memory-type specific code in PCM code gracefully. Link: https://lore.kernel.org/r/20210609162551.7842-5-tiwai@suse.de Signed-off-by: Takashi Iwai <tiwai@suse.de>
Diffstat (limited to 'sound/core/memalloc.c')
-rw-r--r--sound/core/memalloc.c36
1 files changed, 36 insertions, 0 deletions
diff --git a/sound/core/memalloc.c b/sound/core/memalloc.c
index ad68bcdf82cf..bdcb9230acf8 100644
--- a/sound/core/memalloc.c
+++ b/sound/core/memalloc.c
@@ -128,6 +128,23 @@ void snd_dma_free_pages(struct snd_dma_buffer *dmab)
EXPORT_SYMBOL(snd_dma_free_pages);
/**
+ * snd_dma_buffer_mmap - perform mmap of the given DMA buffer
+ * @dmab: buffer allocation information
+ * @area: VM area information
+ */
+int snd_dma_buffer_mmap(struct snd_dma_buffer *dmab,
+ struct vm_area_struct *area)
+{
+ const struct snd_malloc_ops *ops = snd_dma_get_ops(dmab);
+
+ if (ops && ops->mmap)
+ return ops->mmap(dmab, area);
+ else
+ return -ENOENT;
+}
+EXPORT_SYMBOL(snd_dma_buffer_mmap);
+
+/**
* snd_sgbuf_get_addr - return the physical address at the corresponding offset
* @dmab: buffer allocation information
* @offset: offset in the ring buffer
@@ -283,9 +300,20 @@ static void snd_dma_iram_free(struct snd_dma_buffer *dmab)
gen_pool_free(pool, (unsigned long)dmab->area, dmab->bytes);
}
+static int snd_dma_iram_mmap(struct snd_dma_buffer *dmab,
+ struct vm_area_struct *area)
+{
+ area->vm_page_prot = pgprot_writecombine(area->vm_page_prot);
+ return remap_pfn_range(area, area->vm_start,
+ dmab->addr >> PAGE_SHIFT,
+ area->vm_end - area->vm_start,
+ area->vm_page_prot);
+}
+
static const struct snd_malloc_ops snd_dma_iram_ops = {
.alloc = snd_dma_iram_alloc,
.free = snd_dma_iram_free,
+ .mmap = snd_dma_iram_mmap,
};
#endif /* CONFIG_GENERIC_ALLOCATOR */
@@ -320,9 +348,17 @@ static void snd_dma_dev_free(struct snd_dma_buffer *dmab)
dma_free_coherent(dmab->dev.dev, dmab->bytes, dmab->area, dmab->addr);
}
+static int snd_dma_dev_mmap(struct snd_dma_buffer *dmab,
+ struct vm_area_struct *area)
+{
+ return dma_mmap_coherent(dmab->dev.dev, area,
+ dmab->area, dmab->addr, dmab->bytes);
+}
+
static const struct snd_malloc_ops snd_dma_dev_ops = {
.alloc = snd_dma_dev_alloc,
.free = snd_dma_dev_free,
+ .mmap = snd_dma_dev_mmap,
};
#endif /* CONFIG_HAS_DMA */