mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-21 12:36:45 +07:00
55897af630
While the dma-direct code is (relatively) clean and simple we actually have to use the swiotlb ops for the mapping on many architectures due to devices with addressing limits. Instead of keeping two implementations around this commit allows the dma-direct implementation to call the swiotlb bounce buffering functions and thus share the guts of the mapping implementation. This also simplified the dma-mapping setup on a few architectures where we don't have to differenciate which implementation to use. Signed-off-by: Christoph Hellwig <hch@lst.de> Acked-by: Jesper Dangaard Brouer <brouer@redhat.com> Tested-by: Jesper Dangaard Brouer <brouer@redhat.com> Tested-by: Tony Luck <tony.luck@intel.com>
104 lines
2.7 KiB
C
104 lines
2.7 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef __LINUX_SWIOTLB_H
|
|
#define __LINUX_SWIOTLB_H
|
|
|
|
#include <linux/dma-direction.h>
|
|
#include <linux/init.h>
|
|
#include <linux/types.h>
|
|
|
|
struct device;
|
|
struct page;
|
|
struct scatterlist;
|
|
|
|
enum swiotlb_force {
|
|
SWIOTLB_NORMAL, /* Default - depending on HW DMA mask etc. */
|
|
SWIOTLB_FORCE, /* swiotlb=force */
|
|
SWIOTLB_NO_FORCE, /* swiotlb=noforce */
|
|
};
|
|
|
|
/*
|
|
* Maximum allowable number of contiguous slabs to map,
|
|
* must be a power of 2. What is the appropriate value ?
|
|
* The complexity of {map,unmap}_single is linearly dependent on this value.
|
|
*/
|
|
#define IO_TLB_SEGSIZE 128
|
|
|
|
/*
|
|
* log of the size of each IO TLB slab. The number of slabs is command line
|
|
* controllable.
|
|
*/
|
|
#define IO_TLB_SHIFT 11
|
|
|
|
extern void swiotlb_init(int verbose);
|
|
int swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose);
|
|
extern unsigned long swiotlb_nr_tbl(void);
|
|
unsigned long swiotlb_size_or_default(void);
|
|
extern int swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs);
|
|
extern void __init swiotlb_update_mem_attributes(void);
|
|
|
|
/*
|
|
* Enumeration for sync targets
|
|
*/
|
|
enum dma_sync_target {
|
|
SYNC_FOR_CPU = 0,
|
|
SYNC_FOR_DEVICE = 1,
|
|
};
|
|
|
|
extern phys_addr_t swiotlb_tbl_map_single(struct device *hwdev,
|
|
dma_addr_t tbl_dma_addr,
|
|
phys_addr_t phys, size_t size,
|
|
enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
|
|
extern void swiotlb_tbl_unmap_single(struct device *hwdev,
|
|
phys_addr_t tlb_addr,
|
|
size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs);
|
|
|
|
extern void swiotlb_tbl_sync_single(struct device *hwdev,
|
|
phys_addr_t tlb_addr,
|
|
size_t size, enum dma_data_direction dir,
|
|
enum dma_sync_target target);
|
|
|
|
extern int
|
|
swiotlb_dma_supported(struct device *hwdev, u64 mask);
|
|
|
|
#ifdef CONFIG_SWIOTLB
|
|
extern enum swiotlb_force swiotlb_force;
|
|
extern phys_addr_t io_tlb_start, io_tlb_end;
|
|
|
|
static inline bool is_swiotlb_buffer(phys_addr_t paddr)
|
|
{
|
|
return paddr >= io_tlb_start && paddr < io_tlb_end;
|
|
}
|
|
|
|
bool swiotlb_map(struct device *dev, phys_addr_t *phys, dma_addr_t *dma_addr,
|
|
size_t size, enum dma_data_direction dir, unsigned long attrs);
|
|
void __init swiotlb_exit(void);
|
|
unsigned int swiotlb_max_segment(void);
|
|
#else
|
|
#define swiotlb_force SWIOTLB_NO_FORCE
|
|
static inline bool is_swiotlb_buffer(phys_addr_t paddr)
|
|
{
|
|
return false;
|
|
}
|
|
static inline bool swiotlb_map(struct device *dev, phys_addr_t *phys,
|
|
dma_addr_t *dma_addr, size_t size, enum dma_data_direction dir,
|
|
unsigned long attrs)
|
|
{
|
|
return false;
|
|
}
|
|
static inline void swiotlb_exit(void)
|
|
{
|
|
}
|
|
static inline unsigned int swiotlb_max_segment(void)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif /* CONFIG_SWIOTLB */
|
|
|
|
extern void swiotlb_print_info(void);
|
|
extern void swiotlb_set_max_segment(unsigned int);
|
|
|
|
#endif /* __LINUX_SWIOTLB_H */
|