343 lines
		
	
	
	
		
			8.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			343 lines
		
	
	
	
		
			8.6 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (C) 2007-2008 Advanced Micro Devices, Inc.
 | |
|  * Author: Joerg Roedel <joerg.roedel@amd.com>
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify it
 | |
|  * under the terms of the GNU General Public License version 2 as published
 | |
|  * by the Free Software Foundation.
 | |
|  *
 | |
|  * This program is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 | |
|  * GNU General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU General Public License
 | |
|  * along with this program; if not, write to the Free Software
 | |
|  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307 USA
 | |
|  */
 | |
| 
 | |
| #define pr_fmt(fmt)    "%s: " fmt, __func__
 | |
| 
 | |
| #include <linux/device.h>
 | |
| #include <linux/kernel.h>
 | |
| #include <linux/bug.h>
 | |
| #include <linux/types.h>
 | |
| #include <linux/module.h>
 | |
| #include <linux/slab.h>
 | |
| #include <linux/errno.h>
 | |
| #include <linux/iommu.h>
 | |
| 
 | |
| static ssize_t show_iommu_group(struct device *dev,
 | |
| 				struct device_attribute *attr, char *buf)
 | |
| {
 | |
| 	unsigned int groupid;
 | |
| 
 | |
| 	if (iommu_device_group(dev, &groupid))
 | |
| 		return 0;
 | |
| 
 | |
| 	return sprintf(buf, "%u", groupid);
 | |
| }
 | |
| static DEVICE_ATTR(iommu_group, S_IRUGO, show_iommu_group, NULL);
 | |
| 
 | |
| static int add_iommu_group(struct device *dev, void *data)
 | |
| {
 | |
| 	unsigned int groupid;
 | |
| 
 | |
| 	if (iommu_device_group(dev, &groupid) == 0)
 | |
| 		return device_create_file(dev, &dev_attr_iommu_group);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int remove_iommu_group(struct device *dev)
 | |
| {
 | |
| 	unsigned int groupid;
 | |
| 
 | |
| 	if (iommu_device_group(dev, &groupid) == 0)
 | |
| 		device_remove_file(dev, &dev_attr_iommu_group);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int iommu_device_notifier(struct notifier_block *nb,
 | |
| 				 unsigned long action, void *data)
 | |
| {
 | |
| 	struct device *dev = data;
 | |
| 
 | |
| 	if (action == BUS_NOTIFY_ADD_DEVICE)
 | |
| 		return add_iommu_group(dev, NULL);
 | |
| 	else if (action == BUS_NOTIFY_DEL_DEVICE)
 | |
| 		return remove_iommu_group(dev);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static struct notifier_block iommu_device_nb = {
 | |
| 	.notifier_call = iommu_device_notifier,
 | |
| };
 | |
| 
 | |
| static void iommu_bus_init(struct bus_type *bus, struct iommu_ops *ops)
 | |
| {
 | |
| 	bus_register_notifier(bus, &iommu_device_nb);
 | |
| 	bus_for_each_dev(bus, NULL, NULL, add_iommu_group);
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * bus_set_iommu - set iommu-callbacks for the bus
 | |
|  * @bus: bus.
 | |
|  * @ops: the callbacks provided by the iommu-driver
 | |
|  *
 | |
|  * This function is called by an iommu driver to set the iommu methods
 | |
|  * used for a particular bus. Drivers for devices on that bus can use
 | |
|  * the iommu-api after these ops are registered.
 | |
|  * This special function is needed because IOMMUs are usually devices on
 | |
|  * the bus itself, so the iommu drivers are not initialized when the bus
 | |
|  * is set up. With this function the iommu-driver can set the iommu-ops
 | |
|  * afterwards.
 | |
|  */
 | |
| int bus_set_iommu(struct bus_type *bus, struct iommu_ops *ops)
 | |
| {
 | |
| 	if (bus->iommu_ops != NULL)
 | |
| 		return -EBUSY;
 | |
| 
 | |
| 	bus->iommu_ops = ops;
 | |
| 
 | |
| 	/* Do IOMMU specific setup for this bus-type */
 | |
| 	iommu_bus_init(bus, ops);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(bus_set_iommu);
 | |
| 
 | |
| bool iommu_present(struct bus_type *bus)
 | |
| {
 | |
| 	return bus->iommu_ops != NULL;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_present);
 | |
| 
 | |
| /**
 | |
|  * iommu_set_fault_handler() - set a fault handler for an iommu domain
 | |
|  * @domain: iommu domain
 | |
|  * @handler: fault handler
 | |
|  *
 | |
|  * This function should be used by IOMMU users which want to be notified
 | |
|  * whenever an IOMMU fault happens.
 | |
|  *
 | |
|  * The fault handler itself should return 0 on success, and an appropriate
 | |
|  * error code otherwise.
 | |
|  */
 | |
| void iommu_set_fault_handler(struct iommu_domain *domain,
 | |
| 					iommu_fault_handler_t handler)
 | |
| {
 | |
| 	BUG_ON(!domain);
 | |
| 
 | |
| 	domain->handler = handler;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_set_fault_handler);
 | |
| 
 | |
| struct iommu_domain *iommu_domain_alloc(struct bus_type *bus)
 | |
| {
 | |
| 	struct iommu_domain *domain;
 | |
| 	int ret;
 | |
| 
 | |
| 	if (bus == NULL || bus->iommu_ops == NULL)
 | |
| 		return NULL;
 | |
| 
 | |
| 	domain = kzalloc(sizeof(*domain), GFP_KERNEL);
 | |
| 	if (!domain)
 | |
| 		return NULL;
 | |
| 
 | |
| 	domain->ops = bus->iommu_ops;
 | |
| 
 | |
| 	ret = domain->ops->domain_init(domain);
 | |
| 	if (ret)
 | |
| 		goto out_free;
 | |
| 
 | |
| 	return domain;
 | |
| 
 | |
| out_free:
 | |
| 	kfree(domain);
 | |
| 
 | |
| 	return NULL;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_domain_alloc);
 | |
| 
 | |
| void iommu_domain_free(struct iommu_domain *domain)
 | |
| {
 | |
| 	if (likely(domain->ops->domain_destroy != NULL))
 | |
| 		domain->ops->domain_destroy(domain);
 | |
| 
 | |
| 	kfree(domain);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_domain_free);
 | |
| 
 | |
| int iommu_attach_device(struct iommu_domain *domain, struct device *dev)
 | |
| {
 | |
| 	if (unlikely(domain->ops->attach_dev == NULL))
 | |
| 		return -ENODEV;
 | |
| 
 | |
| 	return domain->ops->attach_dev(domain, dev);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_attach_device);
 | |
| 
 | |
| void iommu_detach_device(struct iommu_domain *domain, struct device *dev)
 | |
| {
 | |
| 	if (unlikely(domain->ops->detach_dev == NULL))
 | |
| 		return;
 | |
| 
 | |
| 	domain->ops->detach_dev(domain, dev);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_detach_device);
 | |
| 
 | |
| phys_addr_t iommu_iova_to_phys(struct iommu_domain *domain,
 | |
| 			       unsigned long iova)
 | |
| {
 | |
| 	if (unlikely(domain->ops->iova_to_phys == NULL))
 | |
| 		return 0;
 | |
| 
 | |
| 	return domain->ops->iova_to_phys(domain, iova);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_iova_to_phys);
 | |
| 
 | |
| int iommu_domain_has_cap(struct iommu_domain *domain,
 | |
| 			 unsigned long cap)
 | |
| {
 | |
| 	if (unlikely(domain->ops->domain_has_cap == NULL))
 | |
| 		return 0;
 | |
| 
 | |
| 	return domain->ops->domain_has_cap(domain, cap);
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_domain_has_cap);
 | |
| 
 | |
| int iommu_map(struct iommu_domain *domain, unsigned long iova,
 | |
| 	      phys_addr_t paddr, size_t size, int prot)
 | |
| {
 | |
| 	unsigned long orig_iova = iova;
 | |
| 	unsigned int min_pagesz;
 | |
| 	size_t orig_size = size;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	if (unlikely(domain->ops->map == NULL))
 | |
| 		return -ENODEV;
 | |
| 
 | |
| 	/* find out the minimum page size supported */
 | |
| 	min_pagesz = 1 << __ffs(domain->ops->pgsize_bitmap);
 | |
| 
 | |
| 	/*
 | |
| 	 * both the virtual address and the physical one, as well as
 | |
| 	 * the size of the mapping, must be aligned (at least) to the
 | |
| 	 * size of the smallest page supported by the hardware
 | |
| 	 */
 | |
| 	if (!IS_ALIGNED(iova | paddr | size, min_pagesz)) {
 | |
| 		pr_err("unaligned: iova 0x%lx pa 0x%lx size 0x%lx min_pagesz "
 | |
| 			"0x%x\n", iova, (unsigned long)paddr,
 | |
| 			(unsigned long)size, min_pagesz);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	pr_debug("map: iova 0x%lx pa 0x%lx size 0x%lx\n", iova,
 | |
| 				(unsigned long)paddr, (unsigned long)size);
 | |
| 
 | |
| 	while (size) {
 | |
| 		unsigned long pgsize, addr_merge = iova | paddr;
 | |
| 		unsigned int pgsize_idx;
 | |
| 
 | |
| 		/* Max page size that still fits into 'size' */
 | |
| 		pgsize_idx = __fls(size);
 | |
| 
 | |
| 		/* need to consider alignment requirements ? */
 | |
| 		if (likely(addr_merge)) {
 | |
| 			/* Max page size allowed by both iova and paddr */
 | |
| 			unsigned int align_pgsize_idx = __ffs(addr_merge);
 | |
| 
 | |
| 			pgsize_idx = min(pgsize_idx, align_pgsize_idx);
 | |
| 		}
 | |
| 
 | |
| 		/* build a mask of acceptable page sizes */
 | |
| 		pgsize = (1UL << (pgsize_idx + 1)) - 1;
 | |
| 
 | |
| 		/* throw away page sizes not supported by the hardware */
 | |
| 		pgsize &= domain->ops->pgsize_bitmap;
 | |
| 
 | |
| 		/* make sure we're still sane */
 | |
| 		BUG_ON(!pgsize);
 | |
| 
 | |
| 		/* pick the biggest page */
 | |
| 		pgsize_idx = __fls(pgsize);
 | |
| 		pgsize = 1UL << pgsize_idx;
 | |
| 
 | |
| 		pr_debug("mapping: iova 0x%lx pa 0x%lx pgsize %lu\n", iova,
 | |
| 					(unsigned long)paddr, pgsize);
 | |
| 
 | |
| 		ret = domain->ops->map(domain, iova, paddr, pgsize, prot);
 | |
| 		if (ret)
 | |
| 			break;
 | |
| 
 | |
| 		iova += pgsize;
 | |
| 		paddr += pgsize;
 | |
| 		size -= pgsize;
 | |
| 	}
 | |
| 
 | |
| 	/* unroll mapping in case something went wrong */
 | |
| 	if (ret)
 | |
| 		iommu_unmap(domain, orig_iova, orig_size - size);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_map);
 | |
| 
 | |
| size_t iommu_unmap(struct iommu_domain *domain, unsigned long iova, size_t size)
 | |
| {
 | |
| 	size_t unmapped_page, unmapped = 0;
 | |
| 	unsigned int min_pagesz;
 | |
| 
 | |
| 	if (unlikely(domain->ops->unmap == NULL))
 | |
| 		return -ENODEV;
 | |
| 
 | |
| 	/* find out the minimum page size supported */
 | |
| 	min_pagesz = 1 << __ffs(domain->ops->pgsize_bitmap);
 | |
| 
 | |
| 	/*
 | |
| 	 * The virtual address, as well as the size of the mapping, must be
 | |
| 	 * aligned (at least) to the size of the smallest page supported
 | |
| 	 * by the hardware
 | |
| 	 */
 | |
| 	if (!IS_ALIGNED(iova | size, min_pagesz)) {
 | |
| 		pr_err("unaligned: iova 0x%lx size 0x%lx min_pagesz 0x%x\n",
 | |
| 					iova, (unsigned long)size, min_pagesz);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	pr_debug("unmap this: iova 0x%lx size 0x%lx\n", iova,
 | |
| 							(unsigned long)size);
 | |
| 
 | |
| 	/*
 | |
| 	 * Keep iterating until we either unmap 'size' bytes (or more)
 | |
| 	 * or we hit an area that isn't mapped.
 | |
| 	 */
 | |
| 	while (unmapped < size) {
 | |
| 		size_t left = size - unmapped;
 | |
| 
 | |
| 		unmapped_page = domain->ops->unmap(domain, iova, left);
 | |
| 		if (!unmapped_page)
 | |
| 			break;
 | |
| 
 | |
| 		pr_debug("unmapped: iova 0x%lx size %lx\n", iova,
 | |
| 					(unsigned long)unmapped_page);
 | |
| 
 | |
| 		iova += unmapped_page;
 | |
| 		unmapped += unmapped_page;
 | |
| 	}
 | |
| 
 | |
| 	return unmapped;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_unmap);
 | |
| 
 | |
| int iommu_device_group(struct device *dev, unsigned int *groupid)
 | |
| {
 | |
| 	if (iommu_present(dev->bus) && dev->bus->iommu_ops->device_group)
 | |
| 		return dev->bus->iommu_ops->device_group(dev, groupid);
 | |
| 
 | |
| 	return -ENODEV;
 | |
| }
 | |
| EXPORT_SYMBOL_GPL(iommu_device_group);
 | 
