summaryrefslogtreecommitdiff
path: root/arch/arm/mach-tegra/iovmm-gart.c
AgeCommit message (Collapse)Author
2012-05-30arm: tegra: resolve compilation time warningsSanjay Singh Rawat
Warnings removed are related to unused variables/labels, structure/argument type mismatch, copyright update, function return type mismatch and wrong C coding style. Bug 949219 Change-Id: Ib748d12d5ab3cfc35118be28c29983081cca6cbb Signed-off-by: Sanjay Singh Rawat <srawat@nvidia.com> Reviewed-on: http://git-master/r/103770 Reviewed-by: Automatic_Commit_Validation_User Reviewed-by: Bharat Nihalani <bnihalani@nvidia.com> Reviewed-by: Sachin Nikam <snikam@nvidia.com>
2012-02-09ARM: tegra: gart: Introduce GART_PAGE_SIZE macroHiroshi DOYU
A bit nicer for readibility. Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66362 (cherry picked from commit b8f0aad888dd6b72462a207520d3585951c065f9) Change-Id: I4e4fa5b32a06c8ffee6a9ad002ae0c93b1bf7e80 Signed-off-by: Pritesh Raithatha <praithatha@nvidia.com> Reviewed-on: http://git-master/r/79989 Reviewed-by: Automatic_Commit_Validation_User Reviewed-by: Bharat Nihalani <bnihalani@nvidia.com> Tested-by: Bharat Nihalani <bnihalani@nvidia.com>
2012-02-09ARM: tegra: gart: Fix wrong initialization orderHiroshi DOYU
GART H/W should be enabled after a page table is initialized. Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66361 (cherry picked from commit 223c0ba1413aeb88742b8f9849a9967e4680f126) Change-Id: Iafc26560716273e3871a4555081f22b8bb51bc67 Signed-off-by: Pritesh Raithatha <praithatha@nvidia.com> Reviewed-on: http://git-master/r/79988 Reviewed-by: Automatic_Commit_Validation_User Reviewed-by: Bharat Nihalani <bnihalani@nvidia.com> Tested-by: Bharat Nihalani <bnihalani@nvidia.com>
2012-02-09ARM: tegra: gart: Replace memory barrier with register read-backHiroshi DOYU
Any interaction between any block on PPSB and a block on APB or AHB must have these read-back to ensure the APB/AHB bus transaction is complete before initiating activity on the PPSB block. wmb() won't ensure the above transaction but only register read-back does. So now there's not much benefit ioremap_wc() + wmb() only for GART registers because the number of that access isn't so huge. Replaced memory barrier with register read-back, and also introduced gart_set_pte() for simplification of those register access. Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66358 (cherry picked from commit 4c68b93498ee4912e5f98ddb25d4511107d4d35a) Change-Id: I9e847624f7321bb36c1209340ff6e12b66f7e792 Signed-off-by: Pritesh Raithatha <praithatha@nvidia.com> Reviewed-on: http://git-master/r/79987 Reviewed-by: Automatic_Commit_Validation_User Reviewed-by: Bharat Nihalani <bnihalani@nvidia.com> Tested-by: Bharat Nihalani <bnihalani@nvidia.com>
2011-12-22ARM: tegra: gart: Remove unnecessary wmb()Hiroshi DOYU
Removed wmb()s which are: - Logically unnecessary after while loop. - Right after spinlock which has the same effect as wmb(). Change-Id: I2c4978b12d6bd9d4046a900b5ce7d6f0bdbd70e0 Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66357 Reviewed-by: Rohan Somvanshi <rsomvanshi@nvidia.com> Tested-by: Rohan Somvanshi <rsomvanshi@nvidia.com>
2011-12-15ARM: tegra: gart: Remove unnecessary spin_lockHiroshi DOYU
Nothing else should be using the device until after the probe function returns 0. Change-Id: If676ca004ec8fd87a4fe9472d792a676b83de7cc Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66363 Reviewed-by: Rohan Somvanshi <rsomvanshi@nvidia.com> Tested-by: Rohan Somvanshi <rsomvanshi@nvidia.com>
2011-12-08ARM: tegra: gart: Use resource_size()Hiroshi DOYU
Use resource_size() instead of manual calculation Change-Id: I0c2165b534bdec45fbba7ea0b69ccf10430d3259 Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66365 Reviewed-by: Rohan Somvanshi <rsomvanshi@nvidia.com> Tested-by: Rohan Somvanshi <rsomvanshi@nvidia.com>
2011-12-08ARM: tegra: gart: Remove unused struct member from gart_deviceHiroshi DOYU
"needs_barrier" isn't used. Change-Id: Ic277cfc88449d00f2b26d41ac5104b6bff305461 Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66359 Reviewed-by: Rohan Somvanshi <rsomvanshi@nvidia.com> Tested-by: Rohan Somvanshi <rsomvanshi@nvidia.com>
2011-12-08ARM: tegra: gart: Fix wrong return statementHiroshi DOYU
Remove "return" in void module_exit func(). Change-Id: I0b9a97b27c7dfda17b4607e6f3662cecbe989bd9 Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66360 Reviewed-by: Varun Colbert <vcolbert@nvidia.com> Tested-by: Varun Colbert <vcolbert@nvidia.com>
2011-12-08ARM: tegra: iommu: Clean up checkpatch and CodingStyle errorsHiroshi DOYU
- Fixed checkpatch.pl --strict errors. - Inserted one space around binary operators From Documentation/CodingStyle "3.1: Spaces". - Removed a file path line in the head of file. - Updated Copyright year - Removed duplicated header inclusions Change-Id: I750e31cf6e90a9f36e707a6278da6137e1a8ba05 Signed-off-by: Hiroshi DOYU <hdoyu@nvidia.com> Reviewed-on: http://git-master/r/66351 Reviewed-by: Sachin Nikam <snikam@nvidia.com> Reviewed-by: Krishna Reddy <vdumpa@nvidia.com>
2011-11-30[ARM/tegra] Add Tegra3 supportScott Williams
Bug 764354 Original-Change-Id: I8a390eb4dae87dceacb97461f23d13554868b046 Reviewed-on: http://git-master/r/12228 Reviewed-by: Scott Williams <scwilliams@nvidia.com> Tested-by: Scott Williams <scwilliams@nvidia.com> Original-Change-Id: I8e6b8303898796419fb5a759cd16edff9aeac081 Rebase-Id: R2866240384c6c24f46bd7ef54bc3dc9140d9e96b
2011-11-30tegra: iovmm-gart: move init call to subsys_initcallGary King
nvmap and nvhost will behave improperly if iovmm is not present when their respective devices are probed; however, the probe ordering depends on the order the initcalls are made to register the drivers. move iovmm-gart into subsys_initcall to ensure that it is registered earlier than other drivers Change-Id: If3e07ce239e593a0833a3381cd1132f5d6ef6786 Signed-off-by: Gary King <gking@nvidia.com>
2011-11-30[ARM] tegra: add I/O virtual memory manager interface (iovmm)Gary King
The Tegra IOVMM is an interface to allow device drivers and subsystems in the kernel to manage the virtual memory spaces visible to I/O devices. The interface has been designed to be scalable to allow for I/O virtual memory hardware which exists in one or more limited apertures of the address space (e.g., a small aperture in physical address space which can perform MMU-like remapping) up to complete virtual addressing with multiple address spaces and memory protection. The interface has been designed to be similar to the Linux virtual memory system; however, operations which would be difficult to implement or nonsensical for DMA devices (e.g., copy-on-write) are not present, and APIs have been added to allow for management of multiple simultaneous active address spaces. The API is broken into four principal objects: areas, clients, domains and devices. Areas ===== An area is a contiguous region of the virtual address space which can be filled with virtual-to-physical translations (and, optionally, protection attributes). The virtual address of the area can be queried and used for DMA operations by the client which created it. As with the Linux vm_area structures, it is the responsibility of whichever code creates an area to ensure that it is populated with appropriate translations. Domains ======= A domain in the IOVMM system is similar to a process in a standard CPU virtual memory system; it represents the entire range of virtual addresses which may be allocated and used for translation. Depending on hardware capabilities, one or more domains may be resident and available for translation. IOVMM areas are allocated from IOVMM domains. Whenever a DMA operation is performed to or from an IOVMM area, its parent domain must be made resident prior to commencing the operation. Clients ======= I/O VMM clients represent any entity which needs to be able to allocate and map system memory into I/O virtual space. Clients are created by name and may be created as part of a "share group," where all clients created in the same share group will observe the same I/O virtual space (i.e., all will use the same IOVMM domain). This is similar to threads inside a process in the CPU virtual memory manager. The callers of the I/O VMM system are responsible for deciding on the granularity of client creation and share group definition; depending on the specific usage model expected by the caller, it may be appropriate to create an IOVMM client per task (if the caller represents an ioctl'able interface to user land), an IOVMM client per driver instance, a common IOVMM client for an entire bus, or a global IOVMM client for an OS subsystem (e.g., the DMA mapping interface). Each client is responsible for ensuring that its IOVMM client's translation is resident on the system prior to performing DMA operations using the IOVMM addresses. This is accomplished by preceding all DMA operations for the client with a call to tegra_iovmm_client_lock (or tegra_iovmm_client_trylock), and following all operations (once complete) with a call to tegra_iovmm_client_unlock. In this regard, clients are cooperatively context- switched, and are expected to behave appropriately. Devices ======= I/O VMM devices are the physical hardware which is responsible for performing the I/O virtual-to-physical translation. Devices are responsible for domain management: the mapping and unmapping operations needed to make translations resident in the domain (including any TLB shootdown or cache invalidation needed to ensure coherency), locking and unlocking domains as they are made resident by clients into the devices' address space(s), and allocating and deallocating the domain objects. Devices are responsible for the allocation and deallocation of domains to allow coalescing of multiple client share groups into a single domain. For example, if the device's hardware only allows a single address space to be translated system-wide, performing full flushes and invalidates of the translation at every client switch may be prohibitively expensive. In these circumstances, a legal implementation of the IOVMM interface includes returning the same domain for all clients on the system (regardless of the originally-specified share group). In this respect, a client can be assured that it will share an address space with all of the other clients in its share group; however, it may also share this address space with other clients, too. Multiple devices may be present in a system; a device should return a NULL domain if it is incapable of servicing the client when it is asked to allocate a domain. ---------------------------------------------------------------------------- IOVMM Client API ================ tegra_iovmm_alloc_client - Called to create a new IOVMM client object; the implementation may create a new domain or return an existing one depending on both the device and the share group. tegra_iovmm_free_client - Frees a client. tegra_iovmm_client_lock - Makes a client's translations resident in the IOVMM device for subsequent DMA operations. May block if the device is incapable of context-switching the client when it is called. Returns -EINTR if the waiting thread is interrupted before the client is locked. tegra_iovmm_client_trylock - Non-blocking version of tegra_iovmm_client_lock tegra_iovmm_client_unlock - Called by clients after DMA operations on IOVMM- translated addresses is complete; allows IOVMM system to context-switch the current client out of the device if needed. tegra_iovmm_create_vm - Called to allocate an IOVMM area. If lazy / demand-loading of pages is desired, clients should supply a pointer to a tegra_iovmm_area_ops structure providing callback functions to load, pin and unpin the physical pages which will be mapped into this IOVMM region. tegra_iovmm_get_vm_size - Called to query the total size of an IOVMM client tegra_iovmm_free_vm - Called to free a IOVMM area, releasing any pinned physical pages mapped by it and to decommit any resources (memory for PTEs / PDEs) required by the VM area. tegra_iovmm_vm_insert_pfn - Called to insert an exact pfn (system memory physical page) into the area at a specific virtual address. Illegal to call if the IOVMM area was originally created with lazy / demand-loading. tegra_iovmm_zap_vm - Called to mark all mappings in the IOVMM area as invalid / no-access, but continues to consume the I/O virtual address space. For lazy / demand-loaded IOVMM areas, a zapped region will not be reloaded until it has been unzapped; DMA operations using the affected translations may fault (if supported by the device). tegra_iovmm_unzap_vm - Called to re-enable lazy / demand-loading of pages for a previously-zapped IOVMM area. tegra_iovmm_find_area_get - Called to find the IOVMM area object corresponding to the specified I/O virtual address, or NULL if the address is not allocated in the client's address space. Increases the reference count on the IOVMM area object tegra_iovmm_area_get - Called to increase the reference count on the IOVMM area object tegra_iovmm_area_put - Called to decrease the reference count on the IOVMM area object IOVMM Device API ================ tegra_iovmm_register - Called to register a new IOVMM device with the IOVMM manager tegra_iovmm_unregister - Called to remove an IOVMM device from the IOVMM manager (unspecified behavior if called while a translation is active and / or in-use) tegra_iovmm_domain_init - Called to initialize all of the IOVMM manager's data structures (block trees, etc.) after allocating a new domain IOVMM Device HAL ================ map - Called to inform the device about a new lazy-mapped IOVMM area. Devices may load the entire VM area when this is called, or at any time prior to the completion of the first read or write operation using the translation. unmap - Called to zap or to decommit translations map_pfn - Called to insert a specific virtual-to-physical translation in the IOVMM area lock_domain - Called to make a domain resident; should return 0 if the domain was successfully context-switched, non-zero if the operation can not be completed (e.g., all available simultaneous hardware translations are locked). If the device can guarantee that every domain it allocates is always usable, this function may be NULL. unlock_domain - Releases a domain from residency, allows the hardware translation to be used by other domains. alloc_domain - Called to allocate a new domain; allowed to return an existing domain free_domain - Called to free a domain. Change-Id: Ic65788777b7aba50ee323fe16fd553ce66c4b87c Signed-off-by: Gary King <gking@nvidia.com>