 2f10ffcfb2
			
		
	
	
	2f10ffcfb2
	
	
	
		
			
			* 'drm-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/airlied/drm-2.6: (29 commits) drm/nouveau: bail out of auxch transaction if we repeatedly recieve defers drm/nv50: implement gpio set/get routines drm/nv50: parse/use some more de-magiced parts of gpio table entries drm/nouveau: store raw gpio table entry in bios gpio structs drm/nv40: Init some tiling-related PGRAPH state. drm/nv50: Add NVA3 support in ctxprog/ctxvals generator. drm/nv50: another dodgy DP hack drm/nv50: punt hotplug irq handling out to workqueue drm/nv50: preserve an unknown SOR_MODECTRL value for DP encoders drm/nv50: Allow using the NVA3 new compute class. drm/nv50: cleanup properly if PDISPLAY init fails drm/nouveau: fixup the init failure paths some more drm/nv50: fix instmem init on IGPs if stolen mem crosses 4GiB mark drm/nv40: add LVDS table quirk for Dell Latitude D620 drm/nv40: rework lvds table parsing drm/nouveau: detect vram amount once, and save the value drm/nouveau: remove some unused members from drm_nouveau_private drm/nouveau: Make use of TTM busy_placements. drm/nv50: add more 0x100c80 flushy magic drm/nv50: fix fbcon when framebuffer above 4GiB mark ...
		
			
				
	
	
		
			923 lines
		
	
	
	
		
			28 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			923 lines
		
	
	
	
		
			28 KiB
			
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright 2005 Stephane Marchesin
 | |
|  * Copyright 2008 Stuart Bennett
 | |
|  * All Rights Reserved.
 | |
|  *
 | |
|  * Permission is hereby granted, free of charge, to any person obtaining a
 | |
|  * copy of this software and associated documentation files (the "Software"),
 | |
|  * to deal in the Software without restriction, including without limitation
 | |
|  * the rights to use, copy, modify, merge, publish, distribute, sublicense,
 | |
|  * and/or sell copies of the Software, and to permit persons to whom the
 | |
|  * Software is furnished to do so, subject to the following conditions:
 | |
|  *
 | |
|  * The above copyright notice and this permission notice (including the next
 | |
|  * paragraph) shall be included in all copies or substantial portions of the
 | |
|  * Software.
 | |
|  *
 | |
|  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
 | |
|  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
 | |
|  * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO EVENT SHALL
 | |
|  * PRECISION INSIGHT AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
 | |
|  * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
 | |
|  * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
 | |
|  * DEALINGS IN THE SOFTWARE.
 | |
|  */
 | |
| 
 | |
| #include <linux/swab.h>
 | |
| #include <linux/slab.h>
 | |
| #include "drmP.h"
 | |
| #include "drm.h"
 | |
| #include "drm_sarea.h"
 | |
| #include "drm_crtc_helper.h"
 | |
| #include <linux/vgaarb.h>
 | |
| #include <linux/vga_switcheroo.h>
 | |
| 
 | |
| #include "nouveau_drv.h"
 | |
| #include "nouveau_drm.h"
 | |
| #include "nv50_display.h"
 | |
| 
 | |
| static void nouveau_stub_takedown(struct drm_device *dev) {}
 | |
| 
 | |
| static int nouveau_init_engine_ptrs(struct drm_device *dev)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 	struct nouveau_engine *engine = &dev_priv->engine;
 | |
| 
 | |
| 	switch (dev_priv->chipset & 0xf0) {
 | |
| 	case 0x00:
 | |
| 		engine->instmem.init		= nv04_instmem_init;
 | |
| 		engine->instmem.takedown	= nv04_instmem_takedown;
 | |
| 		engine->instmem.suspend		= nv04_instmem_suspend;
 | |
| 		engine->instmem.resume		= nv04_instmem_resume;
 | |
| 		engine->instmem.populate	= nv04_instmem_populate;
 | |
| 		engine->instmem.clear		= nv04_instmem_clear;
 | |
| 		engine->instmem.bind		= nv04_instmem_bind;
 | |
| 		engine->instmem.unbind		= nv04_instmem_unbind;
 | |
| 		engine->instmem.prepare_access	= nv04_instmem_prepare_access;
 | |
| 		engine->instmem.finish_access	= nv04_instmem_finish_access;
 | |
| 		engine->mc.init			= nv04_mc_init;
 | |
| 		engine->mc.takedown		= nv04_mc_takedown;
 | |
| 		engine->timer.init		= nv04_timer_init;
 | |
| 		engine->timer.read		= nv04_timer_read;
 | |
| 		engine->timer.takedown		= nv04_timer_takedown;
 | |
| 		engine->fb.init			= nv04_fb_init;
 | |
| 		engine->fb.takedown		= nv04_fb_takedown;
 | |
| 		engine->graph.grclass		= nv04_graph_grclass;
 | |
| 		engine->graph.init		= nv04_graph_init;
 | |
| 		engine->graph.takedown		= nv04_graph_takedown;
 | |
| 		engine->graph.fifo_access	= nv04_graph_fifo_access;
 | |
| 		engine->graph.channel		= nv04_graph_channel;
 | |
| 		engine->graph.create_context	= nv04_graph_create_context;
 | |
| 		engine->graph.destroy_context	= nv04_graph_destroy_context;
 | |
| 		engine->graph.load_context	= nv04_graph_load_context;
 | |
| 		engine->graph.unload_context	= nv04_graph_unload_context;
 | |
| 		engine->fifo.channels		= 16;
 | |
| 		engine->fifo.init		= nv04_fifo_init;
 | |
| 		engine->fifo.takedown		= nouveau_stub_takedown;
 | |
| 		engine->fifo.disable		= nv04_fifo_disable;
 | |
| 		engine->fifo.enable		= nv04_fifo_enable;
 | |
| 		engine->fifo.reassign		= nv04_fifo_reassign;
 | |
| 		engine->fifo.cache_flush	= nv04_fifo_cache_flush;
 | |
| 		engine->fifo.cache_pull		= nv04_fifo_cache_pull;
 | |
| 		engine->fifo.channel_id		= nv04_fifo_channel_id;
 | |
| 		engine->fifo.create_context	= nv04_fifo_create_context;
 | |
| 		engine->fifo.destroy_context	= nv04_fifo_destroy_context;
 | |
| 		engine->fifo.load_context	= nv04_fifo_load_context;
 | |
| 		engine->fifo.unload_context	= nv04_fifo_unload_context;
 | |
| 		break;
 | |
| 	case 0x10:
 | |
| 		engine->instmem.init		= nv04_instmem_init;
 | |
| 		engine->instmem.takedown	= nv04_instmem_takedown;
 | |
| 		engine->instmem.suspend		= nv04_instmem_suspend;
 | |
| 		engine->instmem.resume		= nv04_instmem_resume;
 | |
| 		engine->instmem.populate	= nv04_instmem_populate;
 | |
| 		engine->instmem.clear		= nv04_instmem_clear;
 | |
| 		engine->instmem.bind		= nv04_instmem_bind;
 | |
| 		engine->instmem.unbind		= nv04_instmem_unbind;
 | |
| 		engine->instmem.prepare_access	= nv04_instmem_prepare_access;
 | |
| 		engine->instmem.finish_access	= nv04_instmem_finish_access;
 | |
| 		engine->mc.init			= nv04_mc_init;
 | |
| 		engine->mc.takedown		= nv04_mc_takedown;
 | |
| 		engine->timer.init		= nv04_timer_init;
 | |
| 		engine->timer.read		= nv04_timer_read;
 | |
| 		engine->timer.takedown		= nv04_timer_takedown;
 | |
| 		engine->fb.init			= nv10_fb_init;
 | |
| 		engine->fb.takedown		= nv10_fb_takedown;
 | |
| 		engine->fb.set_region_tiling	= nv10_fb_set_region_tiling;
 | |
| 		engine->graph.grclass		= nv10_graph_grclass;
 | |
| 		engine->graph.init		= nv10_graph_init;
 | |
| 		engine->graph.takedown		= nv10_graph_takedown;
 | |
| 		engine->graph.channel		= nv10_graph_channel;
 | |
| 		engine->graph.create_context	= nv10_graph_create_context;
 | |
| 		engine->graph.destroy_context	= nv10_graph_destroy_context;
 | |
| 		engine->graph.fifo_access	= nv04_graph_fifo_access;
 | |
| 		engine->graph.load_context	= nv10_graph_load_context;
 | |
| 		engine->graph.unload_context	= nv10_graph_unload_context;
 | |
| 		engine->graph.set_region_tiling	= nv10_graph_set_region_tiling;
 | |
| 		engine->fifo.channels		= 32;
 | |
| 		engine->fifo.init		= nv10_fifo_init;
 | |
| 		engine->fifo.takedown		= nouveau_stub_takedown;
 | |
| 		engine->fifo.disable		= nv04_fifo_disable;
 | |
| 		engine->fifo.enable		= nv04_fifo_enable;
 | |
| 		engine->fifo.reassign		= nv04_fifo_reassign;
 | |
| 		engine->fifo.cache_flush	= nv04_fifo_cache_flush;
 | |
| 		engine->fifo.cache_pull		= nv04_fifo_cache_pull;
 | |
| 		engine->fifo.channel_id		= nv10_fifo_channel_id;
 | |
| 		engine->fifo.create_context	= nv10_fifo_create_context;
 | |
| 		engine->fifo.destroy_context	= nv10_fifo_destroy_context;
 | |
| 		engine->fifo.load_context	= nv10_fifo_load_context;
 | |
| 		engine->fifo.unload_context	= nv10_fifo_unload_context;
 | |
| 		break;
 | |
| 	case 0x20:
 | |
| 		engine->instmem.init		= nv04_instmem_init;
 | |
| 		engine->instmem.takedown	= nv04_instmem_takedown;
 | |
| 		engine->instmem.suspend		= nv04_instmem_suspend;
 | |
| 		engine->instmem.resume		= nv04_instmem_resume;
 | |
| 		engine->instmem.populate	= nv04_instmem_populate;
 | |
| 		engine->instmem.clear		= nv04_instmem_clear;
 | |
| 		engine->instmem.bind		= nv04_instmem_bind;
 | |
| 		engine->instmem.unbind		= nv04_instmem_unbind;
 | |
| 		engine->instmem.prepare_access	= nv04_instmem_prepare_access;
 | |
| 		engine->instmem.finish_access	= nv04_instmem_finish_access;
 | |
| 		engine->mc.init			= nv04_mc_init;
 | |
| 		engine->mc.takedown		= nv04_mc_takedown;
 | |
| 		engine->timer.init		= nv04_timer_init;
 | |
| 		engine->timer.read		= nv04_timer_read;
 | |
| 		engine->timer.takedown		= nv04_timer_takedown;
 | |
| 		engine->fb.init			= nv10_fb_init;
 | |
| 		engine->fb.takedown		= nv10_fb_takedown;
 | |
| 		engine->fb.set_region_tiling	= nv10_fb_set_region_tiling;
 | |
| 		engine->graph.grclass		= nv20_graph_grclass;
 | |
| 		engine->graph.init		= nv20_graph_init;
 | |
| 		engine->graph.takedown		= nv20_graph_takedown;
 | |
| 		engine->graph.channel		= nv10_graph_channel;
 | |
| 		engine->graph.create_context	= nv20_graph_create_context;
 | |
| 		engine->graph.destroy_context	= nv20_graph_destroy_context;
 | |
| 		engine->graph.fifo_access	= nv04_graph_fifo_access;
 | |
| 		engine->graph.load_context	= nv20_graph_load_context;
 | |
| 		engine->graph.unload_context	= nv20_graph_unload_context;
 | |
| 		engine->graph.set_region_tiling	= nv20_graph_set_region_tiling;
 | |
| 		engine->fifo.channels		= 32;
 | |
| 		engine->fifo.init		= nv10_fifo_init;
 | |
| 		engine->fifo.takedown		= nouveau_stub_takedown;
 | |
| 		engine->fifo.disable		= nv04_fifo_disable;
 | |
| 		engine->fifo.enable		= nv04_fifo_enable;
 | |
| 		engine->fifo.reassign		= nv04_fifo_reassign;
 | |
| 		engine->fifo.cache_flush	= nv04_fifo_cache_flush;
 | |
| 		engine->fifo.cache_pull		= nv04_fifo_cache_pull;
 | |
| 		engine->fifo.channel_id		= nv10_fifo_channel_id;
 | |
| 		engine->fifo.create_context	= nv10_fifo_create_context;
 | |
| 		engine->fifo.destroy_context	= nv10_fifo_destroy_context;
 | |
| 		engine->fifo.load_context	= nv10_fifo_load_context;
 | |
| 		engine->fifo.unload_context	= nv10_fifo_unload_context;
 | |
| 		break;
 | |
| 	case 0x30:
 | |
| 		engine->instmem.init		= nv04_instmem_init;
 | |
| 		engine->instmem.takedown	= nv04_instmem_takedown;
 | |
| 		engine->instmem.suspend		= nv04_instmem_suspend;
 | |
| 		engine->instmem.resume		= nv04_instmem_resume;
 | |
| 		engine->instmem.populate	= nv04_instmem_populate;
 | |
| 		engine->instmem.clear		= nv04_instmem_clear;
 | |
| 		engine->instmem.bind		= nv04_instmem_bind;
 | |
| 		engine->instmem.unbind		= nv04_instmem_unbind;
 | |
| 		engine->instmem.prepare_access	= nv04_instmem_prepare_access;
 | |
| 		engine->instmem.finish_access	= nv04_instmem_finish_access;
 | |
| 		engine->mc.init			= nv04_mc_init;
 | |
| 		engine->mc.takedown		= nv04_mc_takedown;
 | |
| 		engine->timer.init		= nv04_timer_init;
 | |
| 		engine->timer.read		= nv04_timer_read;
 | |
| 		engine->timer.takedown		= nv04_timer_takedown;
 | |
| 		engine->fb.init			= nv10_fb_init;
 | |
| 		engine->fb.takedown		= nv10_fb_takedown;
 | |
| 		engine->fb.set_region_tiling	= nv10_fb_set_region_tiling;
 | |
| 		engine->graph.grclass		= nv30_graph_grclass;
 | |
| 		engine->graph.init		= nv30_graph_init;
 | |
| 		engine->graph.takedown		= nv20_graph_takedown;
 | |
| 		engine->graph.fifo_access	= nv04_graph_fifo_access;
 | |
| 		engine->graph.channel		= nv10_graph_channel;
 | |
| 		engine->graph.create_context	= nv20_graph_create_context;
 | |
| 		engine->graph.destroy_context	= nv20_graph_destroy_context;
 | |
| 		engine->graph.load_context	= nv20_graph_load_context;
 | |
| 		engine->graph.unload_context	= nv20_graph_unload_context;
 | |
| 		engine->graph.set_region_tiling	= nv20_graph_set_region_tiling;
 | |
| 		engine->fifo.channels		= 32;
 | |
| 		engine->fifo.init		= nv10_fifo_init;
 | |
| 		engine->fifo.takedown		= nouveau_stub_takedown;
 | |
| 		engine->fifo.disable		= nv04_fifo_disable;
 | |
| 		engine->fifo.enable		= nv04_fifo_enable;
 | |
| 		engine->fifo.reassign		= nv04_fifo_reassign;
 | |
| 		engine->fifo.cache_flush	= nv04_fifo_cache_flush;
 | |
| 		engine->fifo.cache_pull		= nv04_fifo_cache_pull;
 | |
| 		engine->fifo.channel_id		= nv10_fifo_channel_id;
 | |
| 		engine->fifo.create_context	= nv10_fifo_create_context;
 | |
| 		engine->fifo.destroy_context	= nv10_fifo_destroy_context;
 | |
| 		engine->fifo.load_context	= nv10_fifo_load_context;
 | |
| 		engine->fifo.unload_context	= nv10_fifo_unload_context;
 | |
| 		break;
 | |
| 	case 0x40:
 | |
| 	case 0x60:
 | |
| 		engine->instmem.init		= nv04_instmem_init;
 | |
| 		engine->instmem.takedown	= nv04_instmem_takedown;
 | |
| 		engine->instmem.suspend		= nv04_instmem_suspend;
 | |
| 		engine->instmem.resume		= nv04_instmem_resume;
 | |
| 		engine->instmem.populate	= nv04_instmem_populate;
 | |
| 		engine->instmem.clear		= nv04_instmem_clear;
 | |
| 		engine->instmem.bind		= nv04_instmem_bind;
 | |
| 		engine->instmem.unbind		= nv04_instmem_unbind;
 | |
| 		engine->instmem.prepare_access	= nv04_instmem_prepare_access;
 | |
| 		engine->instmem.finish_access	= nv04_instmem_finish_access;
 | |
| 		engine->mc.init			= nv40_mc_init;
 | |
| 		engine->mc.takedown		= nv40_mc_takedown;
 | |
| 		engine->timer.init		= nv04_timer_init;
 | |
| 		engine->timer.read		= nv04_timer_read;
 | |
| 		engine->timer.takedown		= nv04_timer_takedown;
 | |
| 		engine->fb.init			= nv40_fb_init;
 | |
| 		engine->fb.takedown		= nv40_fb_takedown;
 | |
| 		engine->fb.set_region_tiling	= nv40_fb_set_region_tiling;
 | |
| 		engine->graph.grclass		= nv40_graph_grclass;
 | |
| 		engine->graph.init		= nv40_graph_init;
 | |
| 		engine->graph.takedown		= nv40_graph_takedown;
 | |
| 		engine->graph.fifo_access	= nv04_graph_fifo_access;
 | |
| 		engine->graph.channel		= nv40_graph_channel;
 | |
| 		engine->graph.create_context	= nv40_graph_create_context;
 | |
| 		engine->graph.destroy_context	= nv40_graph_destroy_context;
 | |
| 		engine->graph.load_context	= nv40_graph_load_context;
 | |
| 		engine->graph.unload_context	= nv40_graph_unload_context;
 | |
| 		engine->graph.set_region_tiling	= nv40_graph_set_region_tiling;
 | |
| 		engine->fifo.channels		= 32;
 | |
| 		engine->fifo.init		= nv40_fifo_init;
 | |
| 		engine->fifo.takedown		= nouveau_stub_takedown;
 | |
| 		engine->fifo.disable		= nv04_fifo_disable;
 | |
| 		engine->fifo.enable		= nv04_fifo_enable;
 | |
| 		engine->fifo.reassign		= nv04_fifo_reassign;
 | |
| 		engine->fifo.cache_flush	= nv04_fifo_cache_flush;
 | |
| 		engine->fifo.cache_pull		= nv04_fifo_cache_pull;
 | |
| 		engine->fifo.channel_id		= nv10_fifo_channel_id;
 | |
| 		engine->fifo.create_context	= nv40_fifo_create_context;
 | |
| 		engine->fifo.destroy_context	= nv40_fifo_destroy_context;
 | |
| 		engine->fifo.load_context	= nv40_fifo_load_context;
 | |
| 		engine->fifo.unload_context	= nv40_fifo_unload_context;
 | |
| 		break;
 | |
| 	case 0x50:
 | |
| 	case 0x80: /* gotta love NVIDIA's consistency.. */
 | |
| 	case 0x90:
 | |
| 	case 0xA0:
 | |
| 		engine->instmem.init		= nv50_instmem_init;
 | |
| 		engine->instmem.takedown	= nv50_instmem_takedown;
 | |
| 		engine->instmem.suspend		= nv50_instmem_suspend;
 | |
| 		engine->instmem.resume		= nv50_instmem_resume;
 | |
| 		engine->instmem.populate	= nv50_instmem_populate;
 | |
| 		engine->instmem.clear		= nv50_instmem_clear;
 | |
| 		engine->instmem.bind		= nv50_instmem_bind;
 | |
| 		engine->instmem.unbind		= nv50_instmem_unbind;
 | |
| 		engine->instmem.prepare_access	= nv50_instmem_prepare_access;
 | |
| 		engine->instmem.finish_access	= nv50_instmem_finish_access;
 | |
| 		engine->mc.init			= nv50_mc_init;
 | |
| 		engine->mc.takedown		= nv50_mc_takedown;
 | |
| 		engine->timer.init		= nv04_timer_init;
 | |
| 		engine->timer.read		= nv04_timer_read;
 | |
| 		engine->timer.takedown		= nv04_timer_takedown;
 | |
| 		engine->fb.init			= nv50_fb_init;
 | |
| 		engine->fb.takedown		= nv50_fb_takedown;
 | |
| 		engine->graph.grclass		= nv50_graph_grclass;
 | |
| 		engine->graph.init		= nv50_graph_init;
 | |
| 		engine->graph.takedown		= nv50_graph_takedown;
 | |
| 		engine->graph.fifo_access	= nv50_graph_fifo_access;
 | |
| 		engine->graph.channel		= nv50_graph_channel;
 | |
| 		engine->graph.create_context	= nv50_graph_create_context;
 | |
| 		engine->graph.destroy_context	= nv50_graph_destroy_context;
 | |
| 		engine->graph.load_context	= nv50_graph_load_context;
 | |
| 		engine->graph.unload_context	= nv50_graph_unload_context;
 | |
| 		engine->fifo.channels		= 128;
 | |
| 		engine->fifo.init		= nv50_fifo_init;
 | |
| 		engine->fifo.takedown		= nv50_fifo_takedown;
 | |
| 		engine->fifo.disable		= nv04_fifo_disable;
 | |
| 		engine->fifo.enable		= nv04_fifo_enable;
 | |
| 		engine->fifo.reassign		= nv04_fifo_reassign;
 | |
| 		engine->fifo.channel_id		= nv50_fifo_channel_id;
 | |
| 		engine->fifo.create_context	= nv50_fifo_create_context;
 | |
| 		engine->fifo.destroy_context	= nv50_fifo_destroy_context;
 | |
| 		engine->fifo.load_context	= nv50_fifo_load_context;
 | |
| 		engine->fifo.unload_context	= nv50_fifo_unload_context;
 | |
| 		break;
 | |
| 	default:
 | |
| 		NV_ERROR(dev, "NV%02x unsupported\n", dev_priv->chipset);
 | |
| 		return 1;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static unsigned int
 | |
| nouveau_vga_set_decode(void *priv, bool state)
 | |
| {
 | |
| 	struct drm_device *dev = priv;
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 
 | |
| 	if (dev_priv->chipset >= 0x40)
 | |
| 		nv_wr32(dev, 0x88054, state);
 | |
| 	else
 | |
| 		nv_wr32(dev, 0x1854, state);
 | |
| 
 | |
| 	if (state)
 | |
| 		return VGA_RSRC_LEGACY_IO | VGA_RSRC_LEGACY_MEM |
 | |
| 		       VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM;
 | |
| 	else
 | |
| 		return VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM;
 | |
| }
 | |
| 
 | |
| static int
 | |
| nouveau_card_init_channel(struct drm_device *dev)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 	struct nouveau_gpuobj *gpuobj;
 | |
| 	int ret;
 | |
| 
 | |
| 	ret = nouveau_channel_alloc(dev, &dev_priv->channel,
 | |
| 				    (struct drm_file *)-2,
 | |
| 				    NvDmaFB, NvDmaTT);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	gpuobj = NULL;
 | |
| 	ret = nouveau_gpuobj_dma_new(dev_priv->channel, NV_CLASS_DMA_IN_MEMORY,
 | |
| 				     0, dev_priv->vram_size,
 | |
| 				     NV_DMA_ACCESS_RW, NV_DMA_TARGET_VIDMEM,
 | |
| 				     &gpuobj);
 | |
| 	if (ret)
 | |
| 		goto out_err;
 | |
| 
 | |
| 	ret = nouveau_gpuobj_ref_add(dev, dev_priv->channel, NvDmaVRAM,
 | |
| 				     gpuobj, NULL);
 | |
| 	if (ret)
 | |
| 		goto out_err;
 | |
| 
 | |
| 	gpuobj = NULL;
 | |
| 	ret = nouveau_gpuobj_gart_dma_new(dev_priv->channel, 0,
 | |
| 					  dev_priv->gart_info.aper_size,
 | |
| 					  NV_DMA_ACCESS_RW, &gpuobj, NULL);
 | |
| 	if (ret)
 | |
| 		goto out_err;
 | |
| 
 | |
| 	ret = nouveau_gpuobj_ref_add(dev, dev_priv->channel, NvDmaGART,
 | |
| 				     gpuobj, NULL);
 | |
| 	if (ret)
 | |
| 		goto out_err;
 | |
| 
 | |
| 	return 0;
 | |
| out_err:
 | |
| 	nouveau_gpuobj_del(dev, &gpuobj);
 | |
| 	nouveau_channel_free(dev_priv->channel);
 | |
| 	dev_priv->channel = NULL;
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void nouveau_switcheroo_set_state(struct pci_dev *pdev,
 | |
| 					 enum vga_switcheroo_state state)
 | |
| {
 | |
| 	pm_message_t pmm = { .event = PM_EVENT_SUSPEND };
 | |
| 	if (state == VGA_SWITCHEROO_ON) {
 | |
| 		printk(KERN_ERR "VGA switcheroo: switched nouveau on\n");
 | |
| 		nouveau_pci_resume(pdev);
 | |
| 	} else {
 | |
| 		printk(KERN_ERR "VGA switcheroo: switched nouveau off\n");
 | |
| 		nouveau_pci_suspend(pdev, pmm);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| static bool nouveau_switcheroo_can_switch(struct pci_dev *pdev)
 | |
| {
 | |
| 	struct drm_device *dev = pci_get_drvdata(pdev);
 | |
| 	bool can_switch;
 | |
| 
 | |
| 	spin_lock(&dev->count_lock);
 | |
| 	can_switch = (dev->open_count == 0);
 | |
| 	spin_unlock(&dev->count_lock);
 | |
| 	return can_switch;
 | |
| }
 | |
| 
 | |
| int
 | |
| nouveau_card_init(struct drm_device *dev)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 	struct nouveau_engine *engine;
 | |
| 	int ret;
 | |
| 
 | |
| 	NV_DEBUG(dev, "prev state = %d\n", dev_priv->init_state);
 | |
| 
 | |
| 	if (dev_priv->init_state == NOUVEAU_CARD_INIT_DONE)
 | |
| 		return 0;
 | |
| 
 | |
| 	vga_client_register(dev->pdev, dev, NULL, nouveau_vga_set_decode);
 | |
| 	vga_switcheroo_register_client(dev->pdev, nouveau_switcheroo_set_state,
 | |
| 				       nouveau_switcheroo_can_switch);
 | |
| 
 | |
| 	/* Initialise internal driver API hooks */
 | |
| 	ret = nouveau_init_engine_ptrs(dev);
 | |
| 	if (ret)
 | |
| 		goto out;
 | |
| 	engine = &dev_priv->engine;
 | |
| 	dev_priv->init_state = NOUVEAU_CARD_INIT_FAILED;
 | |
| 	spin_lock_init(&dev_priv->context_switch_lock);
 | |
| 
 | |
| 	/* Parse BIOS tables / Run init tables if card not POSTed */
 | |
| 	if (drm_core_check_feature(dev, DRIVER_MODESET)) {
 | |
| 		ret = nouveau_bios_init(dev);
 | |
| 		if (ret)
 | |
| 			goto out;
 | |
| 	}
 | |
| 
 | |
| 	ret = nouveau_mem_detect(dev);
 | |
| 	if (ret)
 | |
| 		goto out_bios;
 | |
| 
 | |
| 	ret = nouveau_gpuobj_early_init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_bios;
 | |
| 
 | |
| 	/* Initialise instance memory, must happen before mem_init so we
 | |
| 	 * know exactly how much VRAM we're able to use for "normal"
 | |
| 	 * purposes.
 | |
| 	 */
 | |
| 	ret = engine->instmem.init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_gpuobj_early;
 | |
| 
 | |
| 	/* Setup the memory manager */
 | |
| 	ret = nouveau_mem_init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_instmem;
 | |
| 
 | |
| 	ret = nouveau_gpuobj_init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_mem;
 | |
| 
 | |
| 	/* PMC */
 | |
| 	ret = engine->mc.init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_gpuobj;
 | |
| 
 | |
| 	/* PTIMER */
 | |
| 	ret = engine->timer.init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_mc;
 | |
| 
 | |
| 	/* PFB */
 | |
| 	ret = engine->fb.init(dev);
 | |
| 	if (ret)
 | |
| 		goto out_timer;
 | |
| 
 | |
| 	if (nouveau_noaccel)
 | |
| 		engine->graph.accel_blocked = true;
 | |
| 	else {
 | |
| 		/* PGRAPH */
 | |
| 		ret = engine->graph.init(dev);
 | |
| 		if (ret)
 | |
| 			goto out_fb;
 | |
| 
 | |
| 		/* PFIFO */
 | |
| 		ret = engine->fifo.init(dev);
 | |
| 		if (ret)
 | |
| 			goto out_graph;
 | |
| 	}
 | |
| 
 | |
| 	/* this call irq_preinstall, register irq handler and
 | |
| 	 * call irq_postinstall
 | |
| 	 */
 | |
| 	ret = drm_irq_install(dev);
 | |
| 	if (ret)
 | |
| 		goto out_fifo;
 | |
| 
 | |
| 	ret = drm_vblank_init(dev, 0);
 | |
| 	if (ret)
 | |
| 		goto out_irq;
 | |
| 
 | |
| 	/* what about PVIDEO/PCRTC/PRAMDAC etc? */
 | |
| 
 | |
| 	if (!engine->graph.accel_blocked) {
 | |
| 		ret = nouveau_card_init_channel(dev);
 | |
| 		if (ret)
 | |
| 			goto out_irq;
 | |
| 	}
 | |
| 
 | |
| 	if (drm_core_check_feature(dev, DRIVER_MODESET)) {
 | |
| 		if (dev_priv->card_type >= NV_50)
 | |
| 			ret = nv50_display_create(dev);
 | |
| 		else
 | |
| 			ret = nv04_display_create(dev);
 | |
| 		if (ret)
 | |
| 			goto out_channel;
 | |
| 	}
 | |
| 
 | |
| 	ret = nouveau_backlight_init(dev);
 | |
| 	if (ret)
 | |
| 		NV_ERROR(dev, "Error %d registering backlight\n", ret);
 | |
| 
 | |
| 	dev_priv->init_state = NOUVEAU_CARD_INIT_DONE;
 | |
| 
 | |
| 	if (drm_core_check_feature(dev, DRIVER_MODESET))
 | |
| 		drm_helper_initial_config(dev);
 | |
| 
 | |
| 	return 0;
 | |
| 
 | |
| out_channel:
 | |
| 	if (dev_priv->channel) {
 | |
| 		nouveau_channel_free(dev_priv->channel);
 | |
| 		dev_priv->channel = NULL;
 | |
| 	}
 | |
| out_irq:
 | |
| 	drm_irq_uninstall(dev);
 | |
| out_fifo:
 | |
| 	if (!nouveau_noaccel)
 | |
| 		engine->fifo.takedown(dev);
 | |
| out_graph:
 | |
| 	if (!nouveau_noaccel)
 | |
| 		engine->graph.takedown(dev);
 | |
| out_fb:
 | |
| 	engine->fb.takedown(dev);
 | |
| out_timer:
 | |
| 	engine->timer.takedown(dev);
 | |
| out_mc:
 | |
| 	engine->mc.takedown(dev);
 | |
| out_gpuobj:
 | |
| 	nouveau_gpuobj_takedown(dev);
 | |
| out_mem:
 | |
| 	nouveau_sgdma_takedown(dev);
 | |
| 	nouveau_mem_close(dev);
 | |
| out_instmem:
 | |
| 	engine->instmem.takedown(dev);
 | |
| out_gpuobj_early:
 | |
| 	nouveau_gpuobj_late_takedown(dev);
 | |
| out_bios:
 | |
| 	nouveau_bios_takedown(dev);
 | |
| out:
 | |
| 	vga_client_register(dev->pdev, NULL, NULL, NULL);
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| static void nouveau_card_takedown(struct drm_device *dev)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 	struct nouveau_engine *engine = &dev_priv->engine;
 | |
| 
 | |
| 	NV_DEBUG(dev, "prev state = %d\n", dev_priv->init_state);
 | |
| 
 | |
| 	if (dev_priv->init_state != NOUVEAU_CARD_INIT_DOWN) {
 | |
| 		nouveau_backlight_exit(dev);
 | |
| 
 | |
| 		if (dev_priv->channel) {
 | |
| 			nouveau_channel_free(dev_priv->channel);
 | |
| 			dev_priv->channel = NULL;
 | |
| 		}
 | |
| 
 | |
| 		if (!nouveau_noaccel) {
 | |
| 			engine->fifo.takedown(dev);
 | |
| 			engine->graph.takedown(dev);
 | |
| 		}
 | |
| 		engine->fb.takedown(dev);
 | |
| 		engine->timer.takedown(dev);
 | |
| 		engine->mc.takedown(dev);
 | |
| 
 | |
| 		mutex_lock(&dev->struct_mutex);
 | |
| 		ttm_bo_clean_mm(&dev_priv->ttm.bdev, TTM_PL_VRAM);
 | |
| 		ttm_bo_clean_mm(&dev_priv->ttm.bdev, TTM_PL_TT);
 | |
| 		mutex_unlock(&dev->struct_mutex);
 | |
| 		nouveau_sgdma_takedown(dev);
 | |
| 
 | |
| 		nouveau_gpuobj_takedown(dev);
 | |
| 		nouveau_mem_close(dev);
 | |
| 		engine->instmem.takedown(dev);
 | |
| 
 | |
| 		if (drm_core_check_feature(dev, DRIVER_MODESET))
 | |
| 			drm_irq_uninstall(dev);
 | |
| 
 | |
| 		nouveau_gpuobj_late_takedown(dev);
 | |
| 		nouveau_bios_takedown(dev);
 | |
| 
 | |
| 		vga_client_register(dev->pdev, NULL, NULL, NULL);
 | |
| 
 | |
| 		dev_priv->init_state = NOUVEAU_CARD_INIT_DOWN;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* here a client dies, release the stuff that was allocated for its
 | |
|  * file_priv */
 | |
| void nouveau_preclose(struct drm_device *dev, struct drm_file *file_priv)
 | |
| {
 | |
| 	nouveau_channel_cleanup(dev, file_priv);
 | |
| }
 | |
| 
 | |
| /* first module load, setup the mmio/fb mapping */
 | |
| /* KMS: we need mmio at load time, not when the first drm client opens. */
 | |
| int nouveau_firstopen(struct drm_device *dev)
 | |
| {
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* if we have an OF card, copy vbios to RAMIN */
 | |
| static void nouveau_OF_copy_vbios_to_ramin(struct drm_device *dev)
 | |
| {
 | |
| #if defined(__powerpc__)
 | |
| 	int size, i;
 | |
| 	const uint32_t *bios;
 | |
| 	struct device_node *dn = pci_device_to_OF_node(dev->pdev);
 | |
| 	if (!dn) {
 | |
| 		NV_INFO(dev, "Unable to get the OF node\n");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	bios = of_get_property(dn, "NVDA,BMP", &size);
 | |
| 	if (bios) {
 | |
| 		for (i = 0; i < size; i += 4)
 | |
| 			nv_wi32(dev, i, bios[i/4]);
 | |
| 		NV_INFO(dev, "OF bios successfully copied (%d bytes)\n", size);
 | |
| 	} else {
 | |
| 		NV_INFO(dev, "Unable to get the OF bios\n");
 | |
| 	}
 | |
| #endif
 | |
| }
 | |
| 
 | |
| int nouveau_load(struct drm_device *dev, unsigned long flags)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv;
 | |
| 	uint32_t reg0;
 | |
| 	resource_size_t mmio_start_offs;
 | |
| 
 | |
| 	dev_priv = kzalloc(sizeof(*dev_priv), GFP_KERNEL);
 | |
| 	if (!dev_priv)
 | |
| 		return -ENOMEM;
 | |
| 	dev->dev_private = dev_priv;
 | |
| 	dev_priv->dev = dev;
 | |
| 
 | |
| 	dev_priv->flags = flags & NOUVEAU_FLAGS;
 | |
| 	dev_priv->init_state = NOUVEAU_CARD_INIT_DOWN;
 | |
| 
 | |
| 	NV_DEBUG(dev, "vendor: 0x%X device: 0x%X class: 0x%X\n",
 | |
| 		 dev->pci_vendor, dev->pci_device, dev->pdev->class);
 | |
| 
 | |
| 	dev_priv->wq = create_workqueue("nouveau");
 | |
| 	if (!dev_priv->wq)
 | |
| 		return -EINVAL;
 | |
| 
 | |
| 	/* resource 0 is mmio regs */
 | |
| 	/* resource 1 is linear FB */
 | |
| 	/* resource 2 is RAMIN (mmio regs + 0x1000000) */
 | |
| 	/* resource 6 is bios */
 | |
| 
 | |
| 	/* map the mmio regs */
 | |
| 	mmio_start_offs = pci_resource_start(dev->pdev, 0);
 | |
| 	dev_priv->mmio = ioremap(mmio_start_offs, 0x00800000);
 | |
| 	if (!dev_priv->mmio) {
 | |
| 		NV_ERROR(dev, "Unable to initialize the mmio mapping. "
 | |
| 			 "Please report your setup to " DRIVER_EMAIL "\n");
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	NV_DEBUG(dev, "regs mapped ok at 0x%llx\n",
 | |
| 					(unsigned long long)mmio_start_offs);
 | |
| 
 | |
| #ifdef __BIG_ENDIAN
 | |
| 	/* Put the card in BE mode if it's not */
 | |
| 	if (nv_rd32(dev, NV03_PMC_BOOT_1))
 | |
| 		nv_wr32(dev, NV03_PMC_BOOT_1, 0x00000001);
 | |
| 
 | |
| 	DRM_MEMORYBARRIER();
 | |
| #endif
 | |
| 
 | |
| 	/* Time to determine the card architecture */
 | |
| 	reg0 = nv_rd32(dev, NV03_PMC_BOOT_0);
 | |
| 
 | |
| 	/* We're dealing with >=NV10 */
 | |
| 	if ((reg0 & 0x0f000000) > 0) {
 | |
| 		/* Bit 27-20 contain the architecture in hex */
 | |
| 		dev_priv->chipset = (reg0 & 0xff00000) >> 20;
 | |
| 	/* NV04 or NV05 */
 | |
| 	} else if ((reg0 & 0xff00fff0) == 0x20004000) {
 | |
| 		if (reg0 & 0x00f00000)
 | |
| 			dev_priv->chipset = 0x05;
 | |
| 		else
 | |
| 			dev_priv->chipset = 0x04;
 | |
| 	} else
 | |
| 		dev_priv->chipset = 0xff;
 | |
| 
 | |
| 	switch (dev_priv->chipset & 0xf0) {
 | |
| 	case 0x00:
 | |
| 	case 0x10:
 | |
| 	case 0x20:
 | |
| 	case 0x30:
 | |
| 		dev_priv->card_type = dev_priv->chipset & 0xf0;
 | |
| 		break;
 | |
| 	case 0x40:
 | |
| 	case 0x60:
 | |
| 		dev_priv->card_type = NV_40;
 | |
| 		break;
 | |
| 	case 0x50:
 | |
| 	case 0x80:
 | |
| 	case 0x90:
 | |
| 	case 0xa0:
 | |
| 		dev_priv->card_type = NV_50;
 | |
| 		break;
 | |
| 	default:
 | |
| 		NV_INFO(dev, "Unsupported chipset 0x%08x\n", reg0);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	NV_INFO(dev, "Detected an NV%2x generation card (0x%08x)\n",
 | |
| 		dev_priv->card_type, reg0);
 | |
| 
 | |
| 	/* map larger RAMIN aperture on NV40 cards */
 | |
| 	dev_priv->ramin  = NULL;
 | |
| 	if (dev_priv->card_type >= NV_40) {
 | |
| 		int ramin_bar = 2;
 | |
| 		if (pci_resource_len(dev->pdev, ramin_bar) == 0)
 | |
| 			ramin_bar = 3;
 | |
| 
 | |
| 		dev_priv->ramin_size = pci_resource_len(dev->pdev, ramin_bar);
 | |
| 		dev_priv->ramin = ioremap(
 | |
| 				pci_resource_start(dev->pdev, ramin_bar),
 | |
| 				dev_priv->ramin_size);
 | |
| 		if (!dev_priv->ramin) {
 | |
| 			NV_ERROR(dev, "Failed to init RAMIN mapping, "
 | |
| 				      "limited instance memory available\n");
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* On older cards (or if the above failed), create a map covering
 | |
| 	 * the BAR0 PRAMIN aperture */
 | |
| 	if (!dev_priv->ramin) {
 | |
| 		dev_priv->ramin_size = 1 * 1024 * 1024;
 | |
| 		dev_priv->ramin = ioremap(mmio_start_offs + NV_RAMIN,
 | |
| 							dev_priv->ramin_size);
 | |
| 		if (!dev_priv->ramin) {
 | |
| 			NV_ERROR(dev, "Failed to map BAR0 PRAMIN.\n");
 | |
| 			return -ENOMEM;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	nouveau_OF_copy_vbios_to_ramin(dev);
 | |
| 
 | |
| 	/* Special flags */
 | |
| 	if (dev->pci_device == 0x01a0)
 | |
| 		dev_priv->flags |= NV_NFORCE;
 | |
| 	else if (dev->pci_device == 0x01f0)
 | |
| 		dev_priv->flags |= NV_NFORCE2;
 | |
| 
 | |
| 	/* For kernel modesetting, init card now and bring up fbcon */
 | |
| 	if (drm_core_check_feature(dev, DRIVER_MODESET)) {
 | |
| 		int ret = nouveau_card_init(dev);
 | |
| 		if (ret)
 | |
| 			return ret;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void nouveau_close(struct drm_device *dev)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 
 | |
| 	/* In the case of an error dev_priv may not be allocated yet */
 | |
| 	if (dev_priv)
 | |
| 		nouveau_card_takedown(dev);
 | |
| }
 | |
| 
 | |
| /* KMS: we need mmio at load time, not when the first drm client opens. */
 | |
| void nouveau_lastclose(struct drm_device *dev)
 | |
| {
 | |
| 	if (drm_core_check_feature(dev, DRIVER_MODESET))
 | |
| 		return;
 | |
| 
 | |
| 	nouveau_close(dev);
 | |
| }
 | |
| 
 | |
| int nouveau_unload(struct drm_device *dev)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 
 | |
| 	if (drm_core_check_feature(dev, DRIVER_MODESET)) {
 | |
| 		if (dev_priv->card_type >= NV_50)
 | |
| 			nv50_display_destroy(dev);
 | |
| 		else
 | |
| 			nv04_display_destroy(dev);
 | |
| 		nouveau_close(dev);
 | |
| 	}
 | |
| 
 | |
| 	iounmap(dev_priv->mmio);
 | |
| 	iounmap(dev_priv->ramin);
 | |
| 
 | |
| 	kfree(dev_priv);
 | |
| 	dev->dev_private = NULL;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int nouveau_ioctl_getparam(struct drm_device *dev, void *data,
 | |
| 						struct drm_file *file_priv)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 	struct drm_nouveau_getparam *getparam = data;
 | |
| 
 | |
| 	NOUVEAU_CHECK_INITIALISED_WITH_RETURN;
 | |
| 
 | |
| 	switch (getparam->param) {
 | |
| 	case NOUVEAU_GETPARAM_CHIPSET_ID:
 | |
| 		getparam->value = dev_priv->chipset;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_PCI_VENDOR:
 | |
| 		getparam->value = dev->pci_vendor;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_PCI_DEVICE:
 | |
| 		getparam->value = dev->pci_device;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_BUS_TYPE:
 | |
| 		if (drm_device_is_agp(dev))
 | |
| 			getparam->value = NV_AGP;
 | |
| 		else if (drm_device_is_pcie(dev))
 | |
| 			getparam->value = NV_PCIE;
 | |
| 		else
 | |
| 			getparam->value = NV_PCI;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_FB_PHYSICAL:
 | |
| 		getparam->value = dev_priv->fb_phys;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_AGP_PHYSICAL:
 | |
| 		getparam->value = dev_priv->gart_info.aper_base;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_PCI_PHYSICAL:
 | |
| 		if (dev->sg) {
 | |
| 			getparam->value = (unsigned long)dev->sg->virtual;
 | |
| 		} else {
 | |
| 			NV_ERROR(dev, "Requested PCIGART address, "
 | |
| 					"while no PCIGART was created\n");
 | |
| 			return -EINVAL;
 | |
| 		}
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_FB_SIZE:
 | |
| 		getparam->value = dev_priv->fb_available_size;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_AGP_SIZE:
 | |
| 		getparam->value = dev_priv->gart_info.aper_size;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_VM_VRAM_BASE:
 | |
| 		getparam->value = dev_priv->vm_vram_base;
 | |
| 		break;
 | |
| 	case NOUVEAU_GETPARAM_GRAPH_UNITS:
 | |
| 		/* NV40 and NV50 versions are quite different, but register
 | |
| 		 * address is the same. User is supposed to know the card
 | |
| 		 * family anyway... */
 | |
| 		if (dev_priv->chipset >= 0x40) {
 | |
| 			getparam->value = nv_rd32(dev, NV40_PMC_GRAPH_UNITS);
 | |
| 			break;
 | |
| 		}
 | |
| 		/* FALLTHRU */
 | |
| 	default:
 | |
| 		NV_ERROR(dev, "unknown parameter %lld\n", getparam->param);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int
 | |
| nouveau_ioctl_setparam(struct drm_device *dev, void *data,
 | |
| 		       struct drm_file *file_priv)
 | |
| {
 | |
| 	struct drm_nouveau_setparam *setparam = data;
 | |
| 
 | |
| 	NOUVEAU_CHECK_INITIALISED_WITH_RETURN;
 | |
| 
 | |
| 	switch (setparam->param) {
 | |
| 	default:
 | |
| 		NV_ERROR(dev, "unknown parameter %lld\n", setparam->param);
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Wait until (value(reg) & mask) == val, up until timeout has hit */
 | |
| bool nouveau_wait_until(struct drm_device *dev, uint64_t timeout,
 | |
| 			uint32_t reg, uint32_t mask, uint32_t val)
 | |
| {
 | |
| 	struct drm_nouveau_private *dev_priv = dev->dev_private;
 | |
| 	struct nouveau_timer_engine *ptimer = &dev_priv->engine.timer;
 | |
| 	uint64_t start = ptimer->read(dev);
 | |
| 
 | |
| 	do {
 | |
| 		if ((nv_rd32(dev, reg) & mask) == val)
 | |
| 			return true;
 | |
| 	} while (ptimer->read(dev) - start < timeout);
 | |
| 
 | |
| 	return false;
 | |
| }
 | |
| 
 | |
| /* Waits for PGRAPH to go completely idle */
 | |
| bool nouveau_wait_for_idle(struct drm_device *dev)
 | |
| {
 | |
| 	if (!nv_wait(NV04_PGRAPH_STATUS, 0xffffffff, 0x00000000)) {
 | |
| 		NV_ERROR(dev, "PGRAPH idle timed out with status 0x%08x\n",
 | |
| 			 nv_rd32(dev, NV04_PGRAPH_STATUS));
 | |
| 		return false;
 | |
| 	}
 | |
| 
 | |
| 	return true;
 | |
| }
 | |
| 
 |