percpu: Resolve ambiguities in __get_cpu_var/cpumask_var_t
__get_cpu_var can paper over differences in the definitions of cpumask_var_t and either use the address of the cpumask variable directly or perform a fetch of the address of the struct cpumask allocated elsewhere. This is important particularly when using per cpu cpumask_var_t declarations because in one case we have an offset into a per cpu area to handle and in the other case we need to fetch a pointer from the offset. This patch introduces a new macro this_cpu_cpumask_var_ptr() that is defined where cpumask_var_t is defined and performs the proper actions. All use cases where __get_cpu_var is used with cpumask_var_t are converted to the use of this_cpu_cpumask_var_ptr(). Signed-off-by: Christoph Lameter <cl@linux.com> Signed-off-by: Tejun Heo <tj@kernel.org>
This commit is contained in:
		
					parent
					
						
							
								23f66e2d66
							
						
					
				
			
			
				commit
				
					
						4ba2968420
					
				
			
		
					 7 changed files with 17 additions and 7 deletions
				
			
		|  | @ -189,7 +189,7 @@ static inline int p4_ht_thread(int cpu) | ||||||
| { | { | ||||||
| #ifdef CONFIG_SMP | #ifdef CONFIG_SMP | ||||||
| 	if (smp_num_siblings == 2) | 	if (smp_num_siblings == 2) | ||||||
| 		return cpu != cpumask_first(__get_cpu_var(cpu_sibling_map)); | 		return cpu != cpumask_first(this_cpu_cpumask_var_ptr(cpu_sibling_map)); | ||||||
| #endif | #endif | ||||||
| 	return 0; | 	return 0; | ||||||
| } | } | ||||||
|  |  | ||||||
|  | @ -42,8 +42,7 @@ __x2apic_send_IPI_mask(const struct cpumask *mask, int vector, int apic_dest) | ||||||
| 	 * We are to modify mask, so we need an own copy | 	 * We are to modify mask, so we need an own copy | ||||||
| 	 * and be sure it's manipulated with irq off. | 	 * and be sure it's manipulated with irq off. | ||||||
| 	 */ | 	 */ | ||||||
| 	ipi_mask_ptr = __raw_get_cpu_var(ipi_mask); | 	ipi_mask_ptr = this_cpu_cpumask_var_ptr(ipi_mask); | ||||||
| 	cpumask_copy(ipi_mask_ptr, mask); |  | ||||||
| 
 | 
 | ||||||
| 	/*
 | 	/*
 | ||||||
| 	 * The idea is to send one IPI per cluster. | 	 * The idea is to send one IPI per cluster. | ||||||
|  |  | ||||||
|  | @ -372,7 +372,7 @@ static unsigned int get_stagger(void) | ||||||
| { | { | ||||||
| #ifdef CONFIG_SMP | #ifdef CONFIG_SMP | ||||||
| 	int cpu = smp_processor_id(); | 	int cpu = smp_processor_id(); | ||||||
| 	return cpu != cpumask_first(__get_cpu_var(cpu_sibling_map)); | 	return cpu != cpumask_first(this_cpu_cpumask_var_ptr(cpu_sibling_map)); | ||||||
| #endif | #endif | ||||||
| 	return 0; | 	return 0; | ||||||
| } | } | ||||||
|  |  | ||||||
|  | @ -666,10 +666,19 @@ static inline size_t cpumask_size(void) | ||||||
|  * |  * | ||||||
|  * This code makes NR_CPUS length memcopy and brings to a memory corruption. |  * This code makes NR_CPUS length memcopy and brings to a memory corruption. | ||||||
|  * cpumask_copy() provide safe copy functionality. |  * cpumask_copy() provide safe copy functionality. | ||||||
|  |  * | ||||||
|  |  * Note that there is another evil here: If you define a cpumask_var_t | ||||||
|  |  * as a percpu variable then the way to obtain the address of the cpumask | ||||||
|  |  * structure differently influences what this_cpu_* operation needs to be | ||||||
|  |  * used. Please use this_cpu_cpumask_var_t in those cases. The direct use | ||||||
|  |  * of this_cpu_ptr() or this_cpu_read() will lead to failures when the | ||||||
|  |  * other type of cpumask_var_t implementation is configured. | ||||||
|  */ |  */ | ||||||
| #ifdef CONFIG_CPUMASK_OFFSTACK | #ifdef CONFIG_CPUMASK_OFFSTACK | ||||||
| typedef struct cpumask *cpumask_var_t; | typedef struct cpumask *cpumask_var_t; | ||||||
| 
 | 
 | ||||||
|  | #define this_cpu_cpumask_var_ptr(x) this_cpu_read(x) | ||||||
|  | 
 | ||||||
| bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node); | bool alloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node); | ||||||
| bool alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags); | bool alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags); | ||||||
| bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node); | bool zalloc_cpumask_var_node(cpumask_var_t *mask, gfp_t flags, int node); | ||||||
|  | @ -681,6 +690,8 @@ void free_bootmem_cpumask_var(cpumask_var_t mask); | ||||||
| #else | #else | ||||||
| typedef struct cpumask cpumask_var_t[1]; | typedef struct cpumask cpumask_var_t[1]; | ||||||
| 
 | 
 | ||||||
|  | #define this_cpu_cpumask_var_ptr(x) this_cpu_ptr(x) | ||||||
|  | 
 | ||||||
| static inline bool alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags) | static inline bool alloc_cpumask_var(cpumask_var_t *mask, gfp_t flags) | ||||||
| { | { | ||||||
| 	return true; | 	return true; | ||||||
|  |  | ||||||
|  | @ -1158,7 +1158,7 @@ static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask_dl); | ||||||
| static int find_later_rq(struct task_struct *task) | static int find_later_rq(struct task_struct *task) | ||||||
| { | { | ||||||
| 	struct sched_domain *sd; | 	struct sched_domain *sd; | ||||||
| 	struct cpumask *later_mask = __get_cpu_var(local_cpu_mask_dl); | 	struct cpumask *later_mask = this_cpu_cpumask_var_ptr(local_cpu_mask_dl); | ||||||
| 	int this_cpu = smp_processor_id(); | 	int this_cpu = smp_processor_id(); | ||||||
| 	int best_cpu, cpu = task_cpu(task); | 	int best_cpu, cpu = task_cpu(task); | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
|  | @ -6539,7 +6539,7 @@ static int load_balance(int this_cpu, struct rq *this_rq, | ||||||
| 	struct sched_group *group; | 	struct sched_group *group; | ||||||
| 	struct rq *busiest; | 	struct rq *busiest; | ||||||
| 	unsigned long flags; | 	unsigned long flags; | ||||||
| 	struct cpumask *cpus = __get_cpu_var(load_balance_mask); | 	struct cpumask *cpus = this_cpu_cpumask_var_ptr(load_balance_mask); | ||||||
| 
 | 
 | ||||||
| 	struct lb_env env = { | 	struct lb_env env = { | ||||||
| 		.sd		= sd, | 		.sd		= sd, | ||||||
|  |  | ||||||
|  | @ -1526,7 +1526,7 @@ static DEFINE_PER_CPU(cpumask_var_t, local_cpu_mask); | ||||||
| static int find_lowest_rq(struct task_struct *task) | static int find_lowest_rq(struct task_struct *task) | ||||||
| { | { | ||||||
| 	struct sched_domain *sd; | 	struct sched_domain *sd; | ||||||
| 	struct cpumask *lowest_mask = __get_cpu_var(local_cpu_mask); | 	struct cpumask *lowest_mask = this_cpu_cpumask_var_ptr(local_cpu_mask); | ||||||
| 	int this_cpu = smp_processor_id(); | 	int this_cpu = smp_processor_id(); | ||||||
| 	int cpu      = task_cpu(task); | 	int cpu      = task_cpu(task); | ||||||
| 
 | 
 | ||||||
|  |  | ||||||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue
	
	 Christoph Lameter
				Christoph Lameter