workqueue: relocate rebind_workers()
rebind_workers() will be reimplemented in a way which makes it mostly decoupled from the rest of worker management. Move rebind_workers() so that it's located with other CPU hotplug related functions. This patch is pure function relocation. Signed-off-by: Tejun Heo <tj@kernel.org> Reviewed-by: Lai Jiangshan <laijs@cn.fujitsu.com>
This commit is contained in:
		
					parent
					
						
							
								822d8405d1
							
						
					
				
			
			
				commit
				
					
						bd7c089eb2
					
				
			
		
					 1 changed files with 71 additions and 71 deletions
				
			
		| 
						 | 
				
			
			@ -1643,77 +1643,6 @@ static void busy_worker_rebind_fn(struct work_struct *work)
 | 
			
		|||
	spin_unlock_irq(&worker->pool->lock);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/**
 | 
			
		||||
 * rebind_workers - rebind all workers of a pool to the associated CPU
 | 
			
		||||
 * @pool: pool of interest
 | 
			
		||||
 *
 | 
			
		||||
 * @pool->cpu is coming online.  Rebind all workers to the CPU.  Rebinding
 | 
			
		||||
 * is different for idle and busy ones.
 | 
			
		||||
 *
 | 
			
		||||
 * Idle ones will be removed from the idle_list and woken up.  They will
 | 
			
		||||
 * add themselves back after completing rebind.  This ensures that the
 | 
			
		||||
 * idle_list doesn't contain any unbound workers when re-bound busy workers
 | 
			
		||||
 * try to perform local wake-ups for concurrency management.
 | 
			
		||||
 *
 | 
			
		||||
 * Busy workers can rebind after they finish their current work items.
 | 
			
		||||
 * Queueing the rebind work item at the head of the scheduled list is
 | 
			
		||||
 * enough.  Note that nr_running will be properly bumped as busy workers
 | 
			
		||||
 * rebind.
 | 
			
		||||
 *
 | 
			
		||||
 * On return, all non-manager workers are scheduled for rebind - see
 | 
			
		||||
 * manage_workers() for the manager special case.  Any idle worker
 | 
			
		||||
 * including the manager will not appear on @idle_list until rebind is
 | 
			
		||||
 * complete, making local wake-ups safe.
 | 
			
		||||
 */
 | 
			
		||||
static void rebind_workers(struct worker_pool *pool)
 | 
			
		||||
{
 | 
			
		||||
	struct worker *worker, *n;
 | 
			
		||||
	int i;
 | 
			
		||||
 | 
			
		||||
	lockdep_assert_held(&pool->manager_mutex);
 | 
			
		||||
	lockdep_assert_held(&pool->lock);
 | 
			
		||||
 | 
			
		||||
	/* dequeue and kick idle ones */
 | 
			
		||||
	list_for_each_entry_safe(worker, n, &pool->idle_list, entry) {
 | 
			
		||||
		/*
 | 
			
		||||
		 * idle workers should be off @pool->idle_list until rebind
 | 
			
		||||
		 * is complete to avoid receiving premature local wake-ups.
 | 
			
		||||
		 */
 | 
			
		||||
		list_del_init(&worker->entry);
 | 
			
		||||
 | 
			
		||||
		/*
 | 
			
		||||
		 * worker_thread() will see the above dequeuing and call
 | 
			
		||||
		 * idle_worker_rebind().
 | 
			
		||||
		 */
 | 
			
		||||
		wake_up_process(worker->task);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	/* rebind busy workers */
 | 
			
		||||
	for_each_busy_worker(worker, i, pool) {
 | 
			
		||||
		struct work_struct *rebind_work = &worker->rebind_work;
 | 
			
		||||
		struct workqueue_struct *wq;
 | 
			
		||||
 | 
			
		||||
		if (test_and_set_bit(WORK_STRUCT_PENDING_BIT,
 | 
			
		||||
				     work_data_bits(rebind_work)))
 | 
			
		||||
			continue;
 | 
			
		||||
 | 
			
		||||
		debug_work_activate(rebind_work);
 | 
			
		||||
 | 
			
		||||
		/*
 | 
			
		||||
		 * wq doesn't really matter but let's keep @worker->pool
 | 
			
		||||
		 * and @pwq->pool consistent for sanity.
 | 
			
		||||
		 */
 | 
			
		||||
		if (worker->pool->attrs->nice < 0)
 | 
			
		||||
			wq = system_highpri_wq;
 | 
			
		||||
		else
 | 
			
		||||
			wq = system_wq;
 | 
			
		||||
 | 
			
		||||
		insert_work(per_cpu_ptr(wq->cpu_pwqs, pool->cpu), rebind_work,
 | 
			
		||||
			    worker->scheduled.next,
 | 
			
		||||
			    work_color_to_flags(WORK_NO_COLOR));
 | 
			
		||||
	}
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
static struct worker *alloc_worker(void)
 | 
			
		||||
{
 | 
			
		||||
	struct worker *worker;
 | 
			
		||||
| 
						 | 
				
			
			@ -4196,6 +4125,77 @@ static void wq_unbind_fn(struct work_struct *work)
 | 
			
		|||
		atomic_set(&pool->nr_running, 0);
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/**
 | 
			
		||||
 * rebind_workers - rebind all workers of a pool to the associated CPU
 | 
			
		||||
 * @pool: pool of interest
 | 
			
		||||
 *
 | 
			
		||||
 * @pool->cpu is coming online.  Rebind all workers to the CPU.  Rebinding
 | 
			
		||||
 * is different for idle and busy ones.
 | 
			
		||||
 *
 | 
			
		||||
 * Idle ones will be removed from the idle_list and woken up.  They will
 | 
			
		||||
 * add themselves back after completing rebind.  This ensures that the
 | 
			
		||||
 * idle_list doesn't contain any unbound workers when re-bound busy workers
 | 
			
		||||
 * try to perform local wake-ups for concurrency management.
 | 
			
		||||
 *
 | 
			
		||||
 * Busy workers can rebind after they finish their current work items.
 | 
			
		||||
 * Queueing the rebind work item at the head of the scheduled list is
 | 
			
		||||
 * enough.  Note that nr_running will be properly bumped as busy workers
 | 
			
		||||
 * rebind.
 | 
			
		||||
 *
 | 
			
		||||
 * On return, all non-manager workers are scheduled for rebind - see
 | 
			
		||||
 * manage_workers() for the manager special case.  Any idle worker
 | 
			
		||||
 * including the manager will not appear on @idle_list until rebind is
 | 
			
		||||
 * complete, making local wake-ups safe.
 | 
			
		||||
 */
 | 
			
		||||
static void rebind_workers(struct worker_pool *pool)
 | 
			
		||||
{
 | 
			
		||||
	struct worker *worker, *n;
 | 
			
		||||
	int i;
 | 
			
		||||
 | 
			
		||||
	lockdep_assert_held(&pool->manager_mutex);
 | 
			
		||||
	lockdep_assert_held(&pool->lock);
 | 
			
		||||
 | 
			
		||||
	/* dequeue and kick idle ones */
 | 
			
		||||
	list_for_each_entry_safe(worker, n, &pool->idle_list, entry) {
 | 
			
		||||
		/*
 | 
			
		||||
		 * idle workers should be off @pool->idle_list until rebind
 | 
			
		||||
		 * is complete to avoid receiving premature local wake-ups.
 | 
			
		||||
		 */
 | 
			
		||||
		list_del_init(&worker->entry);
 | 
			
		||||
 | 
			
		||||
		/*
 | 
			
		||||
		 * worker_thread() will see the above dequeuing and call
 | 
			
		||||
		 * idle_worker_rebind().
 | 
			
		||||
		 */
 | 
			
		||||
		wake_up_process(worker->task);
 | 
			
		||||
	}
 | 
			
		||||
 | 
			
		||||
	/* rebind busy workers */
 | 
			
		||||
	for_each_busy_worker(worker, i, pool) {
 | 
			
		||||
		struct work_struct *rebind_work = &worker->rebind_work;
 | 
			
		||||
		struct workqueue_struct *wq;
 | 
			
		||||
 | 
			
		||||
		if (test_and_set_bit(WORK_STRUCT_PENDING_BIT,
 | 
			
		||||
				     work_data_bits(rebind_work)))
 | 
			
		||||
			continue;
 | 
			
		||||
 | 
			
		||||
		debug_work_activate(rebind_work);
 | 
			
		||||
 | 
			
		||||
		/*
 | 
			
		||||
		 * wq doesn't really matter but let's keep @worker->pool
 | 
			
		||||
		 * and @pwq->pool consistent for sanity.
 | 
			
		||||
		 */
 | 
			
		||||
		if (worker->pool->attrs->nice < 0)
 | 
			
		||||
			wq = system_highpri_wq;
 | 
			
		||||
		else
 | 
			
		||||
			wq = system_wq;
 | 
			
		||||
 | 
			
		||||
		insert_work(per_cpu_ptr(wq->cpu_pwqs, pool->cpu), rebind_work,
 | 
			
		||||
			    worker->scheduled.next,
 | 
			
		||||
			    work_color_to_flags(WORK_NO_COLOR));
 | 
			
		||||
	}
 | 
			
		||||
}
 | 
			
		||||
 | 
			
		||||
/*
 | 
			
		||||
 * Workqueues should be brought up before normal priority CPU notifiers.
 | 
			
		||||
 * This will be registered high priority CPU notifier.
 | 
			
		||||
| 
						 | 
				
			
			
 | 
			
		|||
		Loading…
	
	Add table
		Add a link
		
	
		Reference in a new issue