blob: 69272620a6b75e3a46fe644e472009b2cc8c7140 [file] [log] [blame]
Arnd Bergmann8b3d6662005-11-15 15:53:52 -05001/* sched.c - SPU scheduler.
2 *
3 * Copyright (C) IBM 2005
4 * Author: Mark Nutter <mnutter@us.ibm.com>
5 *
Mark Nuttera68cf982006-10-04 17:26:12 +02006 * 2006-03-31 NUMA domains added.
Arnd Bergmann8b3d6662005-11-15 15:53:52 -05007 *
8 * This program is free software; you can redistribute it and/or modify
9 * it under the terms of the GNU General Public License as published by
10 * the Free Software Foundation; either version 2, or (at your option)
11 * any later version.
12 *
13 * This program is distributed in the hope that it will be useful,
14 * but WITHOUT ANY WARRANTY; without even the implied warranty of
15 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 * GNU General Public License for more details.
17 *
18 * You should have received a copy of the GNU General Public License
19 * along with this program; if not, write to the Free Software
20 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
21 */
22
Arnd Bergmann3b3d22c2005-12-05 22:52:24 -050023#undef DEBUG
24
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050025#include <linux/module.h>
26#include <linux/errno.h>
27#include <linux/sched.h>
28#include <linux/kernel.h>
29#include <linux/mm.h>
30#include <linux/completion.h>
31#include <linux/vmalloc.h>
32#include <linux/smp.h>
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050033#include <linux/stddef.h>
34#include <linux/unistd.h>
Mark Nuttera68cf982006-10-04 17:26:12 +020035#include <linux/numa.h>
36#include <linux/mutex.h>
Arnd Bergmann86767272006-10-04 17:26:21 +020037#include <linux/notifier.h>
Christoph Hellwig37901802007-06-29 10:57:51 +100038#include <linux/kthread.h>
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050039
40#include <asm/io.h>
41#include <asm/mmu_context.h>
42#include <asm/spu.h>
43#include <asm/spu_csa.h>
Geoff Levanda91942a2006-06-19 20:33:30 +020044#include <asm/spu_priv1.h>
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050045#include "spufs.h"
46
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050047struct spu_prio_array {
Christoph Hellwig72cb3602007-02-13 21:54:28 +010048 DECLARE_BITMAP(bitmap, MAX_PRIO);
Christoph Hellwig079cdb62007-02-13 21:54:23 +010049 struct list_head runq[MAX_PRIO];
50 spinlock_t runq_lock;
Mark Nuttera68cf982006-10-04 17:26:12 +020051 struct list_head active_list[MAX_NUMNODES];
52 struct mutex active_mutex[MAX_NUMNODES];
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050053};
54
Mark Nuttera68cf982006-10-04 17:26:12 +020055static struct spu_prio_array *spu_prio;
Christoph Hellwig37901802007-06-29 10:57:51 +100056static struct task_struct *spusched_task;
57static struct timer_list spusched_timer;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -050058
Christoph Hellwigfe443ef2007-06-29 10:57:52 +100059/*
60 * Priority of a normal, non-rt, non-niced'd process (aka nice level 0).
61 */
62#define NORMAL_PRIO 120
63
64/*
65 * Frequency of the spu scheduler tick. By default we do one SPU scheduler
66 * tick for every 10 CPU scheduler ticks.
67 */
68#define SPUSCHED_TICK (10)
69
70/*
71 * These are the 'tuning knobs' of the scheduler:
72 *
Jeremy Kerr60e24232007-06-29 10:57:53 +100073 * Minimum timeslice is 5 msecs (or 1 spu scheduler tick, whichever is
74 * larger), default timeslice is 100 msecs, maximum timeslice is 800 msecs.
Christoph Hellwigfe443ef2007-06-29 10:57:52 +100075 */
Jeremy Kerr60e24232007-06-29 10:57:53 +100076#define MIN_SPU_TIMESLICE max(5 * HZ / (1000 * SPUSCHED_TICK), 1)
77#define DEF_SPU_TIMESLICE (100 * HZ / (1000 * SPUSCHED_TICK))
Christoph Hellwigfe443ef2007-06-29 10:57:52 +100078
79#define MAX_USER_PRIO (MAX_PRIO - MAX_RT_PRIO)
80#define SCALE_PRIO(x, prio) \
81 max(x * (MAX_PRIO - prio) / (MAX_USER_PRIO / 2), MIN_SPU_TIMESLICE)
82
83/*
84 * scale user-nice values [ -20 ... 0 ... 19 ] to time slice values:
85 * [800ms ... 100ms ... 5ms]
86 *
87 * The higher a thread's priority, the bigger timeslices
88 * it gets during one round of execution. But even the lowest
89 * priority thread gets MIN_TIMESLICE worth of execution time.
90 */
91void spu_set_timeslice(struct spu_context *ctx)
92{
93 if (ctx->prio < NORMAL_PRIO)
94 ctx->time_slice = SCALE_PRIO(DEF_SPU_TIMESLICE * 4, ctx->prio);
95 else
96 ctx->time_slice = SCALE_PRIO(DEF_SPU_TIMESLICE, ctx->prio);
97}
98
Christoph Hellwig2cf2b3b2007-06-29 10:57:55 +100099/*
100 * Update scheduling information from the owning thread.
101 */
102void __spu_update_sched_info(struct spu_context *ctx)
103{
104 /*
105 * We do our own priority calculations, so we normally want
106 * ->static_prio to start with. Unfortunately thies field
107 * contains junk for threads with a realtime scheduling
108 * policy so we have to look at ->prio in this case.
109 */
110 if (rt_prio(current->prio))
111 ctx->prio = current->prio;
112 else
113 ctx->prio = current->static_prio;
114 ctx->policy = current->policy;
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000115
116 /*
117 * A lot of places that don't hold active_mutex poke into
118 * cpus_allowed, including grab_runnable_context which
119 * already holds the runq_lock. So abuse runq_lock
120 * to protect this field aswell.
121 */
122 spin_lock(&spu_prio->runq_lock);
123 ctx->cpus_allowed = current->cpus_allowed;
124 spin_unlock(&spu_prio->runq_lock);
Christoph Hellwig2cf2b3b2007-06-29 10:57:55 +1000125}
126
127void spu_update_sched_info(struct spu_context *ctx)
128{
129 int node = ctx->spu->node;
130
131 mutex_lock(&spu_prio->active_mutex[node]);
132 __spu_update_sched_info(ctx);
133 mutex_unlock(&spu_prio->active_mutex[node]);
134}
135
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000136static int __node_allowed(struct spu_context *ctx, int node)
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500137{
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000138 if (nr_cpus_node(node)) {
139 cpumask_t mask = node_to_cpumask(node);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500140
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000141 if (cpus_intersects(mask, ctx->cpus_allowed))
142 return 1;
143 }
144
145 return 0;
146}
147
148static int node_allowed(struct spu_context *ctx, int node)
149{
150 int rval;
151
152 spin_lock(&spu_prio->runq_lock);
153 rval = __node_allowed(ctx, node);
154 spin_unlock(&spu_prio->runq_lock);
155
156 return rval;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500157}
158
Christoph Hellwig202557d2007-02-13 21:36:49 +0100159/**
160 * spu_add_to_active_list - add spu to active list
161 * @spu: spu to add to the active list
162 */
163static void spu_add_to_active_list(struct spu *spu)
164{
165 mutex_lock(&spu_prio->active_mutex[spu->node]);
166 list_add_tail(&spu->list, &spu_prio->active_list[spu->node]);
167 mutex_unlock(&spu_prio->active_mutex[spu->node]);
168}
169
Christoph Hellwig37901802007-06-29 10:57:51 +1000170static void __spu_remove_from_active_list(struct spu *spu)
171{
172 list_del_init(&spu->list);
173}
174
Christoph Hellwig202557d2007-02-13 21:36:49 +0100175/**
176 * spu_remove_from_active_list - remove spu from active list
177 * @spu: spu to remove from the active list
Christoph Hellwig202557d2007-02-13 21:36:49 +0100178 */
Christoph Hellwig678b2ff2007-02-13 21:54:25 +0100179static void spu_remove_from_active_list(struct spu *spu)
Christoph Hellwig202557d2007-02-13 21:36:49 +0100180{
181 int node = spu->node;
Christoph Hellwig202557d2007-02-13 21:36:49 +0100182
183 mutex_lock(&spu_prio->active_mutex[node]);
Christoph Hellwig37901802007-06-29 10:57:51 +1000184 __spu_remove_from_active_list(spu);
Christoph Hellwig202557d2007-02-13 21:36:49 +0100185 mutex_unlock(&spu_prio->active_mutex[node]);
Christoph Hellwig202557d2007-02-13 21:36:49 +0100186}
187
Arnd Bergmann86767272006-10-04 17:26:21 +0200188static BLOCKING_NOTIFIER_HEAD(spu_switch_notifier);
189
190static void spu_switch_notify(struct spu *spu, struct spu_context *ctx)
191{
192 blocking_notifier_call_chain(&spu_switch_notifier,
193 ctx ? ctx->object_id : 0, spu);
194}
195
196int spu_switch_event_register(struct notifier_block * n)
197{
198 return blocking_notifier_chain_register(&spu_switch_notifier, n);
199}
200
201int spu_switch_event_unregister(struct notifier_block * n)
202{
203 return blocking_notifier_chain_unregister(&spu_switch_notifier, n);
204}
205
Christoph Hellwig202557d2007-02-13 21:36:49 +0100206/**
207 * spu_bind_context - bind spu context to physical spu
208 * @spu: physical spu to bind to
209 * @ctx: context to bind
210 */
211static void spu_bind_context(struct spu *spu, struct spu_context *ctx)
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500212{
Mark Nuttera68cf982006-10-04 17:26:12 +0200213 pr_debug("%s: pid=%d SPU=%d NODE=%d\n", __FUNCTION__, current->pid,
214 spu->number, spu->node);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500215 spu->ctx = ctx;
216 spu->flags = 0;
217 ctx->spu = spu;
218 ctx->ops = &spu_hw_ops;
219 spu->pid = current->pid;
Benjamin Herrenschmidt94b2a432007-03-10 00:05:37 +0100220 spu_associate_mm(spu, ctx->owner);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500221 spu->ibox_callback = spufs_ibox_callback;
222 spu->wbox_callback = spufs_wbox_callback;
Arnd Bergmann51104592005-12-05 22:52:25 -0500223 spu->stop_callback = spufs_stop_callback;
Arnd Bergmanna33a7d72006-03-23 00:00:11 +0100224 spu->mfc_callback = spufs_mfc_callback;
Arnd Bergmann9add11d2006-10-04 17:26:14 +0200225 spu->dma_callback = spufs_dma_callback;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500226 mb();
Arnd Bergmann51104592005-12-05 22:52:25 -0500227 spu_unmap_mappings(ctx);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500228 spu_restore(&ctx->csa, spu);
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500229 spu->timestamp = jiffies;
Mark Nuttera68cf982006-10-04 17:26:12 +0200230 spu_cpu_affinity_set(spu, raw_smp_processor_id());
Arnd Bergmann86767272006-10-04 17:26:21 +0200231 spu_switch_notify(spu, ctx);
Christoph Hellwig81998ba2007-02-13 21:36:48 +0100232 ctx->state = SPU_STATE_RUNNABLE;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500233}
234
Christoph Hellwig202557d2007-02-13 21:36:49 +0100235/**
236 * spu_unbind_context - unbind spu context from physical spu
237 * @spu: physical spu to unbind from
238 * @ctx: context to unbind
Christoph Hellwig202557d2007-02-13 21:36:49 +0100239 */
Christoph Hellwig678b2ff2007-02-13 21:54:25 +0100240static void spu_unbind_context(struct spu *spu, struct spu_context *ctx)
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500241{
Mark Nuttera68cf982006-10-04 17:26:12 +0200242 pr_debug("%s: unbind pid=%d SPU=%d NODE=%d\n", __FUNCTION__,
243 spu->pid, spu->number, spu->node);
Christoph Hellwig202557d2007-02-13 21:36:49 +0100244
Arnd Bergmann86767272006-10-04 17:26:21 +0200245 spu_switch_notify(spu, NULL);
Arnd Bergmann51104592005-12-05 22:52:25 -0500246 spu_unmap_mappings(ctx);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500247 spu_save(&ctx->csa, spu);
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500248 spu->timestamp = jiffies;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500249 ctx->state = SPU_STATE_SAVED;
250 spu->ibox_callback = NULL;
251 spu->wbox_callback = NULL;
Arnd Bergmann51104592005-12-05 22:52:25 -0500252 spu->stop_callback = NULL;
Arnd Bergmanna33a7d72006-03-23 00:00:11 +0100253 spu->mfc_callback = NULL;
Arnd Bergmann9add11d2006-10-04 17:26:14 +0200254 spu->dma_callback = NULL;
Benjamin Herrenschmidt94b2a432007-03-10 00:05:37 +0100255 spu_associate_mm(spu, NULL);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500256 spu->pid = 0;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500257 ctx->ops = &spu_backing_ops;
258 ctx->spu = NULL;
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500259 spu->flags = 0;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500260 spu->ctx = NULL;
261}
262
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100263/**
264 * spu_add_to_rq - add a context to the runqueue
265 * @ctx: context to add
266 */
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200267static void __spu_add_to_rq(struct spu_context *ctx)
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500268{
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200269 int prio = ctx->prio;
270
271 list_add_tail(&ctx->rq, &spu_prio->runq[prio]);
272 set_bit(prio, spu_prio->bitmap);
Mark Nuttera68cf982006-10-04 17:26:12 +0200273}
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500274
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200275static void __spu_del_from_rq(struct spu_context *ctx)
Christoph Hellwiga475c2f2007-04-23 21:08:11 +0200276{
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200277 int prio = ctx->prio;
278
Christoph Hellwiga475c2f2007-04-23 21:08:11 +0200279 if (!list_empty(&ctx->rq))
280 list_del_init(&ctx->rq);
281 if (list_empty(&spu_prio->runq[prio]))
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200282 clear_bit(prio, spu_prio->bitmap);
Mark Nuttera68cf982006-10-04 17:26:12 +0200283}
284
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100285static void spu_prio_wait(struct spu_context *ctx)
286{
Mark Nuttera68cf982006-10-04 17:26:12 +0200287 DEFINE_WAIT(wait);
288
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200289 spin_lock(&spu_prio->runq_lock);
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100290 prepare_to_wait_exclusive(&ctx->stop_wq, &wait, TASK_INTERRUPTIBLE);
Mark Nuttera68cf982006-10-04 17:26:12 +0200291 if (!signal_pending(current)) {
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200292 __spu_add_to_rq(ctx);
293 spin_unlock(&spu_prio->runq_lock);
Christoph Hellwig650f8b02007-02-13 21:36:50 +0100294 mutex_unlock(&ctx->state_mutex);
Mark Nuttera68cf982006-10-04 17:26:12 +0200295 schedule();
Christoph Hellwig650f8b02007-02-13 21:36:50 +0100296 mutex_lock(&ctx->state_mutex);
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200297 spin_lock(&spu_prio->runq_lock);
298 __spu_del_from_rq(ctx);
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500299 }
Luke Browning4e0f4ed2007-04-23 21:08:13 +0200300 spin_unlock(&spu_prio->runq_lock);
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100301 __set_current_state(TASK_RUNNING);
302 remove_wait_queue(&ctx->stop_wq, &wait);
Arnd Bergmann2a911f02005-12-05 22:52:26 -0500303}
304
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100305static struct spu *spu_get_idle(struct spu_context *ctx)
Mark Nuttera68cf982006-10-04 17:26:12 +0200306{
307 struct spu *spu = NULL;
308 int node = cpu_to_node(raw_smp_processor_id());
309 int n;
310
311 for (n = 0; n < MAX_NUMNODES; n++, node++) {
312 node = (node < MAX_NUMNODES) ? node : 0;
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000313 if (!node_allowed(ctx, node))
Mark Nuttera68cf982006-10-04 17:26:12 +0200314 continue;
315 spu = spu_alloc_node(node);
316 if (spu)
317 break;
318 }
319 return spu;
320}
321
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100322/**
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100323 * find_victim - find a lower priority context to preempt
324 * @ctx: canidate context for running
325 *
326 * Returns the freed physical spu to run the new context on.
327 */
328static struct spu *find_victim(struct spu_context *ctx)
329{
330 struct spu_context *victim = NULL;
331 struct spu *spu;
332 int node, n;
333
334 /*
335 * Look for a possible preemption candidate on the local node first.
336 * If there is no candidate look at the other nodes. This isn't
337 * exactly fair, but so far the whole spu schedule tries to keep
338 * a strong node affinity. We might want to fine-tune this in
339 * the future.
340 */
341 restart:
342 node = cpu_to_node(raw_smp_processor_id());
343 for (n = 0; n < MAX_NUMNODES; n++, node++) {
344 node = (node < MAX_NUMNODES) ? node : 0;
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000345 if (!node_allowed(ctx, node))
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100346 continue;
347
348 mutex_lock(&spu_prio->active_mutex[node]);
349 list_for_each_entry(spu, &spu_prio->active_list[node], list) {
350 struct spu_context *tmp = spu->ctx;
351
Christoph Hellwigfe443ef2007-06-29 10:57:52 +1000352 if (tmp->prio > ctx->prio &&
353 (!victim || tmp->prio > victim->prio))
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100354 victim = spu->ctx;
355 }
356 mutex_unlock(&spu_prio->active_mutex[node]);
357
358 if (victim) {
359 /*
360 * This nests ctx->state_mutex, but we always lock
361 * higher priority contexts before lower priority
362 * ones, so this is safe until we introduce
363 * priority inheritance schemes.
364 */
365 if (!mutex_trylock(&victim->state_mutex)) {
366 victim = NULL;
367 goto restart;
368 }
369
370 spu = victim->spu;
371 if (!spu) {
372 /*
373 * This race can happen because we've dropped
374 * the active list mutex. No a problem, just
375 * restart the search.
376 */
377 mutex_unlock(&victim->state_mutex);
378 victim = NULL;
379 goto restart;
380 }
Christoph Hellwig37901802007-06-29 10:57:51 +1000381 spu_remove_from_active_list(spu);
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100382 spu_unbind_context(spu, victim);
383 mutex_unlock(&victim->state_mutex);
Christoph Hellwige097b512007-04-23 21:08:09 +0200384 /*
385 * We need to break out of the wait loop in spu_run
386 * manually to ensure this context gets put on the
387 * runqueue again ASAP.
388 */
389 wake_up(&victim->stop_wq);
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100390 return spu;
391 }
392 }
393
394 return NULL;
395}
396
397/**
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100398 * spu_activate - find a free spu for a context and execute it
399 * @ctx: spu context to schedule
400 * @flags: flags (currently ignored)
401 *
Christoph Hellwig08873092007-04-23 21:08:06 +0200402 * Tries to find a free spu to run @ctx. If no free spu is available
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100403 * add the context to the runqueue so it gets woken up once an spu
404 * is available.
405 */
Christoph Hellwig26bec672007-02-13 21:54:24 +0100406int spu_activate(struct spu_context *ctx, unsigned long flags)
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500407{
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500408
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100409 if (ctx->spu)
410 return 0;
411
412 do {
413 struct spu *spu;
414
415 spu = spu_get_idle(ctx);
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100416 /*
417 * If this is a realtime thread we try to get it running by
418 * preempting a lower priority thread.
419 */
Christoph Hellwigfe443ef2007-06-29 10:57:52 +1000420 if (!spu && rt_prio(ctx->prio))
Christoph Hellwig52f04fc2007-02-13 21:54:27 +0100421 spu = find_victim(ctx);
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100422 if (spu) {
Christoph Hellwig202557d2007-02-13 21:36:49 +0100423 spu_bind_context(spu, ctx);
Christoph Hellwig37901802007-06-29 10:57:51 +1000424 spu_add_to_active_list(spu);
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100425 return 0;
Mark Nuttera68cf982006-10-04 17:26:12 +0200426 }
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100427
Christoph Hellwig50b520d2007-03-10 00:05:36 +0100428 spu_prio_wait(ctx);
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100429 } while (!signal_pending(current));
430
431 return -ERESTARTSYS;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500432}
433
Christoph Hellwig678b2ff2007-02-13 21:54:25 +0100434/**
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000435 * grab_runnable_context - try to find a runnable context
436 *
437 * Remove the highest priority context on the runqueue and return it
438 * to the caller. Returns %NULL if no runnable context was found.
439 */
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000440static struct spu_context *grab_runnable_context(int prio, int node)
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000441{
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000442 struct spu_context *ctx;
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000443 int best;
444
445 spin_lock(&spu_prio->runq_lock);
446 best = sched_find_first_bit(spu_prio->bitmap);
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000447 while (best < prio) {
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000448 struct list_head *rq = &spu_prio->runq[best];
449
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000450 list_for_each_entry(ctx, rq, rq) {
451 /* XXX(hch): check for affinity here aswell */
452 if (__node_allowed(ctx, node)) {
453 __spu_del_from_rq(ctx);
454 goto found;
455 }
456 }
457 best++;
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000458 }
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000459 ctx = NULL;
460 found:
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000461 spin_unlock(&spu_prio->runq_lock);
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000462 return ctx;
463}
464
465static int __spu_deactivate(struct spu_context *ctx, int force, int max_prio)
466{
467 struct spu *spu = ctx->spu;
468 struct spu_context *new = NULL;
469
470 if (spu) {
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000471 new = grab_runnable_context(max_prio, spu->node);
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000472 if (new || force) {
Christoph Hellwig37901802007-06-29 10:57:51 +1000473 spu_remove_from_active_list(spu);
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000474 spu_unbind_context(spu, ctx);
475 spu_free(spu);
476 if (new)
477 wake_up(&new->stop_wq);
478 }
479
480 }
481
482 return new != NULL;
483}
484
485/**
Christoph Hellwig678b2ff2007-02-13 21:54:25 +0100486 * spu_deactivate - unbind a context from it's physical spu
487 * @ctx: spu context to unbind
488 *
489 * Unbind @ctx from the physical spu it is running on and schedule
490 * the highest priority context to run on the freed physical spu.
491 */
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500492void spu_deactivate(struct spu_context *ctx)
493{
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000494 __spu_deactivate(ctx, 1, MAX_PRIO);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500495}
496
Christoph Hellwigae7b4c52007-02-13 21:54:26 +0100497/**
498 * spu_yield - yield a physical spu if others are waiting
499 * @ctx: spu context to yield
500 *
501 * Check if there is a higher priority context waiting and if yes
502 * unbind @ctx from the physical spu and schedule the highest
503 * priority context to run on the freed physical spu instead.
504 */
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500505void spu_yield(struct spu_context *ctx)
506{
Christoph Hellwige5c0b9e2007-06-05 11:25:59 +1000507 if (!(ctx->flags & SPU_CREATE_NOSCHED)) {
508 mutex_lock(&ctx->state_mutex);
509 __spu_deactivate(ctx, 0, MAX_PRIO);
510 mutex_unlock(&ctx->state_mutex);
511 }
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000512}
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500513
Christoph Hellwig37901802007-06-29 10:57:51 +1000514static void spusched_tick(struct spu_context *ctx)
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000515{
Christoph Hellwigfe443ef2007-06-29 10:57:52 +1000516 if (ctx->policy == SCHED_FIFO || --ctx->time_slice)
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000517 return;
518
Christoph Hellwig37901802007-06-29 10:57:51 +1000519 /*
520 * Unfortunately active_mutex ranks outside of state_mutex, so
521 * we have to trylock here. If we fail give the context another
522 * tick and try again.
523 */
524 if (mutex_trylock(&ctx->state_mutex)) {
Christoph Hellwigea1ae592007-06-29 10:57:56 +1000525 struct spu *spu = ctx->spu;
526 struct spu_context *new;
527
528 new = grab_runnable_context(ctx->prio + 1, spu->node);
Christoph Hellwig37901802007-06-29 10:57:51 +1000529 if (new) {
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000530
Christoph Hellwig37901802007-06-29 10:57:51 +1000531 __spu_remove_from_active_list(spu);
532 spu_unbind_context(spu, ctx);
533 spu_free(spu);
534 wake_up(&new->stop_wq);
535 /*
536 * We need to break out of the wait loop in
537 * spu_run manually to ensure this context
538 * gets put on the runqueue again ASAP.
539 */
540 wake_up(&ctx->stop_wq);
541 }
Christoph Hellwigfe443ef2007-06-29 10:57:52 +1000542 spu_set_timeslice(ctx);
Christoph Hellwig37901802007-06-29 10:57:51 +1000543 mutex_unlock(&ctx->state_mutex);
Christoph Hellwigbb5db292007-06-04 23:26:51 +1000544 } else {
Christoph Hellwig37901802007-06-29 10:57:51 +1000545 ctx->time_slice++;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500546 }
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500547}
548
Christoph Hellwig37901802007-06-29 10:57:51 +1000549static void spusched_wake(unsigned long data)
550{
551 mod_timer(&spusched_timer, jiffies + SPUSCHED_TICK);
552 wake_up_process(spusched_task);
553}
554
555static int spusched_thread(void *unused)
556{
557 struct spu *spu, *next;
558 int node;
559
560 setup_timer(&spusched_timer, spusched_wake, 0);
561 __mod_timer(&spusched_timer, jiffies + SPUSCHED_TICK);
562
563 while (!kthread_should_stop()) {
564 set_current_state(TASK_INTERRUPTIBLE);
565 schedule();
566 for (node = 0; node < MAX_NUMNODES; node++) {
567 mutex_lock(&spu_prio->active_mutex[node]);
568 list_for_each_entry_safe(spu, next,
569 &spu_prio->active_list[node],
570 list)
571 spusched_tick(spu->ctx);
572 mutex_unlock(&spu_prio->active_mutex[node]);
573 }
574 }
575
576 del_timer_sync(&spusched_timer);
577 return 0;
578}
579
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500580int __init spu_sched_init(void)
581{
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500582 int i;
583
Mark Nuttera68cf982006-10-04 17:26:12 +0200584 spu_prio = kzalloc(sizeof(struct spu_prio_array), GFP_KERNEL);
Christoph Hellwig37901802007-06-29 10:57:51 +1000585 if (!spu_prio)
586 return -ENOMEM;
587
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500588 for (i = 0; i < MAX_PRIO; i++) {
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100589 INIT_LIST_HEAD(&spu_prio->runq[i]);
Mark Nuttera68cf982006-10-04 17:26:12 +0200590 __clear_bit(i, spu_prio->bitmap);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500591 }
Mark Nuttera68cf982006-10-04 17:26:12 +0200592 __set_bit(MAX_PRIO, spu_prio->bitmap);
593 for (i = 0; i < MAX_NUMNODES; i++) {
594 mutex_init(&spu_prio->active_mutex[i]);
595 INIT_LIST_HEAD(&spu_prio->active_list[i]);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500596 }
Christoph Hellwig079cdb62007-02-13 21:54:23 +0100597 spin_lock_init(&spu_prio->runq_lock);
Christoph Hellwig37901802007-06-29 10:57:51 +1000598
599 spusched_task = kthread_run(spusched_thread, NULL, "spusched");
600 if (IS_ERR(spusched_task)) {
601 kfree(spu_prio);
602 return PTR_ERR(spusched_task);
603 }
Jeremy Kerrf3f59be2007-06-29 10:57:54 +1000604
605 pr_debug("spusched: tick: %d, min ticks: %d, default ticks: %d\n",
606 SPUSCHED_TICK, MIN_SPU_TIMESLICE, DEF_SPU_TIMESLICE);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500607 return 0;
Christoph Hellwig37901802007-06-29 10:57:51 +1000608
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500609}
610
611void __exit spu_sched_exit(void)
612{
Mark Nuttera68cf982006-10-04 17:26:12 +0200613 struct spu *spu, *tmp;
614 int node;
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500615
Christoph Hellwig37901802007-06-29 10:57:51 +1000616 kthread_stop(spusched_task);
617
Mark Nuttera68cf982006-10-04 17:26:12 +0200618 for (node = 0; node < MAX_NUMNODES; node++) {
619 mutex_lock(&spu_prio->active_mutex[node]);
620 list_for_each_entry_safe(spu, tmp, &spu_prio->active_list[node],
621 list) {
622 list_del_init(&spu->list);
623 spu_free(spu);
624 }
625 mutex_unlock(&spu_prio->active_mutex[node]);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500626 }
Mark Nuttera68cf982006-10-04 17:26:12 +0200627 kfree(spu_prio);
Arnd Bergmann8b3d6662005-11-15 15:53:52 -0500628}