* the searched id is returned
* returns NULL if not found.
*/
-static struct cpupool *__cpupool_find_by_id(int id, bool exact)
+static struct cpupool *__cpupool_find_by_id(unsigned int id, bool exact)
{
struct cpupool **q;
return (!exact || (*q == NULL) || ((*q)->cpupool_id == id)) ? *q : NULL;
}
-static struct cpupool *cpupool_find_by_id(int poolid)
+static struct cpupool *cpupool_find_by_id(unsigned int poolid)
{
return __cpupool_find_by_id(poolid, true);
}
-static struct cpupool *__cpupool_get_by_id(int poolid, bool exact)
+static struct cpupool *__cpupool_get_by_id(unsigned int poolid, bool exact)
{
struct cpupool *c;
spin_lock(&cpupool_lock);
return c;
}
-struct cpupool *cpupool_get_by_id(int poolid)
+struct cpupool *cpupool_get_by_id(unsigned int poolid)
{
return __cpupool_get_by_id(poolid, true);
}
-static struct cpupool *cpupool_get_next_by_id(int poolid)
+static struct cpupool *cpupool_get_next_by_id(unsigned int poolid)
{
return __cpupool_get_by_id(poolid, false);
}
* - unknown scheduler
*/
static struct cpupool *cpupool_create(
- int poolid, unsigned int sched_id, int *perr)
+ unsigned int poolid, unsigned int sched_id, int *perr)
{
struct cpupool *c;
struct cpupool **q;
- int last = 0;
+ unsigned int last = 0;
*perr = -ENOMEM;
if ( (c = alloc_cpupool_struct()) == NULL )
/* One reference for caller, one reference for cpupool_destroy(). */
atomic_set(&c->refcnt, 2);
- debugtrace_printk("cpupool_create(pool=%d,sched=%u)\n", poolid, sched_id);
+ debugtrace_printk("cpupool_create(pool=%u,sched=%u)\n", poolid, sched_id);
spin_lock(&cpupool_lock);
spin_unlock(&cpupool_lock);
- debugtrace_printk("Created cpupool %d with scheduler %s (%s)\n",
+ debugtrace_printk("Created cpupool %u with scheduler %s (%s)\n",
c->cpupool_id, c->sched->name, c->sched->opt_name);
*perr = 0;
cpupool_put(c);
- debugtrace_printk("cpupool_destroy(pool=%d)\n", c->cpupool_id);
+ debugtrace_printk("cpupool_destroy(pool=%u)\n", c->cpupool_id);
return 0;
}
struct cpupool *c = info;
long ret;
- debugtrace_printk("cpupool_unassign_cpu(pool=%d,cpu=%d)\n",
+ debugtrace_printk("cpupool_unassign_cpu(pool=%u,cpu=%d)\n",
cpupool_cpu_moving->cpupool_id, cpupool_moving_cpu);
spin_lock(&cpupool_lock);
int ret;
unsigned int master_cpu;
- debugtrace_printk("cpupool_unassign_cpu(pool=%d,cpu=%d)\n",
+ debugtrace_printk("cpupool_unassign_cpu(pool=%u,cpu=%d)\n",
c->cpupool_id, cpu);
if ( !cpu_online(cpu) )
ret = cpupool_unassign_cpu_start(c, master_cpu);
if ( ret )
{
- debugtrace_printk("cpupool_unassign_cpu(pool=%d,cpu=%d) ret %d\n",
+ debugtrace_printk("cpupool_unassign_cpu(pool=%u,cpu=%d) ret %d\n",
c->cpupool_id, cpu, ret);
return ret;
}
* - pool does not exist
* - no cpu assigned to pool
*/
-int cpupool_add_domain(struct domain *d, int poolid)
+int cpupool_add_domain(struct domain *d, unsigned int poolid)
{
struct cpupool *c;
int rc;
rc = 0;
}
spin_unlock(&cpupool_lock);
- debugtrace_printk("cpupool_add_domain(dom=%d,pool=%d) n_dom %d rc %d\n",
+ debugtrace_printk("cpupool_add_domain(dom=%d,pool=%u) n_dom %d rc %d\n",
d->domain_id, poolid, n_dom, rc);
return rc;
}
*/
void cpupool_rm_domain(struct domain *d)
{
- int cpupool_id;
+ unsigned int cpupool_id;
int n_dom;
if ( d->cpupool == NULL )
n_dom = d->cpupool->n_dom;
d->cpupool = NULL;
spin_unlock(&cpupool_lock);
- debugtrace_printk("cpupool_rm_domain(dom=%d,pool=%d) n_dom %d\n",
+ debugtrace_printk("cpupool_rm_domain(dom=%d,pool=%u) n_dom %d\n",
d->domain_id, cpupool_id, n_dom);
return;
}
case XEN_SYSCTL_CPUPOOL_OP_CREATE:
{
- int poolid;
+ unsigned int poolid;
poolid = (op->cpupool_id == XEN_SYSCTL_CPUPOOL_PAR_ANY) ?
CPUPOOLID_NONE: op->cpupool_id;
const cpumask_t *cpus;
cpu = op->cpu;
- debugtrace_printk("cpupool_assign_cpu(pool=%d,cpu=%d)\n",
+ debugtrace_printk("cpupool_assign_cpu(pool=%u,cpu=%u)\n",
op->cpupool_id, cpu);
spin_lock(&cpupool_lock);
addcpu_out:
spin_unlock(&cpupool_lock);
- debugtrace_printk("cpupool_assign_cpu(pool=%d,cpu=%d) ret %d\n",
+ debugtrace_printk("cpupool_assign_cpu(pool=%u,cpu=%u) ret %d\n",
op->cpupool_id, cpu, ret);
}
rcu_unlock_domain(d);
break;
}
- debugtrace_printk("cpupool move_domain(dom=%d)->pool=%d\n",
+ debugtrace_printk("cpupool move_domain(dom=%d)->pool=%u\n",
d->domain_id, op->cpupool_id);
ret = -ENOENT;
spin_lock(&cpupool_lock);
ret = cpupool_move_domain_locked(d, c);
spin_unlock(&cpupool_lock);
- debugtrace_printk("cpupool move_domain(dom=%d)->pool=%d ret %d\n",
+ debugtrace_printk("cpupool move_domain(dom=%d)->pool=%u ret %d\n",
d->domain_id, op->cpupool_id, ret);
rcu_unlock_domain(d);
}
return ret;
}
-int cpupool_get_id(const struct domain *d)
+unsigned int cpupool_get_id(const struct domain *d)
{
return d->cpupool ? d->cpupool->cpupool_id : CPUPOOLID_NONE;
}
for_each_cpupool(c)
{
- printk("Cpupool %d:\n", (*c)->cpupool_id);
+ printk("Cpupool %u:\n", (*c)->cpupool_id);
printk("Cpus: %*pbl\n", CPUMASK_PR((*c)->cpu_valid));
sched_gran_print((*c)->gran, cpupool_get_granularity(*c));
schedule_dump(*c);