Lines Matching refs:grp
321 static inline struct cfs_rq *group_cfs_rq(struct sched_entity *grp)
323 return grp->my_q;
506 static inline struct cfs_rq *group_cfs_rq(struct sched_entity *grp)
2558 static inline int get_numa_group(struct numa_group *grp)
2560 return refcount_inc_not_zero(&grp->refcount);
2563 static inline void put_numa_group(struct numa_group *grp)
2565 if (refcount_dec_and_test(&grp->refcount)) {
2566 kfree_rcu(grp, rcu);
2572 struct numa_group *grp, *my_grp;
2581 grp = kzalloc(size, GFP_KERNEL | __GFP_NOWARN);
2582 if (!grp) {
2586 refcount_set(&grp->refcount, 1);
2587 grp->active_nodes = 1;
2588 grp->max_faults_cpu = 0;
2589 spin_lock_init(&grp->lock);
2590 grp->gid = p->pid;
2592 grp->faults_cpu = grp->faults + NR_NUMA_HINT_FAULT_TYPES * nr_node_ids;
2595 grp->faults[i] = p->numa_faults[i];
2598 grp->total_faults = p->total_numa_faults;
2600 grp->nr_tasks++;
2601 rcu_assign_pointer(p->numa_group, grp);
2610 grp = rcu_dereference(tsk->numa_group);
2611 if (!grp) {
2616 if (grp == my_grp) {
2624 if (my_grp->nr_tasks > grp->nr_tasks) {
2629 * Tie-break on the grp address.
2631 if (my_grp->nr_tasks == grp->nr_tasks && my_grp > grp) {
2648 if (join && !get_numa_group(grp)) {
2659 double_lock_irq(&my_grp->lock, &grp->lock);
2663 grp->faults[i] += p->numa_faults[i];
2666 grp->total_faults += p->total_numa_faults;
2669 grp->nr_tasks++;
2672 spin_unlock_irq(&grp->lock);
2674 rcu_assign_pointer(p->numa_group, grp);
2694 struct numa_group *grp = rcu_dereference_raw(p->numa_group);
2703 if (grp) {
2704 spin_lock_irqsave(&grp->lock, flags);
2706 grp->faults[i] -= p->numa_faults[i];
2708 grp->total_faults -= p->total_numa_faults;
2710 grp->nr_tasks--;
2711 spin_unlock_irqrestore(&grp->lock, flags);
2713 put_numa_group(grp);
3284 * grp->load.weight