Lines Matching refs:list

53  * - the task that performs a successful semop() scans the list of all
67 * - There are two lists of the pending operations: a per-array list
68 * and per-semaphore list (stored in the array). This allows to achieve FIFO
131 struct list_head list; /* queue of pending operations */
143 /* Each task has a list of undo requests. They are executed automatically
147 struct list_head list_proc; /* per-process list: *
152 struct list_head list_id; /* per semaphore array list:
159 /* sem_undo_list controls shared access to the list of sem_undo structures
293 list_for_each_entry_safe(q, tq, &sma->pending_alter, list) {
297 list_add_tail(&q->list, &curr->pending_alter);
799 list_del(&q->list);
864 list_for_each_entry_safe(q, tmp, pending_list, list) {
959 list_for_each_entry_safe(q, tmp, pending_list, list) {
962 /* If we are scanning the single sop, per-semaphore list of
1118 list_for_each_entry(q, l, list) {
1119 /* all task on a per-semaphore list sleep on exactly
1126 list_for_each_entry(q, &sma->pending_alter, list) {
1130 list_for_each_entry(q, &sma->pending_const, list) {
1161 list_for_each_entry_safe(q, tq, &sma->pending_const, list) {
1166 list_for_each_entry_safe(q, tq, &sma->pending_alter, list) {
1172 list_for_each_entry_safe(q, tq, &sem->pending_const, list) {
1176 list_for_each_entry_safe(q, tq, &sem->pending_alter, list) {
1839 * here. We guarantee there is only one thread using this undo list,
2145 list_add_tail(&queue.list,
2149 list_add_tail(&queue.list,
2153 list_add_tail(&queue.list, &curr->pending_const);
2160 list_add_tail(&queue.list, &sma->pending_alter);
2162 list_add_tail(&queue.list, &sma->pending_const);