Lines Matching refs:binder_node
18 * 2) node->lock : protects most fields of binder_node.
269 * struct binder_node - binder node bookkeeping
328 struct binder_node {
403 * @node: binder_node of target node. When cleaning up a
422 struct binder_node *node;
727 * binder_node_lock() - Acquire spinlock for given binder_node
728 * @node: struct binder_node to acquire
730 * Acquires node->lock. Used to protect binder_node fields
734 _binder_node_lock(struct binder_node *node, int line)
744 * @node: struct binder_node to acquire
750 _binder_node_unlock(struct binder_node *node, int line)
760 * @node: struct binder_node to acquire
763 * proc->inner_lock. Used to protect binder_node fields
767 _binder_node_inner_lock(struct binder_node *node, int line)
782 * @node: struct binder_node to acquire
788 _binder_node_inner_unlock(struct binder_node *node, int line)
947 static void binder_inc_node_tmpref_ilocked(struct binder_node *node);
1104 static struct binder_node *binder_get_node_ilocked(struct binder_proc *proc,
1108 struct binder_node *node;
1113 node = rb_entry(n, struct binder_node, rb_node);
1132 static struct binder_node *binder_get_node(struct binder_proc *proc,
1135 struct binder_node *node;
1143 static struct binder_node *binder_init_node_ilocked(
1145 struct binder_node *new_node,
1150 struct binder_node *node;
1160 node = rb_entry(parent, struct binder_node, rb_node);
1200 static struct binder_node *binder_new_node(struct binder_proc *proc,
1203 struct binder_node *node;
1204 struct binder_node *new_node = kzalloc(sizeof(*node), GFP_KERNEL);
1220 static void binder_free_node(struct binder_node *node)
1226 static int binder_inc_node_nilocked(struct binder_node *node, int strong,
1275 static int binder_inc_node(struct binder_node *node, int strong, int internal,
1287 static bool binder_dec_node_nilocked(struct binder_node *node,
1347 static void binder_dec_node(struct binder_node *node, int strong, int internal)
1358 static void binder_inc_node_tmpref_ilocked(struct binder_node *node)
1381 static void binder_inc_node_tmpref(struct binder_node *node)
1402 static void binder_dec_node_tmpref(struct binder_node *node)
1429 static void binder_put_node(struct binder_node *node)
1460 * @node: binder_node of target
1477 struct binder_node *node,
1662 * Given a proc and ref handle, return the associated binder_node
1664 * Return: a binder_node or NULL if not found or not strong when strong required
1666 static struct binder_node *binder_get_node_from_ref(
1671 struct binder_node *node;
1699 * Free the binder_ref. Free the binder_node indicated by ref->node
1787 struct binder_node *node,
2372 struct binder_node *node;
2528 struct binder_node *node;
2583 struct binder_node *node;
3095 struct binder_node *node = t->buffer->target_node;
3138 * @node: struct binder_node for which to get refs
3157 static struct binder_node *binder_get_node_refs_for_txn(
3158 struct binder_node *node,
3162 struct binder_node *target_node = NULL;
3194 struct binder_node *target_node = NULL;
4010 struct binder_node *buf_node;
4073 struct binder_node *ctx_mgr_node;
4130 struct binder_node *node;
4736 struct binder_node *node = container_of(w, struct binder_node, work);
4875 struct binder_node *target_node = t->buffer->target_node;
5370 struct binder_node *new_node;
5415 struct binder_node *node;
5458 struct binder_node *node = rb_entry(n, struct binder_node,
5878 static int binder_node_release(struct binder_node *node, int refs)
5985 struct binder_node *node;
5987 node = rb_entry(n, struct binder_node, rb_node);
6115 struct binder_node *node;
6135 node = container_of(w, struct binder_node, work);
6194 struct binder_node *node)
6241 struct binder_node *last_node = NULL;
6253 struct binder_node *node = rb_entry(n, struct binder_node,
6465 struct binder_node *node;
6466 struct binder_node *last_node = NULL;
6660 struct binder_node *last_node = NULL;
6683 struct binder_node *node = rb_entry(n, struct binder_node, rb_node);