Lines Matching refs:new_ref
1461 * @new_ref: newly allocated binder_ref to be initialized or %NULL
1472 * new_ref. new_ref must be kfree'd by the caller in
1478 struct binder_ref *new_ref)
1497 if (!new_ref)
1501 new_ref->data.debug_id = atomic_inc_return(&binder_last_id);
1502 new_ref->proc = proc;
1503 new_ref->node = node;
1504 rb_link_node(&new_ref->rb_node_node, parent, p);
1505 rb_insert_color(&new_ref->rb_node_node, &proc->refs_by_node);
1507 new_ref->data.desc = (node == context->binder_context_mgr_node) ? 0 : 1;
1510 if (ref->data.desc > new_ref->data.desc)
1512 new_ref->data.desc = ref->data.desc + 1;
1520 if (new_ref->data.desc < ref->data.desc)
1522 else if (new_ref->data.desc > ref->data.desc)
1527 rb_link_node(&new_ref->rb_node_desc, parent, p);
1528 rb_insert_color(&new_ref->rb_node_desc, &proc->refs_by_desc);
1531 hlist_add_head(&new_ref->node_entry, &node->refs);
1535 proc->pid, new_ref->data.debug_id, new_ref->data.desc,
1538 return new_ref;
1793 struct binder_ref *new_ref = NULL;
1800 new_ref = kzalloc(sizeof(*ref), GFP_KERNEL);
1801 if (!new_ref)
1804 ref = binder_get_ref_for_node_olocked(proc, node, new_ref);
1808 if (ret && ref == new_ref) {
1814 * the node. The new_ref gets kfree'd below.
1816 binder_cleanup_ref_olocked(new_ref);
1821 if (new_ref && ref != new_ref)
1826 kfree(new_ref);