Lines Matching refs:event

1034  * queue or arm its CQ for event generation), no further harm is expected.
1360 struct devx_event *event;
1374 event = xa_load(&dev->devx_event_table.event_xa,
1376 WARN_ON(!event);
1378 xa_val_level2 = xa_load(&event->object_ids, sub->xa_key_level2);
1380 xa_erase(&event->object_ids,
1433 struct devx_event *event;
1439 event = xa_load(&table->event_xa, MLX5_EVENT_TYPE_COMP);
1440 if (!event)
1443 obj_event = xa_load(&event->object_ids, obj_id);
1859 struct devx_event *event;
1866 event = xa_load(&devx_event_table->event_xa, key_level1);
1867 WARN_ON(!event);
1869 xa_val_level2 = xa_load(&event->object_ids,
1872 xa_erase(&event->object_ids,
1885 struct devx_event *event;
1888 event = xa_load(&devx_event_table->event_xa, key_level1);
1889 if (!event) {
1890 event = kzalloc(sizeof(*event), GFP_KERNEL);
1891 if (!event)
1894 INIT_LIST_HEAD(&event->unaffiliated_list);
1895 xa_init(&event->object_ids);
1899 event,
1902 kfree(event);
1910 obj_event = xa_load(&event->object_ids, key_level2);
1917 err = xa_insert(&event->object_ids,
2130 struct devx_event *event;
2140 event = xa_load(&devx_event_table->event_xa,
2142 WARN_ON(!event);
2146 &event->unaffiliated_list);
2150 obj_event = xa_load(&event->object_ids, obj_id);
2512 struct devx_event *event;
2531 event = xa_load(&table->event_xa, event_type | (obj_type << 16));
2532 if (!event) {
2538 dispatch_event_fd(&event->unaffiliated_list, data);
2544 obj_event = xa_load(&event->object_ids, obj_id);
2577 struct devx_event *event;
2585 event = entry;
2587 sub, tmp, &event->unaffiliated_list, xa_list)
2603 struct devx_async_data *event;
2629 event = list_entry(ev_queue->event_list.next,
2631 eventsz = event->cmd_out_len +
2642 if (copy_to_user(buf, &event->hdr, eventsz))
2647 atomic_sub(event->cmd_out_len, &ev_queue->bytes_in_use);
2648 kvfree(event);
2684 struct devx_async_event_data *event;
2727 event = list_first_entry(&ev_file->event_list,
2731 event_data = &event->hdr;
2742 list_del(&event->list);
2753 kfree(event);