xref: /kernel/linux/linux-6.6/drivers/misc/mei/bus.c (revision 62306a36)
1// SPDX-License-Identifier: GPL-2.0
2/*
3 * Copyright (c) 2012-2023, Intel Corporation. All rights reserved.
4 * Intel Management Engine Interface (Intel MEI) Linux driver
5 */
6
7#include <linux/module.h>
8#include <linux/device.h>
9#include <linux/kernel.h>
10#include <linux/sched/signal.h>
11#include <linux/init.h>
12#include <linux/errno.h>
13#include <linux/slab.h>
14#include <linux/mutex.h>
15#include <linux/interrupt.h>
16#include <linux/scatterlist.h>
17#include <linux/mei_cl_bus.h>
18
19#include "mei_dev.h"
20#include "client.h"
21
22#define to_mei_cl_driver(d) container_of(d, struct mei_cl_driver, driver)
23
24/**
25 * __mei_cl_send - internal client send (write)
26 *
27 * @cl: host client
28 * @buf: buffer to send
29 * @length: buffer length
30 * @vtag: virtual tag
31 * @mode: sending mode
32 *
33 * Return: written size bytes or < 0 on error
34 */
35ssize_t __mei_cl_send(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag,
36		      unsigned int mode)
37{
38	return __mei_cl_send_timeout(cl, buf, length, vtag, mode, MAX_SCHEDULE_TIMEOUT);
39}
40
41/**
42 * __mei_cl_send_timeout - internal client send (write)
43 *
44 * @cl: host client
45 * @buf: buffer to send
46 * @length: buffer length
47 * @vtag: virtual tag
48 * @mode: sending mode
49 * @timeout: send timeout in milliseconds.
50 *           effective only for blocking writes: the MEI_CL_IO_TX_BLOCKING mode bit is set.
51 *           set timeout to the MAX_SCHEDULE_TIMEOUT to maixum allowed wait.
52 *
53 * Return: written size bytes or < 0 on error
54 */
55ssize_t __mei_cl_send_timeout(struct mei_cl *cl, const u8 *buf, size_t length, u8 vtag,
56			      unsigned int mode, unsigned long timeout)
57{
58	struct mei_device *bus;
59	struct mei_cl_cb *cb;
60	ssize_t rets;
61
62	if (WARN_ON(!cl || !cl->dev))
63		return -ENODEV;
64
65	bus = cl->dev;
66
67	mutex_lock(&bus->device_lock);
68	if (bus->dev_state != MEI_DEV_ENABLED &&
69	    bus->dev_state != MEI_DEV_POWERING_DOWN) {
70		rets = -ENODEV;
71		goto out;
72	}
73
74	if (!mei_cl_is_connected(cl)) {
75		rets = -ENODEV;
76		goto out;
77	}
78
79	/* Check if we have an ME client device */
80	if (!mei_me_cl_is_active(cl->me_cl)) {
81		rets = -ENOTTY;
82		goto out;
83	}
84
85	if (vtag) {
86		/* Check if vtag is supported by client */
87		rets = mei_cl_vt_support_check(cl);
88		if (rets)
89			goto out;
90	}
91
92	if (length > mei_cl_mtu(cl)) {
93		rets = -EFBIG;
94		goto out;
95	}
96
97	while (cl->tx_cb_queued >= bus->tx_queue_limit) {
98		mutex_unlock(&bus->device_lock);
99		rets = wait_event_interruptible(cl->tx_wait,
100				cl->writing_state == MEI_WRITE_COMPLETE ||
101				(!mei_cl_is_connected(cl)));
102		mutex_lock(&bus->device_lock);
103		if (rets) {
104			if (signal_pending(current))
105				rets = -EINTR;
106			goto out;
107		}
108		if (!mei_cl_is_connected(cl)) {
109			rets = -ENODEV;
110			goto out;
111		}
112	}
113
114	cb = mei_cl_alloc_cb(cl, length, MEI_FOP_WRITE, NULL);
115	if (!cb) {
116		rets = -ENOMEM;
117		goto out;
118	}
119	cb->vtag = vtag;
120
121	cb->internal = !!(mode & MEI_CL_IO_TX_INTERNAL);
122	cb->blocking = !!(mode & MEI_CL_IO_TX_BLOCKING);
123	memcpy(cb->buf.data, buf, length);
124	/* hack we point data to header */
125	if (mode & MEI_CL_IO_SGL) {
126		cb->ext_hdr = (struct mei_ext_hdr *)cb->buf.data;
127		cb->buf.data = NULL;
128		cb->buf.size = 0;
129	}
130
131	rets = mei_cl_write(cl, cb, timeout);
132
133	if (mode & MEI_CL_IO_SGL && rets == 0)
134		rets = length;
135
136out:
137	mutex_unlock(&bus->device_lock);
138
139	return rets;
140}
141
142/**
143 * __mei_cl_recv - internal client receive (read)
144 *
145 * @cl: host client
146 * @buf: buffer to receive
147 * @length: buffer length
148 * @mode: io mode
149 * @vtag: virtual tag
150 * @timeout: recv timeout, 0 for infinite timeout
151 *
152 * Return: read size in bytes of < 0 on error
153 */
154ssize_t __mei_cl_recv(struct mei_cl *cl, u8 *buf, size_t length, u8 *vtag,
155		      unsigned int mode, unsigned long timeout)
156{
157	struct mei_device *bus;
158	struct mei_cl_cb *cb;
159	size_t r_length;
160	ssize_t rets;
161	bool nonblock = !!(mode & MEI_CL_IO_RX_NONBLOCK);
162
163	if (WARN_ON(!cl || !cl->dev))
164		return -ENODEV;
165
166	bus = cl->dev;
167
168	mutex_lock(&bus->device_lock);
169	if (bus->dev_state != MEI_DEV_ENABLED &&
170	    bus->dev_state != MEI_DEV_POWERING_DOWN) {
171		rets = -ENODEV;
172		goto out;
173	}
174
175	cb = mei_cl_read_cb(cl, NULL);
176	if (cb)
177		goto copy;
178
179	rets = mei_cl_read_start(cl, length, NULL);
180	if (rets && rets != -EBUSY)
181		goto out;
182
183	if (nonblock) {
184		rets = -EAGAIN;
185		goto out;
186	}
187
188	/* wait on event only if there is no other waiter */
189	/* synchronized under device mutex */
190	if (!waitqueue_active(&cl->rx_wait)) {
191
192		mutex_unlock(&bus->device_lock);
193
194		if (timeout) {
195			rets = wait_event_interruptible_timeout
196					(cl->rx_wait,
197					mei_cl_read_cb(cl, NULL) ||
198					(!mei_cl_is_connected(cl)),
199					msecs_to_jiffies(timeout));
200			if (rets == 0)
201				return -ETIME;
202			if (rets < 0) {
203				if (signal_pending(current))
204					return -EINTR;
205				return -ERESTARTSYS;
206			}
207		} else {
208			if (wait_event_interruptible
209					(cl->rx_wait,
210					mei_cl_read_cb(cl, NULL) ||
211					(!mei_cl_is_connected(cl)))) {
212				if (signal_pending(current))
213					return -EINTR;
214				return -ERESTARTSYS;
215			}
216		}
217
218		mutex_lock(&bus->device_lock);
219
220		if (!mei_cl_is_connected(cl)) {
221			rets = -ENODEV;
222			goto out;
223		}
224	}
225
226	cb = mei_cl_read_cb(cl, NULL);
227	if (!cb) {
228		rets = 0;
229		goto out;
230	}
231
232copy:
233	if (cb->status) {
234		rets = cb->status;
235		goto free;
236	}
237
238	/* for the GSC type - copy the extended header to the buffer */
239	if (cb->ext_hdr && cb->ext_hdr->type == MEI_EXT_HDR_GSC) {
240		r_length = min_t(size_t, length, cb->ext_hdr->length * sizeof(u32));
241		memcpy(buf, cb->ext_hdr, r_length);
242	} else {
243		r_length = min_t(size_t, length, cb->buf_idx);
244		memcpy(buf, cb->buf.data, r_length);
245	}
246	rets = r_length;
247
248	if (vtag)
249		*vtag = cb->vtag;
250
251free:
252	mei_cl_del_rd_completed(cl, cb);
253out:
254	mutex_unlock(&bus->device_lock);
255
256	return rets;
257}
258
259/**
260 * mei_cldev_send_vtag - me device send with vtag  (write)
261 *
262 * @cldev: me client device
263 * @buf: buffer to send
264 * @length: buffer length
265 * @vtag: virtual tag
266 *
267 * Return:
268 *  * written size in bytes
269 *  * < 0 on error
270 */
271
272ssize_t mei_cldev_send_vtag(struct mei_cl_device *cldev, const u8 *buf,
273			    size_t length, u8 vtag)
274{
275	struct mei_cl *cl = cldev->cl;
276
277	return __mei_cl_send(cl, buf, length, vtag, MEI_CL_IO_TX_BLOCKING);
278}
279EXPORT_SYMBOL_GPL(mei_cldev_send_vtag);
280
281/**
282 * mei_cldev_recv_vtag - client receive with vtag (read)
283 *
284 * @cldev: me client device
285 * @buf: buffer to receive
286 * @length: buffer length
287 * @vtag: virtual tag
288 *
289 * Return:
290 * * read size in bytes
291 * *  < 0 on error
292 */
293
294ssize_t mei_cldev_recv_vtag(struct mei_cl_device *cldev, u8 *buf, size_t length,
295			    u8 *vtag)
296{
297	struct mei_cl *cl = cldev->cl;
298
299	return __mei_cl_recv(cl, buf, length, vtag, 0, 0);
300}
301EXPORT_SYMBOL_GPL(mei_cldev_recv_vtag);
302
303/**
304 * mei_cldev_recv_nonblock_vtag - non block client receive with vtag (read)
305 *
306 * @cldev: me client device
307 * @buf: buffer to receive
308 * @length: buffer length
309 * @vtag: virtual tag
310 *
311 * Return:
312 * * read size in bytes
313 * * -EAGAIN if function will block.
314 * * < 0 on other error
315 */
316ssize_t mei_cldev_recv_nonblock_vtag(struct mei_cl_device *cldev, u8 *buf,
317				     size_t length, u8 *vtag)
318{
319	struct mei_cl *cl = cldev->cl;
320
321	return __mei_cl_recv(cl, buf, length, vtag, MEI_CL_IO_RX_NONBLOCK, 0);
322}
323EXPORT_SYMBOL_GPL(mei_cldev_recv_nonblock_vtag);
324
325/**
326 * mei_cldev_send - me device send  (write)
327 *
328 * @cldev: me client device
329 * @buf: buffer to send
330 * @length: buffer length
331 *
332 * Return:
333 *  * written size in bytes
334 *  * < 0 on error
335 */
336ssize_t mei_cldev_send(struct mei_cl_device *cldev, const u8 *buf, size_t length)
337{
338	return mei_cldev_send_vtag(cldev, buf, length, 0);
339}
340EXPORT_SYMBOL_GPL(mei_cldev_send);
341
342/**
343 * mei_cldev_recv - client receive (read)
344 *
345 * @cldev: me client device
346 * @buf: buffer to receive
347 * @length: buffer length
348 *
349 * Return: read size in bytes of < 0 on error
350 */
351ssize_t mei_cldev_recv(struct mei_cl_device *cldev, u8 *buf, size_t length)
352{
353	return mei_cldev_recv_vtag(cldev, buf, length, NULL);
354}
355EXPORT_SYMBOL_GPL(mei_cldev_recv);
356
357/**
358 * mei_cldev_recv_nonblock - non block client receive (read)
359 *
360 * @cldev: me client device
361 * @buf: buffer to receive
362 * @length: buffer length
363 *
364 * Return: read size in bytes of < 0 on error
365 *         -EAGAIN if function will block.
366 */
367ssize_t mei_cldev_recv_nonblock(struct mei_cl_device *cldev, u8 *buf,
368				size_t length)
369{
370	return mei_cldev_recv_nonblock_vtag(cldev, buf, length, NULL);
371}
372EXPORT_SYMBOL_GPL(mei_cldev_recv_nonblock);
373
374/**
375 * mei_cl_bus_rx_work - dispatch rx event for a bus device
376 *
377 * @work: work
378 */
379static void mei_cl_bus_rx_work(struct work_struct *work)
380{
381	struct mei_cl_device *cldev;
382	struct mei_device *bus;
383
384	cldev = container_of(work, struct mei_cl_device, rx_work);
385
386	bus = cldev->bus;
387
388	if (cldev->rx_cb)
389		cldev->rx_cb(cldev);
390
391	mutex_lock(&bus->device_lock);
392	if (mei_cl_is_connected(cldev->cl))
393		mei_cl_read_start(cldev->cl, mei_cl_mtu(cldev->cl), NULL);
394	mutex_unlock(&bus->device_lock);
395}
396
397/**
398 * mei_cl_bus_notif_work - dispatch FW notif event for a bus device
399 *
400 * @work: work
401 */
402static void mei_cl_bus_notif_work(struct work_struct *work)
403{
404	struct mei_cl_device *cldev;
405
406	cldev = container_of(work, struct mei_cl_device, notif_work);
407
408	if (cldev->notif_cb)
409		cldev->notif_cb(cldev);
410}
411
412/**
413 * mei_cl_bus_notify_event - schedule notify cb on bus client
414 *
415 * @cl: host client
416 *
417 * Return: true if event was scheduled
418 *         false if the client is not waiting for event
419 */
420bool mei_cl_bus_notify_event(struct mei_cl *cl)
421{
422	struct mei_cl_device *cldev = cl->cldev;
423
424	if (!cldev || !cldev->notif_cb)
425		return false;
426
427	if (!cl->notify_ev)
428		return false;
429
430	schedule_work(&cldev->notif_work);
431
432	cl->notify_ev = false;
433
434	return true;
435}
436
437/**
438 * mei_cl_bus_rx_event - schedule rx event
439 *
440 * @cl: host client
441 *
442 * Return: true if event was scheduled
443 *         false if the client is not waiting for event
444 */
445bool mei_cl_bus_rx_event(struct mei_cl *cl)
446{
447	struct mei_cl_device *cldev = cl->cldev;
448
449	if (!cldev || !cldev->rx_cb)
450		return false;
451
452	schedule_work(&cldev->rx_work);
453
454	return true;
455}
456
457/**
458 * mei_cldev_register_rx_cb - register Rx event callback
459 *
460 * @cldev: me client devices
461 * @rx_cb: callback function
462 *
463 * Return: 0 on success
464 *         -EALREADY if an callback is already registered
465 *         <0 on other errors
466 */
467int mei_cldev_register_rx_cb(struct mei_cl_device *cldev, mei_cldev_cb_t rx_cb)
468{
469	struct mei_device *bus = cldev->bus;
470	int ret;
471
472	if (!rx_cb)
473		return -EINVAL;
474	if (cldev->rx_cb)
475		return -EALREADY;
476
477	cldev->rx_cb = rx_cb;
478	INIT_WORK(&cldev->rx_work, mei_cl_bus_rx_work);
479
480	mutex_lock(&bus->device_lock);
481	if (mei_cl_is_connected(cldev->cl))
482		ret = mei_cl_read_start(cldev->cl, mei_cl_mtu(cldev->cl), NULL);
483	else
484		ret = -ENODEV;
485	mutex_unlock(&bus->device_lock);
486	if (ret && ret != -EBUSY) {
487		cancel_work_sync(&cldev->rx_work);
488		cldev->rx_cb = NULL;
489		return ret;
490	}
491
492	return 0;
493}
494EXPORT_SYMBOL_GPL(mei_cldev_register_rx_cb);
495
496/**
497 * mei_cldev_register_notif_cb - register FW notification event callback
498 *
499 * @cldev: me client devices
500 * @notif_cb: callback function
501 *
502 * Return: 0 on success
503 *         -EALREADY if an callback is already registered
504 *         <0 on other errors
505 */
506int mei_cldev_register_notif_cb(struct mei_cl_device *cldev,
507				mei_cldev_cb_t notif_cb)
508{
509	struct mei_device *bus = cldev->bus;
510	int ret;
511
512	if (!notif_cb)
513		return -EINVAL;
514
515	if (cldev->notif_cb)
516		return -EALREADY;
517
518	cldev->notif_cb = notif_cb;
519	INIT_WORK(&cldev->notif_work, mei_cl_bus_notif_work);
520
521	mutex_lock(&bus->device_lock);
522	ret = mei_cl_notify_request(cldev->cl, NULL, 1);
523	mutex_unlock(&bus->device_lock);
524	if (ret) {
525		cancel_work_sync(&cldev->notif_work);
526		cldev->notif_cb = NULL;
527		return ret;
528	}
529
530	return 0;
531}
532EXPORT_SYMBOL_GPL(mei_cldev_register_notif_cb);
533
534/**
535 * mei_cldev_get_drvdata - driver data getter
536 *
537 * @cldev: mei client device
538 *
539 * Return: driver private data
540 */
541void *mei_cldev_get_drvdata(const struct mei_cl_device *cldev)
542{
543	return dev_get_drvdata(&cldev->dev);
544}
545EXPORT_SYMBOL_GPL(mei_cldev_get_drvdata);
546
547/**
548 * mei_cldev_set_drvdata - driver data setter
549 *
550 * @cldev: mei client device
551 * @data: data to store
552 */
553void mei_cldev_set_drvdata(struct mei_cl_device *cldev, void *data)
554{
555	dev_set_drvdata(&cldev->dev, data);
556}
557EXPORT_SYMBOL_GPL(mei_cldev_set_drvdata);
558
559/**
560 * mei_cldev_uuid - return uuid of the underlying me client
561 *
562 * @cldev: mei client device
563 *
564 * Return: me client uuid
565 */
566const uuid_le *mei_cldev_uuid(const struct mei_cl_device *cldev)
567{
568	return mei_me_cl_uuid(cldev->me_cl);
569}
570EXPORT_SYMBOL_GPL(mei_cldev_uuid);
571
572/**
573 * mei_cldev_ver - return protocol version of the underlying me client
574 *
575 * @cldev: mei client device
576 *
577 * Return: me client protocol version
578 */
579u8 mei_cldev_ver(const struct mei_cl_device *cldev)
580{
581	return mei_me_cl_ver(cldev->me_cl);
582}
583EXPORT_SYMBOL_GPL(mei_cldev_ver);
584
585/**
586 * mei_cldev_enabled - check whether the device is enabled
587 *
588 * @cldev: mei client device
589 *
590 * Return: true if me client is initialized and connected
591 */
592bool mei_cldev_enabled(const struct mei_cl_device *cldev)
593{
594	return mei_cl_is_connected(cldev->cl);
595}
596EXPORT_SYMBOL_GPL(mei_cldev_enabled);
597
598/**
599 * mei_cl_bus_module_get - acquire module of the underlying
600 *    hw driver.
601 *
602 * @cldev: mei client device
603 *
604 * Return: true on success; false if the module was removed.
605 */
606static bool mei_cl_bus_module_get(struct mei_cl_device *cldev)
607{
608	return try_module_get(cldev->bus->dev->driver->owner);
609}
610
611/**
612 * mei_cl_bus_module_put -  release the underlying hw module.
613 *
614 * @cldev: mei client device
615 */
616static void mei_cl_bus_module_put(struct mei_cl_device *cldev)
617{
618	module_put(cldev->bus->dev->driver->owner);
619}
620
621/**
622 * mei_cl_bus_vtag - get bus vtag entry wrapper
623 *     The tag for bus client is always first.
624 *
625 * @cl: host client
626 *
627 * Return: bus vtag or NULL
628 */
629static inline struct mei_cl_vtag *mei_cl_bus_vtag(struct mei_cl *cl)
630{
631	return list_first_entry_or_null(&cl->vtag_map,
632					struct mei_cl_vtag, list);
633}
634
635/**
636 * mei_cl_bus_vtag_alloc - add bus client entry to vtag map
637 *
638 * @cldev: me client device
639 *
640 * Return:
641 * * 0 on success
642 * * -ENOMEM if memory allocation failed
643 */
644static int mei_cl_bus_vtag_alloc(struct mei_cl_device *cldev)
645{
646	struct mei_cl *cl = cldev->cl;
647	struct mei_cl_vtag *cl_vtag;
648
649	/*
650	 * Bail out if the client does not supports vtags
651	 * or has already allocated one
652	 */
653	if (mei_cl_vt_support_check(cl) || mei_cl_bus_vtag(cl))
654		return 0;
655
656	cl_vtag = mei_cl_vtag_alloc(NULL, 0);
657	if (IS_ERR(cl_vtag))
658		return -ENOMEM;
659
660	list_add_tail(&cl_vtag->list, &cl->vtag_map);
661
662	return 0;
663}
664
665/**
666 * mei_cl_bus_vtag_free - remove the bus entry from vtag map
667 *
668 * @cldev: me client device
669 */
670static void mei_cl_bus_vtag_free(struct mei_cl_device *cldev)
671{
672	struct mei_cl *cl = cldev->cl;
673	struct mei_cl_vtag *cl_vtag;
674
675	cl_vtag = mei_cl_bus_vtag(cl);
676	if (!cl_vtag)
677		return;
678
679	list_del(&cl_vtag->list);
680	kfree(cl_vtag);
681}
682
683void *mei_cldev_dma_map(struct mei_cl_device *cldev, u8 buffer_id, size_t size)
684{
685	struct mei_device *bus;
686	struct mei_cl *cl;
687	int ret;
688
689	if (!cldev || !buffer_id || !size)
690		return ERR_PTR(-EINVAL);
691
692	if (!IS_ALIGNED(size, MEI_FW_PAGE_SIZE)) {
693		dev_err(&cldev->dev, "Map size should be aligned to %lu\n",
694			MEI_FW_PAGE_SIZE);
695		return ERR_PTR(-EINVAL);
696	}
697
698	cl = cldev->cl;
699	bus = cldev->bus;
700
701	mutex_lock(&bus->device_lock);
702	if (cl->state == MEI_FILE_UNINITIALIZED) {
703		ret = mei_cl_link(cl);
704		if (ret)
705			goto notlinked;
706		/* update pointers */
707		cl->cldev = cldev;
708	}
709
710	ret = mei_cl_dma_alloc_and_map(cl, NULL, buffer_id, size);
711	if (ret)
712		mei_cl_unlink(cl);
713notlinked:
714	mutex_unlock(&bus->device_lock);
715	if (ret)
716		return ERR_PTR(ret);
717	return cl->dma.vaddr;
718}
719EXPORT_SYMBOL_GPL(mei_cldev_dma_map);
720
721int mei_cldev_dma_unmap(struct mei_cl_device *cldev)
722{
723	struct mei_device *bus;
724	struct mei_cl *cl;
725	int ret;
726
727	if (!cldev)
728		return -EINVAL;
729
730	cl = cldev->cl;
731	bus = cldev->bus;
732
733	mutex_lock(&bus->device_lock);
734	ret = mei_cl_dma_unmap(cl, NULL);
735
736	mei_cl_flush_queues(cl, NULL);
737	mei_cl_unlink(cl);
738	mutex_unlock(&bus->device_lock);
739	return ret;
740}
741EXPORT_SYMBOL_GPL(mei_cldev_dma_unmap);
742
743/**
744 * mei_cldev_enable - enable me client device
745 *     create connection with me client
746 *
747 * @cldev: me client device
748 *
749 * Return: 0 on success and < 0 on error
750 */
751int mei_cldev_enable(struct mei_cl_device *cldev)
752{
753	struct mei_device *bus = cldev->bus;
754	struct mei_cl *cl;
755	int ret;
756
757	cl = cldev->cl;
758
759	mutex_lock(&bus->device_lock);
760	if (cl->state == MEI_FILE_UNINITIALIZED) {
761		ret = mei_cl_link(cl);
762		if (ret)
763			goto notlinked;
764		/* update pointers */
765		cl->cldev = cldev;
766	}
767
768	if (mei_cl_is_connected(cl)) {
769		ret = 0;
770		goto out;
771	}
772
773	if (!mei_me_cl_is_active(cldev->me_cl)) {
774		dev_err(&cldev->dev, "me client is not active\n");
775		ret = -ENOTTY;
776		goto out;
777	}
778
779	ret = mei_cl_bus_vtag_alloc(cldev);
780	if (ret)
781		goto out;
782
783	ret = mei_cl_connect(cl, cldev->me_cl, NULL);
784	if (ret < 0) {
785		dev_err(&cldev->dev, "cannot connect\n");
786		mei_cl_bus_vtag_free(cldev);
787	}
788
789out:
790	if (ret)
791		mei_cl_unlink(cl);
792notlinked:
793	mutex_unlock(&bus->device_lock);
794
795	return ret;
796}
797EXPORT_SYMBOL_GPL(mei_cldev_enable);
798
799/**
800 * mei_cldev_unregister_callbacks - internal wrapper for unregistering
801 *  callbacks.
802 *
803 * @cldev: client device
804 */
805static void mei_cldev_unregister_callbacks(struct mei_cl_device *cldev)
806{
807	if (cldev->rx_cb) {
808		cancel_work_sync(&cldev->rx_work);
809		cldev->rx_cb = NULL;
810	}
811
812	if (cldev->notif_cb) {
813		cancel_work_sync(&cldev->notif_work);
814		cldev->notif_cb = NULL;
815	}
816}
817
818/**
819 * mei_cldev_disable - disable me client device
820 *     disconnect form the me client
821 *
822 * @cldev: me client device
823 *
824 * Return: 0 on success and < 0 on error
825 */
826int mei_cldev_disable(struct mei_cl_device *cldev)
827{
828	struct mei_device *bus;
829	struct mei_cl *cl;
830	int err;
831
832	if (!cldev)
833		return -ENODEV;
834
835	cl = cldev->cl;
836
837	bus = cldev->bus;
838
839	mei_cldev_unregister_callbacks(cldev);
840
841	mutex_lock(&bus->device_lock);
842
843	mei_cl_bus_vtag_free(cldev);
844
845	if (!mei_cl_is_connected(cl)) {
846		dev_dbg(bus->dev, "Already disconnected\n");
847		err = 0;
848		goto out;
849	}
850
851	err = mei_cl_disconnect(cl);
852	if (err < 0)
853		dev_err(bus->dev, "Could not disconnect from the ME client\n");
854
855out:
856	/* Flush queues and remove any pending read unless we have mapped DMA */
857	if (!cl->dma_mapped) {
858		mei_cl_flush_queues(cl, NULL);
859		mei_cl_unlink(cl);
860	}
861
862	mutex_unlock(&bus->device_lock);
863	return err;
864}
865EXPORT_SYMBOL_GPL(mei_cldev_disable);
866
867/**
868 * mei_cldev_send_gsc_command - sends a gsc command, by sending
869 * a gsl mei message to gsc and receiving reply from gsc
870 *
871 * @cldev: me client device
872 * @client_id: client id to send the command to
873 * @fence_id: fence id to send the command to
874 * @sg_in: scatter gather list containing addresses for rx message buffer
875 * @total_in_len: total length of data in 'in' sg, can be less than the sum of buffers sizes
876 * @sg_out: scatter gather list containing addresses for tx message buffer
877 *
878 * Return:
879 *  * written size in bytes
880 *  * < 0 on error
881 */
882ssize_t mei_cldev_send_gsc_command(struct mei_cl_device *cldev,
883				   u8 client_id, u32 fence_id,
884				   struct scatterlist *sg_in,
885				   size_t total_in_len,
886				   struct scatterlist *sg_out)
887{
888	struct mei_cl *cl;
889	struct mei_device *bus;
890	ssize_t ret = 0;
891
892	struct mei_ext_hdr_gsc_h2f *ext_hdr;
893	size_t buf_sz = sizeof(struct mei_ext_hdr_gsc_h2f);
894	int sg_out_nents, sg_in_nents;
895	int i;
896	struct scatterlist *sg;
897	struct mei_ext_hdr_gsc_f2h rx_msg;
898	unsigned int sg_len;
899
900	if (!cldev || !sg_in || !sg_out)
901		return -EINVAL;
902
903	cl = cldev->cl;
904	bus = cldev->bus;
905
906	dev_dbg(bus->dev, "client_id %u, fence_id %u\n", client_id, fence_id);
907
908	if (!bus->hbm_f_gsc_supported)
909		return -EOPNOTSUPP;
910
911	sg_out_nents = sg_nents(sg_out);
912	sg_in_nents = sg_nents(sg_in);
913	/* at least one entry in tx and rx sgls must be present */
914	if (sg_out_nents <= 0 || sg_in_nents <= 0)
915		return -EINVAL;
916
917	buf_sz += (sg_out_nents + sg_in_nents) * sizeof(struct mei_gsc_sgl);
918	ext_hdr = kzalloc(buf_sz, GFP_KERNEL);
919	if (!ext_hdr)
920		return -ENOMEM;
921
922	/* construct the GSC message */
923	ext_hdr->hdr.type = MEI_EXT_HDR_GSC;
924	ext_hdr->hdr.length = buf_sz / sizeof(u32); /* length is in dw */
925
926	ext_hdr->client_id = client_id;
927	ext_hdr->addr_type = GSC_ADDRESS_TYPE_PHYSICAL_SGL;
928	ext_hdr->fence_id = fence_id;
929	ext_hdr->input_address_count = sg_in_nents;
930	ext_hdr->output_address_count = sg_out_nents;
931	ext_hdr->reserved[0] = 0;
932	ext_hdr->reserved[1] = 0;
933
934	/* copy in-sgl to the message */
935	for (i = 0, sg = sg_in; i < sg_in_nents; i++, sg++) {
936		ext_hdr->sgl[i].low = lower_32_bits(sg_dma_address(sg));
937		ext_hdr->sgl[i].high = upper_32_bits(sg_dma_address(sg));
938		sg_len = min_t(unsigned int, sg_dma_len(sg), PAGE_SIZE);
939		ext_hdr->sgl[i].length = (sg_len <= total_in_len) ? sg_len : total_in_len;
940		total_in_len -= ext_hdr->sgl[i].length;
941	}
942
943	/* copy out-sgl to the message */
944	for (i = sg_in_nents, sg = sg_out; i < sg_in_nents + sg_out_nents; i++, sg++) {
945		ext_hdr->sgl[i].low = lower_32_bits(sg_dma_address(sg));
946		ext_hdr->sgl[i].high = upper_32_bits(sg_dma_address(sg));
947		sg_len = min_t(unsigned int, sg_dma_len(sg), PAGE_SIZE);
948		ext_hdr->sgl[i].length = sg_len;
949	}
950
951	/* send the message to GSC */
952	ret = __mei_cl_send(cl, (u8 *)ext_hdr, buf_sz, 0, MEI_CL_IO_SGL);
953	if (ret < 0) {
954		dev_err(bus->dev, "__mei_cl_send failed, returned %zd\n", ret);
955		goto end;
956	}
957	if (ret != buf_sz) {
958		dev_err(bus->dev, "__mei_cl_send returned %zd instead of expected %zd\n",
959			ret, buf_sz);
960		ret = -EIO;
961		goto end;
962	}
963
964	/* receive the reply from GSC, note that at this point sg_in should contain the reply */
965	ret = __mei_cl_recv(cl, (u8 *)&rx_msg, sizeof(rx_msg), NULL, MEI_CL_IO_SGL, 0);
966
967	if (ret != sizeof(rx_msg)) {
968		dev_err(bus->dev, "__mei_cl_recv returned %zd instead of expected %zd\n",
969			ret, sizeof(rx_msg));
970		if (ret >= 0)
971			ret = -EIO;
972		goto end;
973	}
974
975	/* check rx_msg.client_id and rx_msg.fence_id match the ones we send */
976	if (rx_msg.client_id != client_id || rx_msg.fence_id != fence_id) {
977		dev_err(bus->dev, "received client_id/fence_id  %u/%u  instead of %u/%u sent\n",
978			rx_msg.client_id, rx_msg.fence_id, client_id, fence_id);
979		ret = -EFAULT;
980		goto end;
981	}
982
983	dev_dbg(bus->dev, "gsc command: successfully written %u bytes\n",  rx_msg.written);
984	ret = rx_msg.written;
985
986end:
987	kfree(ext_hdr);
988	return ret;
989}
990EXPORT_SYMBOL_GPL(mei_cldev_send_gsc_command);
991
992/**
993 * mei_cl_device_find - find matching entry in the driver id table
994 *
995 * @cldev: me client device
996 * @cldrv: me client driver
997 *
998 * Return: id on success; NULL if no id is matching
999 */
1000static const
1001struct mei_cl_device_id *mei_cl_device_find(const struct mei_cl_device *cldev,
1002					    const struct mei_cl_driver *cldrv)
1003{
1004	const struct mei_cl_device_id *id;
1005	const uuid_le *uuid;
1006	u8 version;
1007	bool match;
1008
1009	uuid = mei_me_cl_uuid(cldev->me_cl);
1010	version = mei_me_cl_ver(cldev->me_cl);
1011
1012	id = cldrv->id_table;
1013	while (uuid_le_cmp(NULL_UUID_LE, id->uuid)) {
1014		if (!uuid_le_cmp(*uuid, id->uuid)) {
1015			match = true;
1016
1017			if (cldev->name[0])
1018				if (strncmp(cldev->name, id->name,
1019					    sizeof(id->name)))
1020					match = false;
1021
1022			if (id->version != MEI_CL_VERSION_ANY)
1023				if (id->version != version)
1024					match = false;
1025			if (match)
1026				return id;
1027		}
1028
1029		id++;
1030	}
1031
1032	return NULL;
1033}
1034
1035/**
1036 * mei_cl_device_match  - device match function
1037 *
1038 * @dev: device
1039 * @drv: driver
1040 *
1041 * Return:  1 if matching device was found 0 otherwise
1042 */
1043static int mei_cl_device_match(struct device *dev, struct device_driver *drv)
1044{
1045	const struct mei_cl_device *cldev = to_mei_cl_device(dev);
1046	const struct mei_cl_driver *cldrv = to_mei_cl_driver(drv);
1047	const struct mei_cl_device_id *found_id;
1048
1049	if (!cldev->do_match)
1050		return 0;
1051
1052	if (!cldrv || !cldrv->id_table)
1053		return 0;
1054
1055	found_id = mei_cl_device_find(cldev, cldrv);
1056	if (found_id)
1057		return 1;
1058
1059	return 0;
1060}
1061
1062/**
1063 * mei_cl_device_probe - bus probe function
1064 *
1065 * @dev: device
1066 *
1067 * Return:  0 on success; < 0 otherwise
1068 */
1069static int mei_cl_device_probe(struct device *dev)
1070{
1071	struct mei_cl_device *cldev;
1072	struct mei_cl_driver *cldrv;
1073	const struct mei_cl_device_id *id;
1074	int ret;
1075
1076	cldev = to_mei_cl_device(dev);
1077	cldrv = to_mei_cl_driver(dev->driver);
1078
1079	if (!cldrv || !cldrv->probe)
1080		return -ENODEV;
1081
1082	id = mei_cl_device_find(cldev, cldrv);
1083	if (!id)
1084		return -ENODEV;
1085
1086	if (!mei_cl_bus_module_get(cldev)) {
1087		dev_err(&cldev->dev, "get hw module failed");
1088		return -ENODEV;
1089	}
1090
1091	ret = cldrv->probe(cldev, id);
1092	if (ret) {
1093		mei_cl_bus_module_put(cldev);
1094		return ret;
1095	}
1096
1097	__module_get(THIS_MODULE);
1098	return 0;
1099}
1100
1101/**
1102 * mei_cl_device_remove - remove device from the bus
1103 *
1104 * @dev: device
1105 *
1106 * Return:  0 on success; < 0 otherwise
1107 */
1108static void mei_cl_device_remove(struct device *dev)
1109{
1110	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1111	struct mei_cl_driver *cldrv = to_mei_cl_driver(dev->driver);
1112
1113	if (cldrv->remove)
1114		cldrv->remove(cldev);
1115
1116	mei_cldev_unregister_callbacks(cldev);
1117
1118	mei_cl_bus_module_put(cldev);
1119	module_put(THIS_MODULE);
1120}
1121
1122static ssize_t name_show(struct device *dev, struct device_attribute *a,
1123			     char *buf)
1124{
1125	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1126
1127	return scnprintf(buf, PAGE_SIZE, "%s", cldev->name);
1128}
1129static DEVICE_ATTR_RO(name);
1130
1131static ssize_t uuid_show(struct device *dev, struct device_attribute *a,
1132			     char *buf)
1133{
1134	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1135	const uuid_le *uuid = mei_me_cl_uuid(cldev->me_cl);
1136
1137	return sprintf(buf, "%pUl", uuid);
1138}
1139static DEVICE_ATTR_RO(uuid);
1140
1141static ssize_t version_show(struct device *dev, struct device_attribute *a,
1142			     char *buf)
1143{
1144	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1145	u8 version = mei_me_cl_ver(cldev->me_cl);
1146
1147	return sprintf(buf, "%02X", version);
1148}
1149static DEVICE_ATTR_RO(version);
1150
1151static ssize_t modalias_show(struct device *dev, struct device_attribute *a,
1152			     char *buf)
1153{
1154	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1155	const uuid_le *uuid = mei_me_cl_uuid(cldev->me_cl);
1156	u8 version = mei_me_cl_ver(cldev->me_cl);
1157
1158	return scnprintf(buf, PAGE_SIZE, "mei:%s:%pUl:%02X:",
1159			 cldev->name, uuid, version);
1160}
1161static DEVICE_ATTR_RO(modalias);
1162
1163static ssize_t max_conn_show(struct device *dev, struct device_attribute *a,
1164			     char *buf)
1165{
1166	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1167	u8 maxconn = mei_me_cl_max_conn(cldev->me_cl);
1168
1169	return sprintf(buf, "%d", maxconn);
1170}
1171static DEVICE_ATTR_RO(max_conn);
1172
1173static ssize_t fixed_show(struct device *dev, struct device_attribute *a,
1174			  char *buf)
1175{
1176	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1177	u8 fixed = mei_me_cl_fixed(cldev->me_cl);
1178
1179	return sprintf(buf, "%d", fixed);
1180}
1181static DEVICE_ATTR_RO(fixed);
1182
1183static ssize_t vtag_show(struct device *dev, struct device_attribute *a,
1184			 char *buf)
1185{
1186	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1187	bool vt = mei_me_cl_vt(cldev->me_cl);
1188
1189	return sprintf(buf, "%d", vt);
1190}
1191static DEVICE_ATTR_RO(vtag);
1192
1193static ssize_t max_len_show(struct device *dev, struct device_attribute *a,
1194			    char *buf)
1195{
1196	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1197	u32 maxlen = mei_me_cl_max_len(cldev->me_cl);
1198
1199	return sprintf(buf, "%u", maxlen);
1200}
1201static DEVICE_ATTR_RO(max_len);
1202
1203static struct attribute *mei_cldev_attrs[] = {
1204	&dev_attr_name.attr,
1205	&dev_attr_uuid.attr,
1206	&dev_attr_version.attr,
1207	&dev_attr_modalias.attr,
1208	&dev_attr_max_conn.attr,
1209	&dev_attr_fixed.attr,
1210	&dev_attr_vtag.attr,
1211	&dev_attr_max_len.attr,
1212	NULL,
1213};
1214ATTRIBUTE_GROUPS(mei_cldev);
1215
1216/**
1217 * mei_cl_device_uevent - me client bus uevent handler
1218 *
1219 * @dev: device
1220 * @env: uevent kobject
1221 *
1222 * Return: 0 on success -ENOMEM on when add_uevent_var fails
1223 */
1224static int mei_cl_device_uevent(const struct device *dev, struct kobj_uevent_env *env)
1225{
1226	const struct mei_cl_device *cldev = to_mei_cl_device(dev);
1227	const uuid_le *uuid = mei_me_cl_uuid(cldev->me_cl);
1228	u8 version = mei_me_cl_ver(cldev->me_cl);
1229
1230	if (add_uevent_var(env, "MEI_CL_VERSION=%d", version))
1231		return -ENOMEM;
1232
1233	if (add_uevent_var(env, "MEI_CL_UUID=%pUl", uuid))
1234		return -ENOMEM;
1235
1236	if (add_uevent_var(env, "MEI_CL_NAME=%s", cldev->name))
1237		return -ENOMEM;
1238
1239	if (add_uevent_var(env, "MODALIAS=mei:%s:%pUl:%02X:",
1240			   cldev->name, uuid, version))
1241		return -ENOMEM;
1242
1243	return 0;
1244}
1245
1246static struct bus_type mei_cl_bus_type = {
1247	.name		= "mei",
1248	.dev_groups	= mei_cldev_groups,
1249	.match		= mei_cl_device_match,
1250	.probe		= mei_cl_device_probe,
1251	.remove		= mei_cl_device_remove,
1252	.uevent		= mei_cl_device_uevent,
1253};
1254
1255static struct mei_device *mei_dev_bus_get(struct mei_device *bus)
1256{
1257	if (bus)
1258		get_device(bus->dev);
1259
1260	return bus;
1261}
1262
1263static void mei_dev_bus_put(struct mei_device *bus)
1264{
1265	if (bus)
1266		put_device(bus->dev);
1267}
1268
1269static void mei_cl_bus_dev_release(struct device *dev)
1270{
1271	struct mei_cl_device *cldev = to_mei_cl_device(dev);
1272
1273	mei_cl_flush_queues(cldev->cl, NULL);
1274	mei_me_cl_put(cldev->me_cl);
1275	mei_dev_bus_put(cldev->bus);
1276	kfree(cldev->cl);
1277	kfree(cldev);
1278}
1279
1280static const struct device_type mei_cl_device_type = {
1281	.release = mei_cl_bus_dev_release,
1282};
1283
1284/**
1285 * mei_cl_bus_set_name - set device name for me client device
1286 *  <controller>-<client device>
1287 *  Example: 0000:00:16.0-55213584-9a29-4916-badf-0fb7ed682aeb
1288 *
1289 * @cldev: me client device
1290 */
1291static inline void mei_cl_bus_set_name(struct mei_cl_device *cldev)
1292{
1293	dev_set_name(&cldev->dev, "%s-%pUl",
1294		     dev_name(cldev->bus->dev),
1295		     mei_me_cl_uuid(cldev->me_cl));
1296}
1297
1298/**
1299 * mei_cl_bus_dev_alloc - initialize and allocate mei client device
1300 *
1301 * @bus: mei device
1302 * @me_cl: me client
1303 *
1304 * Return: allocated device structur or NULL on allocation failure
1305 */
1306static struct mei_cl_device *mei_cl_bus_dev_alloc(struct mei_device *bus,
1307						  struct mei_me_client *me_cl)
1308{
1309	struct mei_cl_device *cldev;
1310	struct mei_cl *cl;
1311
1312	cldev = kzalloc(sizeof(*cldev), GFP_KERNEL);
1313	if (!cldev)
1314		return NULL;
1315
1316	cl = mei_cl_allocate(bus);
1317	if (!cl) {
1318		kfree(cldev);
1319		return NULL;
1320	}
1321
1322	device_initialize(&cldev->dev);
1323	cldev->dev.parent = bus->dev;
1324	cldev->dev.bus    = &mei_cl_bus_type;
1325	cldev->dev.type   = &mei_cl_device_type;
1326	cldev->bus        = mei_dev_bus_get(bus);
1327	cldev->me_cl      = mei_me_cl_get(me_cl);
1328	cldev->cl         = cl;
1329	mei_cl_bus_set_name(cldev);
1330	cldev->is_added   = 0;
1331	INIT_LIST_HEAD(&cldev->bus_list);
1332	device_enable_async_suspend(&cldev->dev);
1333
1334	return cldev;
1335}
1336
1337/**
1338 * mei_cl_bus_dev_setup - setup me client device
1339 *    run fix up routines and set the device name
1340 *
1341 * @bus: mei device
1342 * @cldev: me client device
1343 *
1344 * Return: true if the device is eligible for enumeration
1345 */
1346static bool mei_cl_bus_dev_setup(struct mei_device *bus,
1347				 struct mei_cl_device *cldev)
1348{
1349	cldev->do_match = 1;
1350	mei_cl_bus_dev_fixup(cldev);
1351
1352	/* the device name can change during fix up */
1353	if (cldev->do_match)
1354		mei_cl_bus_set_name(cldev);
1355
1356	return cldev->do_match == 1;
1357}
1358
1359/**
1360 * mei_cl_bus_dev_add - add me client devices
1361 *
1362 * @cldev: me client device
1363 *
1364 * Return: 0 on success; < 0 on failre
1365 */
1366static int mei_cl_bus_dev_add(struct mei_cl_device *cldev)
1367{
1368	int ret;
1369
1370	dev_dbg(cldev->bus->dev, "adding %pUL:%02X\n",
1371		mei_me_cl_uuid(cldev->me_cl),
1372		mei_me_cl_ver(cldev->me_cl));
1373	ret = device_add(&cldev->dev);
1374	if (!ret)
1375		cldev->is_added = 1;
1376
1377	return ret;
1378}
1379
1380/**
1381 * mei_cl_bus_dev_stop - stop the driver
1382 *
1383 * @cldev: me client device
1384 */
1385static void mei_cl_bus_dev_stop(struct mei_cl_device *cldev)
1386{
1387	cldev->do_match = 0;
1388	if (cldev->is_added)
1389		device_release_driver(&cldev->dev);
1390}
1391
1392/**
1393 * mei_cl_bus_dev_destroy - destroy me client devices object
1394 *
1395 * @cldev: me client device
1396 *
1397 * Locking: called under "dev->cl_bus_lock" lock
1398 */
1399static void mei_cl_bus_dev_destroy(struct mei_cl_device *cldev)
1400{
1401
1402	WARN_ON(!mutex_is_locked(&cldev->bus->cl_bus_lock));
1403
1404	if (!cldev->is_added)
1405		return;
1406
1407	device_del(&cldev->dev);
1408
1409	list_del_init(&cldev->bus_list);
1410
1411	cldev->is_added = 0;
1412	put_device(&cldev->dev);
1413}
1414
1415/**
1416 * mei_cl_bus_remove_device - remove a devices form the bus
1417 *
1418 * @cldev: me client device
1419 */
1420static void mei_cl_bus_remove_device(struct mei_cl_device *cldev)
1421{
1422	mei_cl_bus_dev_stop(cldev);
1423	mei_cl_bus_dev_destroy(cldev);
1424}
1425
1426/**
1427 * mei_cl_bus_remove_devices - remove all devices form the bus
1428 *
1429 * @bus: mei device
1430 */
1431void mei_cl_bus_remove_devices(struct mei_device *bus)
1432{
1433	struct mei_cl_device *cldev, *next;
1434
1435	mutex_lock(&bus->cl_bus_lock);
1436	list_for_each_entry_safe(cldev, next, &bus->device_list, bus_list)
1437		mei_cl_bus_remove_device(cldev);
1438	mutex_unlock(&bus->cl_bus_lock);
1439}
1440
1441
1442/**
1443 * mei_cl_bus_dev_init - allocate and initializes an mei client devices
1444 *     based on me client
1445 *
1446 * @bus: mei device
1447 * @me_cl: me client
1448 *
1449 * Locking: called under "dev->cl_bus_lock" lock
1450 */
1451static void mei_cl_bus_dev_init(struct mei_device *bus,
1452				struct mei_me_client *me_cl)
1453{
1454	struct mei_cl_device *cldev;
1455
1456	WARN_ON(!mutex_is_locked(&bus->cl_bus_lock));
1457
1458	dev_dbg(bus->dev, "initializing %pUl", mei_me_cl_uuid(me_cl));
1459
1460	if (me_cl->bus_added)
1461		return;
1462
1463	cldev = mei_cl_bus_dev_alloc(bus, me_cl);
1464	if (!cldev)
1465		return;
1466
1467	me_cl->bus_added = true;
1468	list_add_tail(&cldev->bus_list, &bus->device_list);
1469
1470}
1471
1472/**
1473 * mei_cl_bus_rescan - scan me clients list and add create
1474 *    devices for eligible clients
1475 *
1476 * @bus: mei device
1477 */
1478static void mei_cl_bus_rescan(struct mei_device *bus)
1479{
1480	struct mei_cl_device *cldev, *n;
1481	struct mei_me_client *me_cl;
1482
1483	mutex_lock(&bus->cl_bus_lock);
1484
1485	down_read(&bus->me_clients_rwsem);
1486	list_for_each_entry(me_cl, &bus->me_clients, list)
1487		mei_cl_bus_dev_init(bus, me_cl);
1488	up_read(&bus->me_clients_rwsem);
1489
1490	list_for_each_entry_safe(cldev, n, &bus->device_list, bus_list) {
1491
1492		if (!mei_me_cl_is_active(cldev->me_cl)) {
1493			mei_cl_bus_remove_device(cldev);
1494			continue;
1495		}
1496
1497		if (cldev->is_added)
1498			continue;
1499
1500		if (mei_cl_bus_dev_setup(bus, cldev))
1501			mei_cl_bus_dev_add(cldev);
1502		else {
1503			list_del_init(&cldev->bus_list);
1504			put_device(&cldev->dev);
1505		}
1506	}
1507	mutex_unlock(&bus->cl_bus_lock);
1508
1509	dev_dbg(bus->dev, "rescan end");
1510}
1511
1512void mei_cl_bus_rescan_work(struct work_struct *work)
1513{
1514	struct mei_device *bus =
1515		container_of(work, struct mei_device, bus_rescan_work);
1516
1517	mei_cl_bus_rescan(bus);
1518}
1519
1520int __mei_cldev_driver_register(struct mei_cl_driver *cldrv,
1521				struct module *owner)
1522{
1523	int err;
1524
1525	cldrv->driver.name = cldrv->name;
1526	cldrv->driver.owner = owner;
1527	cldrv->driver.bus = &mei_cl_bus_type;
1528
1529	err = driver_register(&cldrv->driver);
1530	if (err)
1531		return err;
1532
1533	pr_debug("mei: driver [%s] registered\n", cldrv->driver.name);
1534
1535	return 0;
1536}
1537EXPORT_SYMBOL_GPL(__mei_cldev_driver_register);
1538
1539void mei_cldev_driver_unregister(struct mei_cl_driver *cldrv)
1540{
1541	driver_unregister(&cldrv->driver);
1542
1543	pr_debug("mei: driver [%s] unregistered\n", cldrv->driver.name);
1544}
1545EXPORT_SYMBOL_GPL(mei_cldev_driver_unregister);
1546
1547
1548int __init mei_cl_bus_init(void)
1549{
1550	return bus_register(&mei_cl_bus_type);
1551}
1552
1553void __exit mei_cl_bus_exit(void)
1554{
1555	bus_unregister(&mei_cl_bus_type);
1556}
1557