/kernel/linux/linux-5.10/drivers/input/ |
H A D | ff-memless.c | 112 static void ml_schedule_timer(struct ml_device *ml) in ml_schedule_timer() argument 125 state = &ml->states[i]; in ml_schedule_timer() 142 del_timer(&ml->timer); in ml_schedule_timer() 145 mod_timer(&ml->timer, earliest); in ml_schedule_timer() 326 static int ml_get_combo_effect(struct ml_device *ml, in ml_get_combo_effect() argument 341 state = &ml->states[i]; in ml_get_combo_effect() 355 effect_type = get_compatible_type(ml->dev->ff, effect->type); in ml_get_combo_effect() 383 ml_combine_effects(combo_effect, state, ml->gain); in ml_get_combo_effect() 390 static void ml_play_effects(struct ml_device *ml) in ml_play_effects() argument 397 while (ml_get_combo_effect(ml, handled_b in ml_play_effects() 405 struct ml_device *ml = from_timer(ml, t, timer); ml_effect_timer() local 421 struct ml_device *ml = dev->ff->private; ml_ff_set_gain() local 437 struct ml_device *ml = dev->ff->private; ml_ff_playback() local 468 struct ml_device *ml = dev->ff->private; ml_ff_upload() local 490 struct ml_device *ml = ff->private; ml_ff_destroy() local 513 struct ml_device *ml; input_ff_create_memless() local [all...] |
/kernel/linux/linux-6.6/drivers/input/ |
H A D | ff-memless.c | 109 static void ml_schedule_timer(struct ml_device *ml) in ml_schedule_timer() argument 122 state = &ml->states[i]; in ml_schedule_timer() 139 del_timer(&ml->timer); in ml_schedule_timer() 142 mod_timer(&ml->timer, earliest); in ml_schedule_timer() 323 static int ml_get_combo_effect(struct ml_device *ml, in ml_get_combo_effect() argument 338 state = &ml->states[i]; in ml_get_combo_effect() 352 effect_type = get_compatible_type(ml->dev->ff, effect->type); in ml_get_combo_effect() 380 ml_combine_effects(combo_effect, state, ml->gain); in ml_get_combo_effect() 387 static void ml_play_effects(struct ml_device *ml) in ml_play_effects() argument 394 while (ml_get_combo_effect(ml, handled_b in ml_play_effects() 402 struct ml_device *ml = from_timer(ml, t, timer); ml_effect_timer() local 418 struct ml_device *ml = dev->ff->private; ml_ff_set_gain() local 434 struct ml_device *ml = dev->ff->private; ml_ff_playback() local 465 struct ml_device *ml = dev->ff->private; ml_ff_upload() local 487 struct ml_device *ml = ff->private; ml_ff_destroy() local 510 struct ml_device *ml; input_ff_create_memless() local [all...] |
/kernel/linux/linux-5.10/fs/ocfs2/dlm/ |
H A D | dlmast.c | 54 if (lock->ml.highest_blocked == LKM_IVMODE) in dlm_should_cancel_bast() 56 BUG_ON(lock->ml.highest_blocked == LKM_NLMODE); in dlm_should_cancel_bast() 63 if (lock->ml.type == LKM_EXMODE) in dlm_should_cancel_bast() 66 else if (lock->ml.type == LKM_NLMODE) in dlm_should_cancel_bast() 69 else if (lock->ml.highest_blocked != LKM_EXMODE) in dlm_should_cancel_bast() 91 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 92 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 93 lock->ast_pending, lock->ml.type); in __dlm_queue_ast() 99 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 100 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml in __dlm_queue_ast() [all...] |
H A D | dlmconvert.c | 116 lock->ml.type, lock->ml.convert_type, type); in __dlmconvert_master() 121 if (lock->ml.convert_type != LKM_IVMODE) { in __dlmconvert_master() 137 switch (lock->ml.type) { in __dlmconvert_master() 141 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 151 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 157 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 167 if (type <= lock->ml.type) in __dlmconvert_master() 175 if (!dlm_lock_compatible(tmplock->ml.type, type)) in __dlmconvert_master() 180 if (!dlm_lock_compatible(tmplock->ml in __dlmconvert_master() [all...] |
H A D | dlmthread.c | 367 if (target->ml.convert_type == LKM_IVMODE) { in dlm_shuffle_lists() 375 if (!dlm_lock_compatible(lock->ml.type, in dlm_shuffle_lists() 376 target->ml.convert_type)) { in dlm_shuffle_lists() 379 if (lock->ml.highest_blocked == LKM_IVMODE) { in dlm_shuffle_lists() 384 if (lock->ml.highest_blocked < target->ml.convert_type) in dlm_shuffle_lists() 385 lock->ml.highest_blocked = in dlm_shuffle_lists() 386 target->ml.convert_type; in dlm_shuffle_lists() 393 if (!dlm_lock_compatible(lock->ml.type, in dlm_shuffle_lists() 394 target->ml in dlm_shuffle_lists() [all...] |
H A D | dlmrecovery.c | 1058 if (lock->ml.node == dead_node) { in dlm_move_reco_locks_to_list() 1133 struct_size(mres, ml, mres->num_locks), in dlm_send_mig_lockres_msg() 1192 if (lock->ml.type != LKM_EXMODE && lock->ml.type != LKM_PRMODE) in dlm_prepare_lvb_for_migration() 1206 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dlm_prepare_lvb_for_migration() 1207 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in dlm_prepare_lvb_for_migration() 1209 lock->ml.node); in dlm_prepare_lvb_for_migration() 1219 struct dlm_migratable_lock *ml; in dlm_add_lock_to_array() local 1222 ml = &(mres->ml[lock_nu in dlm_add_lock_to_array() 1254 dlm_is_dummy_lock(struct dlm_ctxt *dlm, struct dlm_migratable_lock *ml, u8 *nodenum) dlm_is_dummy_lock() argument 1800 struct dlm_migratable_lock *ml; dlm_process_recovery_data() local [all...] |
H A D | dlmlock.c | 81 if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) in dlm_can_grant_new_lock() 86 if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) in dlm_can_grant_new_lock() 88 if (!dlm_lock_compatible(tmplock->ml.convert_type, in dlm_can_grant_new_lock() 89 lock->ml.type)) in dlm_can_grant_new_lock() 110 mlog(0, "type=%d\n", lock->ml.type); in dlmlock_master() 117 lock->ml.node != dlm->node_num) { in dlmlock_master() 146 lock->ml.node); in dlmlock_master() 157 lock->ml in dlmlock_master() [all...] |
H A D | dlmunlock.c | 220 lock->ml.convert_type = LKM_IVMODE; in dlmunlock_common() 229 BUG_ON(lock->ml.convert_type != LKM_IVMODE); in dlmunlock_common() 231 BUG_ON(lock->ml.convert_type == LKM_IVMODE); in dlmunlock_common() 252 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dlmunlock_common() 253 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in dlmunlock_common() 279 lock->ml.convert_type = LKM_IVMODE; in dlm_commit_pending_cancel() 338 unlock.cookie = lock->ml.cookie; in dlm_send_remote_unlock_request() 467 if (lock->ml.cookie == unlock->cookie && in dlm_unlock_lock_handler() 468 lock->ml.node == unlock->node_idx) { in dlm_unlock_lock_handler() 488 lock->ml in dlm_unlock_lock_handler() [all...] |
H A D | dlmdebug.c | 66 lock->ml.type, lock->ml.convert_type, lock->ml.node, in __dlm_print_lock() 67 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_print_lock() 68 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in __dlm_print_lock() 473 list_type, lock->ml.type, lock->ml.convert_type, in dump_lock() 474 lock->ml.node, in dump_lock() 475 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dump_lock() 476 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml in dump_lock() [all...] |
/kernel/linux/linux-6.6/fs/ocfs2/dlm/ |
H A D | dlmast.c | 52 if (lock->ml.highest_blocked == LKM_IVMODE) in dlm_should_cancel_bast() 54 BUG_ON(lock->ml.highest_blocked == LKM_NLMODE); in dlm_should_cancel_bast() 61 if (lock->ml.type == LKM_EXMODE) in dlm_should_cancel_bast() 64 else if (lock->ml.type == LKM_NLMODE) in dlm_should_cancel_bast() 67 else if (lock->ml.highest_blocked != LKM_EXMODE) in dlm_should_cancel_bast() 89 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 90 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 91 lock->ast_pending, lock->ml.type); in __dlm_queue_ast() 97 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_queue_ast() 98 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml in __dlm_queue_ast() [all...] |
H A D | dlmconvert.c | 114 lock->ml.type, lock->ml.convert_type, type); in __dlmconvert_master() 119 if (lock->ml.convert_type != LKM_IVMODE) { in __dlmconvert_master() 135 switch (lock->ml.type) { in __dlmconvert_master() 139 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 149 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 155 dlm_lock_mode_name(lock->ml.type), in __dlmconvert_master() 165 if (type <= lock->ml.type) in __dlmconvert_master() 173 if (!dlm_lock_compatible(tmplock->ml.type, type)) in __dlmconvert_master() 178 if (!dlm_lock_compatible(tmplock->ml in __dlmconvert_master() [all...] |
H A D | dlmthread.c | 365 if (target->ml.convert_type == LKM_IVMODE) { in dlm_shuffle_lists() 373 if (!dlm_lock_compatible(lock->ml.type, in dlm_shuffle_lists() 374 target->ml.convert_type)) { in dlm_shuffle_lists() 377 if (lock->ml.highest_blocked == LKM_IVMODE) { in dlm_shuffle_lists() 382 if (lock->ml.highest_blocked < target->ml.convert_type) in dlm_shuffle_lists() 383 lock->ml.highest_blocked = in dlm_shuffle_lists() 384 target->ml.convert_type; in dlm_shuffle_lists() 391 if (!dlm_lock_compatible(lock->ml.type, in dlm_shuffle_lists() 392 target->ml in dlm_shuffle_lists() [all...] |
H A D | dlmrecovery.c | 1049 if (lock->ml.node == dead_node) { in dlm_move_reco_locks_to_list() 1124 struct_size(mres, ml, mres->num_locks), in dlm_send_mig_lockres_msg() 1183 if (lock->ml.type != LKM_EXMODE && lock->ml.type != LKM_PRMODE) in dlm_prepare_lvb_for_migration() 1197 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dlm_prepare_lvb_for_migration() 1198 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in dlm_prepare_lvb_for_migration() 1200 lock->ml.node); in dlm_prepare_lvb_for_migration() 1210 struct dlm_migratable_lock *ml; in dlm_add_lock_to_array() local 1213 ml = &(mres->ml[lock_nu in dlm_add_lock_to_array() 1245 dlm_is_dummy_lock(struct dlm_ctxt *dlm, struct dlm_migratable_lock *ml, u8 *nodenum) dlm_is_dummy_lock() argument 1791 struct dlm_migratable_lock *ml; dlm_process_recovery_data() local [all...] |
H A D | dlmlock.c | 79 if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) in dlm_can_grant_new_lock() 84 if (!dlm_lock_compatible(tmplock->ml.type, lock->ml.type)) in dlm_can_grant_new_lock() 86 if (!dlm_lock_compatible(tmplock->ml.convert_type, in dlm_can_grant_new_lock() 87 lock->ml.type)) in dlm_can_grant_new_lock() 108 mlog(0, "type=%d\n", lock->ml.type); in dlmlock_master() 115 lock->ml.node != dlm->node_num) { in dlmlock_master() 144 lock->ml.node); in dlmlock_master() 155 lock->ml in dlmlock_master() [all...] |
H A D | dlmunlock.c | 218 lock->ml.convert_type = LKM_IVMODE; in dlmunlock_common() 227 BUG_ON(lock->ml.convert_type != LKM_IVMODE); in dlmunlock_common() 229 BUG_ON(lock->ml.convert_type == LKM_IVMODE); in dlmunlock_common() 250 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dlmunlock_common() 251 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in dlmunlock_common() 277 lock->ml.convert_type = LKM_IVMODE; in dlm_commit_pending_cancel() 336 unlock.cookie = lock->ml.cookie; in dlm_send_remote_unlock_request() 464 if (iter->ml.cookie == unlock->cookie && in dlm_unlock_lock_handler() 465 iter->ml.node == unlock->node_idx) { in dlm_unlock_lock_handler() 485 lock->ml in dlm_unlock_lock_handler() [all...] |
H A D | dlmdebug.c | 64 lock->ml.type, lock->ml.convert_type, lock->ml.node, in __dlm_print_lock() 65 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in __dlm_print_lock() 66 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml.cookie)), in __dlm_print_lock() 471 list_type, lock->ml.type, lock->ml.convert_type, in dump_lock() 472 lock->ml.node, in dump_lock() 473 dlm_get_lock_cookie_node(be64_to_cpu(lock->ml.cookie)), in dump_lock() 474 dlm_get_lock_cookie_seq(be64_to_cpu(lock->ml in dump_lock() [all...] |
/kernel/linux/linux-5.10/lib/lz4/ |
H A D | lz4hc_compress.c | 116 size_t ml = 0; in LZ4HC_InsertAndFindBestMatch() local 128 if (*(match + ml) == *(ip + ml) in LZ4HC_InsertAndFindBestMatch() 133 if (mlt > ml) { in LZ4HC_InsertAndFindBestMatch() 134 ml = mlt; in LZ4HC_InsertAndFindBestMatch() 155 if (mlt > ml) { in LZ4HC_InsertAndFindBestMatch() 157 ml = mlt; in LZ4HC_InsertAndFindBestMatch() 165 return (int)ml; in LZ4HC_InsertAndFindBestMatch() 358 int ml, ml2, ml3, ml0; in LZ4HC_compress_generic() local 379 ml in LZ4HC_compress_generic() [all...] |
/kernel/linux/linux-6.6/lib/lz4/ |
H A D | lz4hc_compress.c | 116 size_t ml = 0; in LZ4HC_InsertAndFindBestMatch() local 128 if (*(match + ml) == *(ip + ml) in LZ4HC_InsertAndFindBestMatch() 133 if (mlt > ml) { in LZ4HC_InsertAndFindBestMatch() 134 ml = mlt; in LZ4HC_InsertAndFindBestMatch() 155 if (mlt > ml) { in LZ4HC_InsertAndFindBestMatch() 157 ml = mlt; in LZ4HC_InsertAndFindBestMatch() 165 return (int)ml; in LZ4HC_InsertAndFindBestMatch() 358 int ml, ml2, ml3, ml0; in LZ4HC_compress_generic() local 379 ml in LZ4HC_compress_generic() [all...] |
/kernel/linux/linux-5.10/net/packet/ |
H A D | diag.c | 41 struct packet_mclist *ml; in pdiag_put_mclist() local 48 for (ml = po->mclist; ml; ml = ml->next) { in pdiag_put_mclist() 58 dml->pdmc_index = ml->ifindex; in pdiag_put_mclist() 59 dml->pdmc_type = ml->type; in pdiag_put_mclist() 60 dml->pdmc_alen = ml->alen; in pdiag_put_mclist() 61 dml->pdmc_count = ml->count; in pdiag_put_mclist() 62 BUILD_BUG_ON(sizeof(dml->pdmc_addr) != sizeof(ml in pdiag_put_mclist() [all...] |
/kernel/linux/linux-6.6/net/packet/ |
H A D | diag.c | 41 struct packet_mclist *ml; in pdiag_put_mclist() local 48 for (ml = po->mclist; ml; ml = ml->next) { in pdiag_put_mclist() 58 dml->pdmc_index = ml->ifindex; in pdiag_put_mclist() 59 dml->pdmc_type = ml->type; in pdiag_put_mclist() 60 dml->pdmc_alen = ml->alen; in pdiag_put_mclist() 61 dml->pdmc_count = ml->count; in pdiag_put_mclist() 62 BUILD_BUG_ON(sizeof(dml->pdmc_addr) != sizeof(ml in pdiag_put_mclist() [all...] |
/kernel/linux/linux-5.10/arch/arm/vfp/ |
H A D | vfp.h | 44 static inline void add128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) in add128() argument 51 : "0" (nl), "1" (nh), "r" (ml), "r" (mh) in add128() 57 static inline void sub128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) in sub128() argument 64 : "0" (nl), "1" (nh), "r" (ml), "r" (mh) in sub128() 72 u32 nh, nl, mh, ml; in mul64to128() local 76 ml = m; in mul64to128() 77 rl = (u64)nl * ml; in mul64to128() 80 rma = (u64)nh * ml; in mul64to128() 112 u64 mh, ml, remh, reml, termh, terml, z; in vfp_estimate_div128to64() local 126 ml in vfp_estimate_div128to64() [all...] |
/kernel/linux/linux-6.6/arch/arm/vfp/ |
H A D | vfp.h | 44 static inline void add128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) in add128() argument 51 : "0" (nl), "1" (nh), "r" (ml), "r" (mh) in add128() 57 static inline void sub128(u64 *resh, u64 *resl, u64 nh, u64 nl, u64 mh, u64 ml) in sub128() argument 64 : "0" (nl), "1" (nh), "r" (ml), "r" (mh) in sub128() 72 u32 nh, nl, mh, ml; in mul64to128() local 76 ml = m; in mul64to128() 77 rl = (u64)nl * ml; in mul64to128() 80 rma = (u64)nh * ml; in mul64to128() 112 u64 mh, ml, remh, reml, termh, terml, z; in vfp_estimate_div128to64() local 126 ml in vfp_estimate_div128to64() [all...] |
/kernel/linux/linux-5.10/arch/sh/kernel/cpu/sh2a/ |
H A D | fpu.c | 154 unsigned long long ml, int n) in rshift64() 158 return (mh << (64 - n)) | (ml >> n); in rshift64() 164 unsigned long long mh, ml, nh, nl; in denormal_muld() local 175 mult64(ix, iy, &mh, &ml); in denormal_muld() 177 nl = ml; in denormal_muld() 188 ix = (rshift64(mh, ml, w - 52) & 0x000fffffffffffffLL) in denormal_muld() 191 ix = rshift64(mh, ml, w - 51 - exp) & 0x000fffffffffffffLL; in denormal_muld() 153 rshift64(unsigned long long mh, unsigned long long ml, int n) rshift64() argument
|
/kernel/linux/linux-6.6/arch/sh/kernel/cpu/sh2a/ |
H A D | fpu.c | 154 unsigned long long ml, int n) in rshift64() 158 return (mh << (64 - n)) | (ml >> n); in rshift64() 164 unsigned long long mh, ml, nh, nl; in denormal_muld() local 175 mult64(ix, iy, &mh, &ml); in denormal_muld() 177 nl = ml; in denormal_muld() 188 ix = (rshift64(mh, ml, w - 52) & 0x000fffffffffffffLL) in denormal_muld() 191 ix = rshift64(mh, ml, w - 51 - exp) & 0x000fffffffffffffLL; in denormal_muld() 153 rshift64(unsigned long long mh, unsigned long long ml, int n) rshift64() argument
|
/kernel/linux/linux-6.6/lib/zstd/compress/ |
H A D | zstd_lazy.c | 513 * If none are longer than the argument ml, then ml will be returned. 516 size_t ZSTD_dedicatedDictSearch_lazy_search(size_t* offsetPtr, size_t ml, U32 nbAttempts, in ZSTD_dedicatedDictSearch_lazy_search() argument 549 return ml; in ZSTD_dedicatedDictSearch_lazy_search() 562 if (currentMl > ml) { in ZSTD_dedicatedDictSearch_lazy_search() 563 ml = currentMl; in ZSTD_dedicatedDictSearch_lazy_search() 567 return ml; in ZSTD_dedicatedDictSearch_lazy_search() 599 if (currentMl > ml) { in ZSTD_dedicatedDictSearch_lazy_search() 600 ml = currentMl; in ZSTD_dedicatedDictSearch_lazy_search() 606 return ml; in ZSTD_dedicatedDictSearch_lazy_search() 671 size_t ml=4-1; ZSTD_HcFindBestMatch() local 1147 size_t ml=4-1; ZSTD_RowFindBestMatch() local [all...] |