Lines Matching defs:wmm
13 #include "wmm.h"
188 &priv->wmm.tid_tbl_ptr[i].ra_list);
198 priv->wmm.queue_priority[0] = WMM_AC_VO;
199 priv->wmm.queue_priority[1] = WMM_AC_VI;
200 priv->wmm.queue_priority[2] = WMM_AC_BE;
201 priv->wmm.queue_priority[3] = WMM_AC_BK;
210 struct mwifiex_wmm_desc *wmm = &priv->wmm;
211 u8 *queue_priority = wmm->queue_priority;
222 atomic_set(&wmm->highest_queued_prio, HIGH_PRIO_TID);
256 priv->wmm.queue_priority[ac_idx] = ac_idx;
271 swap(priv->wmm.queue_priority[j - 1],
272 priv->wmm.queue_priority[j]);
274 if (priv->wmm.queue_priority[j - 1]
275 < priv->wmm.queue_priority[j])
276 swap(priv->wmm.queue_priority[j - 1],
277 priv->wmm.queue_priority[j]);
299 ac_status = &priv->wmm.ac_status[eval_ac];
316 ac_status = &priv->wmm.ac_status[down_ac];
341 priv->wmm.ac_down_graded_vals[ac_val] =
345 priv->wmm.ac_down_graded_vals[ac_val]
351 priv->wmm.ac_down_graded_vals[ac_val]);
393 ac_down = priv->wmm.ac_down_graded_vals[ac];
444 priv->wmm.drv_pkt_delay_max = MWIFIEX_WMM_DRV_DELAY_MAX;
445 atomic_set(&priv->wmm.tx_pkts_queued, 0);
446 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
488 if (atomic_read(&priv->wmm.tx_pkts_queued))
539 mwifiex_wmm_del_pkts_in_ralist(priv, &priv->wmm.tid_tbl_ptr[i].
542 atomic_set(&priv->wmm.tx_pkts_queued, 0);
543 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
558 &priv->wmm.tid_tbl_ptr[i].ra_list,
564 INIT_LIST_HEAD(&priv->wmm.tid_tbl_ptr[i].ra_list);
591 spin_lock_bh(&priv->wmm.ra_list_spinlock);
605 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
632 list_for_each_entry(ra_list, &priv->wmm.tid_tbl_ptr[tid].ra_list,
648 spin_lock_bh(&priv->wmm.ra_list_spinlock);
656 priv->wmm.pkts_paused[i] +=
659 priv->wmm.pkts_paused[i] -=
665 tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued);
671 atomic_set(&priv->wmm.tx_pkts_queued, tx_pkts_queued);
672 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
674 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
687 spin_lock_bh(&priv->wmm.ra_list_spinlock);
690 list_for_each_entry(ra_list, &priv->wmm.tid_tbl_ptr[i].ra_list,
699 priv->wmm.pkts_paused[i] +=
702 priv->wmm.pkts_paused[i] -=
709 tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued);
715 atomic_set(&priv->wmm.tx_pkts_queued, tx_pkts_queued);
716 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
718 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
752 spin_lock_bh(&priv->wmm.ra_list_spinlock);
761 priv->wmm.pkts_paused[i] -= ra_list->total_pkt_count;
764 &priv->wmm.tx_pkts_queued);
768 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
781 list_for_each_entry(rlist, &priv->wmm.tid_tbl_ptr[ptr_index].ra_list,
844 spin_lock_bh(&priv->wmm.ra_list_spinlock);
864 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
867 list_head = priv->wmm.tid_tbl_ptr[tid_down].ra_list;
880 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
890 if (atomic_read(&priv->wmm.highest_queued_prio) <
892 atomic_set(&priv->wmm.highest_queued_prio,
896 priv->wmm.pkts_paused[tid_down]++;
898 atomic_inc(&priv->wmm.tx_pkts_queued);
900 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
950 ac_status = &priv->wmm.ac_status[tlv_wmm_qstatus->
1076 ret_val = (u8) (min(queue_delay, priv->wmm.drv_pkt_delay_max) >> 1);
1109 (atomic_read(&priv_tmp->wmm.tx_pkts_queued) == 0))
1117 hqp = &priv_tmp->wmm.highest_queued_prio;
1120 spin_lock_bh(&priv_tmp->wmm.ra_list_spinlock);
1122 tid_ptr = &(priv_tmp)->wmm.
1135 spin_unlock_bh(&priv_tmp->wmm.ra_list_spinlock);
1138 if (atomic_read(&priv_tmp->wmm.tx_pkts_queued) != 0) {
1139 atomic_set(&priv_tmp->wmm.highest_queued_prio,
1146 atomic_set(&priv_tmp->wmm.highest_queued_prio,
1157 spin_unlock_bh(&priv_tmp->wmm.ra_list_spinlock);
1172 * Function also increments wmm.packets_out counter.
1180 struct mwifiex_tid_tbl *tid_ptr = &priv->wmm.tid_tbl_ptr[tid];
1191 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1193 priv->wmm.packets_out[tid]++;
1197 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1235 __releases(&priv->wmm.ra_list_spinlock)
1243 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1261 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1268 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1271 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1281 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1284 atomic_dec(&priv->wmm.tx_pkts_queued);
1318 __releases(&priv->wmm.ra_list_spinlock)
1327 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1335 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1337 atomic_dec(&priv->wmm.tx_pkts_queued);
1349 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1365 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1368 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1376 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1393 atomic_dec(&priv->wmm.tx_pkts_queued);
1394 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1396 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1421 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1423 spin_unlock_bh(&priv->wmm.ra_list_spinlock);