Lines Matching refs:wmm
25 #include "wmm.h"
200 &priv->wmm.tid_tbl_ptr[i].ra_list);
210 priv->wmm.queue_priority[0] = WMM_AC_VO;
211 priv->wmm.queue_priority[1] = WMM_AC_VI;
212 priv->wmm.queue_priority[2] = WMM_AC_BE;
213 priv->wmm.queue_priority[3] = WMM_AC_BK;
222 struct mwifiex_wmm_desc *wmm = &priv->wmm;
223 u8 *queue_priority = wmm->queue_priority;
234 atomic_set(&wmm->highest_queued_prio, HIGH_PRIO_TID);
268 priv->wmm.queue_priority[ac_idx] = ac_idx;
283 swap(priv->wmm.queue_priority[j - 1],
284 priv->wmm.queue_priority[j]);
286 if (priv->wmm.queue_priority[j - 1]
287 < priv->wmm.queue_priority[j])
288 swap(priv->wmm.queue_priority[j - 1],
289 priv->wmm.queue_priority[j]);
311 ac_status = &priv->wmm.ac_status[eval_ac];
328 ac_status = &priv->wmm.ac_status[down_ac];
353 priv->wmm.ac_down_graded_vals[ac_val] =
357 priv->wmm.ac_down_graded_vals[ac_val]
363 priv->wmm.ac_down_graded_vals[ac_val]);
405 ac_down = priv->wmm.ac_down_graded_vals[ac];
456 priv->wmm.drv_pkt_delay_max = MWIFIEX_WMM_DRV_DELAY_MAX;
457 atomic_set(&priv->wmm.tx_pkts_queued, 0);
458 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
500 if (atomic_read(&priv->wmm.tx_pkts_queued))
551 mwifiex_wmm_del_pkts_in_ralist(priv, &priv->wmm.tid_tbl_ptr[i].
554 atomic_set(&priv->wmm.tx_pkts_queued, 0);
555 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
570 &priv->wmm.tid_tbl_ptr[i].ra_list,
576 INIT_LIST_HEAD(&priv->wmm.tid_tbl_ptr[i].ra_list);
603 spin_lock_bh(&priv->wmm.ra_list_spinlock);
617 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
644 list_for_each_entry(ra_list, &priv->wmm.tid_tbl_ptr[tid].ra_list,
660 spin_lock_bh(&priv->wmm.ra_list_spinlock);
668 priv->wmm.pkts_paused[i] +=
671 priv->wmm.pkts_paused[i] -=
677 tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued);
683 atomic_set(&priv->wmm.tx_pkts_queued, tx_pkts_queued);
684 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
686 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
699 spin_lock_bh(&priv->wmm.ra_list_spinlock);
702 list_for_each_entry(ra_list, &priv->wmm.tid_tbl_ptr[i].ra_list,
711 priv->wmm.pkts_paused[i] +=
714 priv->wmm.pkts_paused[i] -=
721 tx_pkts_queued = atomic_read(&priv->wmm.tx_pkts_queued);
727 atomic_set(&priv->wmm.tx_pkts_queued, tx_pkts_queued);
728 atomic_set(&priv->wmm.highest_queued_prio, HIGH_PRIO_TID);
730 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
764 spin_lock_bh(&priv->wmm.ra_list_spinlock);
773 priv->wmm.pkts_paused[i] -= ra_list->total_pkt_count;
776 &priv->wmm.tx_pkts_queued);
780 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
793 list_for_each_entry(rlist, &priv->wmm.tid_tbl_ptr[ptr_index].ra_list,
856 spin_lock_bh(&priv->wmm.ra_list_spinlock);
876 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
879 list_head = priv->wmm.tid_tbl_ptr[tid_down].ra_list;
892 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
902 if (atomic_read(&priv->wmm.highest_queued_prio) <
904 atomic_set(&priv->wmm.highest_queued_prio,
908 priv->wmm.pkts_paused[tid_down]++;
910 atomic_inc(&priv->wmm.tx_pkts_queued);
912 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
962 ac_status = &priv->wmm.ac_status[tlv_wmm_qstatus->
1088 ret_val = (u8) (min(queue_delay, priv->wmm.drv_pkt_delay_max) >> 1);
1121 (atomic_read(&priv_tmp->wmm.tx_pkts_queued) == 0))
1129 hqp = &priv_tmp->wmm.highest_queued_prio;
1132 spin_lock_bh(&priv_tmp->wmm.ra_list_spinlock);
1134 tid_ptr = &(priv_tmp)->wmm.
1147 spin_unlock_bh(&priv_tmp->wmm.ra_list_spinlock);
1150 if (atomic_read(&priv_tmp->wmm.tx_pkts_queued) != 0) {
1151 atomic_set(&priv_tmp->wmm.highest_queued_prio,
1158 atomic_set(&priv_tmp->wmm.highest_queued_prio,
1169 spin_unlock_bh(&priv_tmp->wmm.ra_list_spinlock);
1184 * Function also increments wmm.packets_out counter.
1192 struct mwifiex_tid_tbl *tid_ptr = &priv->wmm.tid_tbl_ptr[tid];
1203 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1205 priv->wmm.packets_out[tid]++;
1209 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1247 __releases(&priv->wmm.ra_list_spinlock)
1255 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1273 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1280 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1283 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1293 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1296 atomic_dec(&priv->wmm.tx_pkts_queued);
1330 __releases(&priv->wmm.ra_list_spinlock)
1339 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1347 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1349 atomic_dec(&priv->wmm.tx_pkts_queued);
1361 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1377 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1380 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1388 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1404 atomic_dec(&priv->wmm.tx_pkts_queued);
1405 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1407 spin_unlock_bh(&priv->wmm.ra_list_spinlock);
1432 spin_lock_bh(&priv->wmm.ra_list_spinlock);
1434 spin_unlock_bh(&priv->wmm.ra_list_spinlock);