/kernel/linux/linux-5.10/net/netfilter/ |
H A D | nft_limit.c | 28 static inline bool nft_limit_eval(struct nft_limit *limit, u64 cost) in nft_limit_eval() argument 33 spin_lock_bh(&limit->lock); in nft_limit_eval() 35 tokens = limit->tokens + now - limit->last; in nft_limit_eval() 36 if (tokens > limit->tokens_max) in nft_limit_eval() 37 tokens = limit->tokens_max; in nft_limit_eval() 39 limit->last = now; in nft_limit_eval() 42 limit->tokens = delta; in nft_limit_eval() 43 spin_unlock_bh(&limit->lock); in nft_limit_eval() 44 return limit in nft_limit_eval() 54 nft_limit_init(struct nft_limit *limit, const struct nlattr * const tb[], bool pkts) nft_limit_init() argument 104 nft_limit_dump(struct sk_buff *skb, const struct nft_limit *limit, enum nft_limit_type type) nft_limit_dump() argument 125 struct nft_limit limit; global() member [all...] |
/kernel/linux/linux-6.6/arch/arm64/lib/ |
H A D | memcmp.S | 22 #define limit x2 define 36 subs limit, limit, 8 44 subs limit, limit, 8 47 ldr data1, [src1, limit] 48 ldr data2, [src2, limit] 59 subs limit, limit, 16 63 try to align, so limit i [all...] |
H A D | strncmp.S | 26 #define limit x2 define 62 cbz limit, L(ret0) 78 subs limit, limit, #8 91 add limit, limit, 8 /* Rewind limit to before last subs. */ 94 is before the limit. */ 100 cmp limit, pos, lsr #3 109 /* Not reached the limit, mus [all...] |
/kernel/linux/linux-5.10/arch/x86/mm/ |
H A D | amdtopology.c | 83 u64 base, limit; in amd_numa_init() local 86 limit = read_pci_config(0, nb, 1, 0x44 + i*8); in amd_numa_init() 88 nodeids[i] = nodeid = limit & 7; in amd_numa_init() 96 base, limit); in amd_numa_init() 100 if (!limit) { in amd_numa_init() 105 if ((base >> 8) & 3 || (limit >> 8) & 3) { in amd_numa_init() 107 nodeid, (base >> 8) & 3, (limit >> 8) & 3); in amd_numa_init() 116 limit >>= 16; in amd_numa_init() 117 limit++; in amd_numa_init() 118 limit << in amd_numa_init() [all...] |
/kernel/linux/linux-6.6/arch/x86/mm/ |
H A D | amdtopology.c | 83 u64 base, limit; in amd_numa_init() local 86 limit = read_pci_config(0, nb, 1, 0x44 + i*8); in amd_numa_init() 88 nodeids[i] = nodeid = limit & 7; in amd_numa_init() 96 base, limit); in amd_numa_init() 100 if (!limit) { in amd_numa_init() 105 if ((base >> 8) & 3 || (limit >> 8) & 3) { in amd_numa_init() 107 nodeid, (base >> 8) & 3, (limit >> 8) & 3); in amd_numa_init() 116 limit >>= 16; in amd_numa_init() 117 limit++; in amd_numa_init() 118 limit << in amd_numa_init() [all...] |
/kernel/linux/linux-5.10/arch/arm64/lib/ |
H A D | memcmp.S | 31 limit .req x2 label 50 cbz limit, .Lret0 56 sub limit_wd, limit, #1 /* limit != 0, so no underflow. */ 71 /* Not reached the limit, must have found a diff. */ 75 ands limit, limit, #7 81 lsl limit, limit, #3 /* bytes-> bits. */ 83 CPU_BE( lsr mask, mask, limit ) [all...] |
H A D | strncmp.S | 35 limit .req x2 label 56 cbz limit, .Lret0 65 * when limit is mulitply of 8, if not sub 1, 68 sub limit_wd, limit, #1 /* limit != 0, so no underflow. */ 89 /*Not reached the limit, must have found the end or a diff. */ 93 ands limit, limit, #7 96 lsl limit, limit, # [all...] |
/kernel/linux/linux-5.10/lib/ |
H A D | dynamic_queue_limits.c | 17 /* Records completed count and recalculates the queue limit */ 20 unsigned int inprogress, prev_inprogress, limit; in dql_completed() local 30 limit = dql->limit; in dql_completed() 31 ovlimit = POSDIFF(num_queued - dql->num_completed, limit); in dql_completed() 40 * - The queue was over-limit in the last interval, in dql_completed() 43 * - The queue was over-limit in the previous interval and in dql_completed() 49 * When queue is starved increase the limit by the amount in dql_completed() 51 * plus any previous over-limit. in dql_completed() 53 limit in dql_completed() [all...] |
/kernel/linux/linux-6.6/lib/ |
H A D | dynamic_queue_limits.c | 17 /* Records completed count and recalculates the queue limit */ 20 unsigned int inprogress, prev_inprogress, limit; in dql_completed() local 30 limit = dql->limit; in dql_completed() 31 ovlimit = POSDIFF(num_queued - dql->num_completed, limit); in dql_completed() 40 * - The queue was over-limit in the last interval, in dql_completed() 43 * - The queue was over-limit in the previous interval and in dql_completed() 49 * When queue is starved increase the limit by the amount in dql_completed() 51 * plus any previous over-limit. in dql_completed() 53 limit in dql_completed() [all...] |
/kernel/linux/linux-6.6/net/netfilter/ |
H A D | nft_limit.c | 24 struct nft_limit *limit; member 37 spin_lock_bh(&priv->limit->lock); in nft_limit_eval() 39 tokens = priv->limit->tokens + now - priv->limit->last; in nft_limit_eval() 43 priv->limit->last = now; in nft_limit_eval() 46 priv->limit->tokens = delta; in nft_limit_eval() 47 spin_unlock_bh(&priv->limit->lock); in nft_limit_eval() 50 priv->limit->tokens = tokens; in nft_limit_eval() 51 spin_unlock_bh(&priv->limit->lock); in nft_limit_eval() 113 priv->limit in nft_limit_init() 173 struct nft_limit_priv limit; global() member [all...] |
/kernel/linux/linux-5.10/arch/powerpc/kernel/ |
H A D | paca.c | 28 unsigned long limit, int cpu) in alloc_paca_data() 46 limit, nid); in alloc_paca_data() 60 static void *__init alloc_shared_lppaca(unsigned long size, unsigned long limit, in alloc_shared_lppaca() argument 81 limit, NUMA_NO_NODE); in alloc_shared_lppaca() 122 static struct lppaca * __init new_lppaca(int cpu, unsigned long limit) in new_lppaca() argument 132 lp = alloc_shared_lppaca(LPPACA_SIZE, limit, cpu); in new_lppaca() 134 lp = alloc_paca_data(LPPACA_SIZE, 0x400, limit, cpu); in new_lppaca() 151 static struct slb_shadow * __init new_slb_shadow(int cpu, unsigned long limit) in new_slb_shadow() argument 165 s = alloc_paca_data(sizeof(*s), L1_CACHE_BYTES, limit, cpu); in new_slb_shadow() 179 * @limit 27 alloc_paca_data(unsigned long size, unsigned long align, unsigned long limit, int cpu) alloc_paca_data() argument 187 new_rtas_args(int cpu, unsigned long limit) new_rtas_args() argument 287 u64 limit; allocate_paca() local [all...] |
/kernel/linux/linux-5.10/drivers/clk/bcm/ |
H A D | clk-kona-setup.c | 29 u32 limit; in ccu_data_offsets_valid() local 31 limit = ccu->range - sizeof(u32); in ccu_data_offsets_valid() 32 limit = round_down(limit, sizeof(u32)); in ccu_data_offsets_valid() 34 if (ccu_policy->enable.offset > limit) { in ccu_data_offsets_valid() 37 ccu->name, ccu_policy->enable.offset, limit); in ccu_data_offsets_valid() 40 if (ccu_policy->control.offset > limit) { in ccu_data_offsets_valid() 43 ccu->name, ccu_policy->control.offset, limit); in ccu_data_offsets_valid() 88 u32 limit; in peri_clk_data_offsets_valid() local 95 limit in peri_clk_data_offsets_valid() 182 u32 limit = BITS_PER_BYTE * sizeof(u32) - 1; bit_posn_valid() local 202 u32 limit = BITS_PER_BYTE * sizeof(u32); bitfield_valid() local 299 u32 limit; sel_valid() local 373 u32 limit; kona_dividers_valid() local [all...] |
/kernel/linux/linux-6.6/drivers/clk/bcm/ |
H A D | clk-kona-setup.c | 21 u32 limit; in ccu_data_offsets_valid() local 23 limit = ccu->range - sizeof(u32); in ccu_data_offsets_valid() 24 limit = round_down(limit, sizeof(u32)); in ccu_data_offsets_valid() 26 if (ccu_policy->enable.offset > limit) { in ccu_data_offsets_valid() 29 ccu->name, ccu_policy->enable.offset, limit); in ccu_data_offsets_valid() 32 if (ccu_policy->control.offset > limit) { in ccu_data_offsets_valid() 35 ccu->name, ccu_policy->control.offset, limit); in ccu_data_offsets_valid() 80 u32 limit; in peri_clk_data_offsets_valid() local 87 limit in peri_clk_data_offsets_valid() 174 u32 limit = BITS_PER_BYTE * sizeof(u32) - 1; bit_posn_valid() local 194 u32 limit = BITS_PER_BYTE * sizeof(u32); bitfield_valid() local 291 u32 limit; sel_valid() local 365 u32 limit; kona_dividers_valid() local [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/display/ |
H A D | intel_dpll.c | 372 const struct intel_limit *limit, in intel_pll_is_valid() 375 if (clock->n < limit->n.min || limit->n.max < clock->n) in intel_pll_is_valid() 377 if (clock->p1 < limit->p1.min || limit->p1.max < clock->p1) in intel_pll_is_valid() 379 if (clock->m2 < limit->m2.min || limit->m2.max < clock->m2) in intel_pll_is_valid() 381 if (clock->m1 < limit->m1.min || limit->m1.max < clock->m1) in intel_pll_is_valid() 389 if (clock->p < limit in intel_pll_is_valid() 371 intel_pll_is_valid(struct drm_i915_private *dev_priv, const struct intel_limit *limit, const struct dpll *clock) intel_pll_is_valid() argument 407 i9xx_select_p2_div(const struct intel_limit *limit, const struct intel_crtc_state *crtc_state, int target) i9xx_select_p2_div() argument 441 i9xx_find_best_dpll(const struct intel_limit *limit, struct intel_crtc_state *crtc_state, int target, int refclk, const struct dpll *match_clock, struct dpll *best_clock) i9xx_find_best_dpll() argument 499 pnv_find_best_dpll(const struct intel_limit *limit, struct intel_crtc_state *crtc_state, int target, int refclk, const struct dpll *match_clock, struct dpll *best_clock) pnv_find_best_dpll() argument 555 g4x_find_best_dpll(const struct intel_limit *limit, struct intel_crtc_state *crtc_state, int target, int refclk, const struct dpll *match_clock, struct dpll *best_clock) g4x_find_best_dpll() argument 649 vlv_find_best_dpll(const struct intel_limit *limit, struct intel_crtc_state *crtc_state, int target, int refclk, const struct dpll *match_clock, struct dpll *best_clock) vlv_find_best_dpll() argument 707 chv_find_best_dpll(const struct intel_limit *limit, struct intel_crtc_state *crtc_state, int target, int refclk, const struct dpll *match_clock, struct dpll *best_clock) chv_find_best_dpll() argument 768 const struct intel_limit *limit = &intel_limits_bxt; bxt_find_best_dpll() local 1146 const struct intel_limit *limit; ilk_crtc_compute_clock() local 1248 const struct intel_limit *limit = &intel_limits_chv; chv_crtc_compute_clock() local 1273 const struct intel_limit *limit = &intel_limits_vlv; vlv_crtc_compute_clock() local 1300 const struct intel_limit *limit; g4x_crtc_compute_clock() local 1347 const struct intel_limit *limit; pnv_crtc_compute_clock() local 1383 const struct intel_limit *limit; i9xx_crtc_compute_clock() local 1421 const struct intel_limit *limit; i8xx_crtc_compute_clock() local [all...] |
/kernel/linux/linux-5.10/fs/romfs/ |
H A D | storage.c | 131 unsigned long pos, size_t limit) in romfs_blk_strnlen() 140 while (limit > 0) { in romfs_blk_strnlen() 142 segment = min_t(size_t, limit, ROMBSIZE - offset); in romfs_blk_strnlen() 151 limit -= segment; in romfs_blk_strnlen() 217 size_t limit; in romfs_dev_read() local 219 limit = romfs_maxsize(sb); in romfs_dev_read() 220 if (pos >= limit || buflen > limit - pos) in romfs_dev_read() 240 size_t limit; in romfs_dev_strnlen() local 242 limit in romfs_dev_strnlen() 130 romfs_blk_strnlen(struct super_block *sb, unsigned long pos, size_t limit) romfs_blk_strnlen() argument 268 size_t limit; romfs_dev_strcmp() local [all...] |
/kernel/linux/linux-6.6/fs/romfs/ |
H A D | storage.c | 131 unsigned long pos, size_t limit) in romfs_blk_strnlen() 140 while (limit > 0) { in romfs_blk_strnlen() 142 segment = min_t(size_t, limit, ROMBSIZE - offset); in romfs_blk_strnlen() 151 limit -= segment; in romfs_blk_strnlen() 217 size_t limit; in romfs_dev_read() local 219 limit = romfs_maxsize(sb); in romfs_dev_read() 220 if (pos >= limit || buflen > limit - pos) in romfs_dev_read() 240 size_t limit; in romfs_dev_strnlen() local 242 limit in romfs_dev_strnlen() 130 romfs_blk_strnlen(struct super_block *sb, unsigned long pos, size_t limit) romfs_blk_strnlen() argument 268 size_t limit; romfs_dev_strcmp() local [all...] |
/kernel/linux/linux-5.10/kernel/cgroup/ |
H A D | pids.c | 6 * after a certain limit is reached. 8 * Since it is trivial to hit the task limit without hitting any kmemcg limits 21 * To set a cgroup to have no limit, set pids.max to "max". This is the default 23 * stringent limit in the hierarchy is followed). 49 atomic64_t limit; member 54 /* Number of times fork failed because limit was hit. */ 78 atomic64_set(&pids->limit, PIDS_MAX); in pids_css_alloc() 123 * This function does *not* follow the pid limit set. It cannot fail and the new 124 * pid count may exceed the limit. This is only used for reverting failed 125 * attaches, where there is no other way out than violating the limit 150 int64_t limit = atomic64_read(&p->limit); pids_try_charge() local 267 int64_t limit; pids_max_write() local 296 int64_t limit = atomic64_read(&pids->limit); pids_max_show() local [all...] |
/kernel/liteos_a/kernel/extended/plimit/ |
H A D | los_memlimit.c | 44 procMemLimiter->limit = OS_NULL_INT; in OsMemLimiterInit() 48 VOID OsMemLimitSetLimit(UINT32 limit) in OsMemLimitSetLimit() argument 50 g_procMemLimiter->limit = limit; in OsMemLimitSetLimit() 77 plimiteDest->limit = plimiteSrc->limit; in OsMemLimiterCopy() 85 if ((currMemLimit->usage + parentMemLimit->usage) >= parentMemLimit->limit) { in MemLimiteMigrateCheck() 110 BOOL OsMemLimitAddProcessCheck(UINTPTR limit, UINTPTR process) in OsMemLimitAddProcessCheck() argument 112 ProcMemLimiter *memLimit = (ProcMemLimiter *)limit; in OsMemLimitAddProcessCheck() 114 if ((memLimit->usage + pcb->limitStat.memUsed) > memLimit->limit) { in OsMemLimitAddProcessCheck() 120 OsMemLimitAddProcess(UINTPTR limit, UINTPTR process) OsMemLimitAddProcess() argument 131 OsMemLimitDelProcess(UINTPTR limit, UINTPTR process) OsMemLimitDelProcess() argument [all...] |
H A D | los_processlimit.c | 41 VOID PidLimiterInit(UINTPTR limit) in PidLimiterInit() argument 43 PidLimit *pidLimit = (PidLimit *)limit; in PidLimiterInit() 63 VOID PidLimterFree(UINTPTR limit) in PidLimterFree() argument 65 PidLimit *pidLimit = (PidLimit *)limit; in PidLimterFree() 70 (VOID)LOS_MemFree(m_aucSysMem1, (VOID *)limit); in PidLimterFree() 87 BOOL OsPidLimitAddProcessCheck(UINTPTR limit, UINTPTR process) in OsPidLimitAddProcessCheck() argument 90 PidLimit *pidLimit = (PidLimit *)limit; in OsPidLimitAddProcessCheck() 97 VOID OsPidLimitAddProcess(UINTPTR limit, UINTPTR process) in OsPidLimitAddProcess() argument 100 PidLimit *plimits = (PidLimit *)limit; in OsPidLimitAddProcess() 106 VOID OsPidLimitDelProcess(UINTPTR limit, UINTPT argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/amd/pm/powerplay/ |
H A D | kv_smc.c | 78 u32 smc_address, u32 limit) in kv_set_smc_sram_address() 82 if ((smc_address + 3) > limit) in kv_set_smc_sram_address() 93 u32 *value, u32 limit) in amdgpu_kv_read_smc_sram_dword() 97 ret = kv_set_smc_sram_address(adev, smc_address, limit); in amdgpu_kv_read_smc_sram_dword() 123 const u8 *src, u32 byte_count, u32 limit) in amdgpu_kv_copy_bytes_to_smc() 128 if ((smc_start_address + byte_count) > limit) in amdgpu_kv_copy_bytes_to_smc() 138 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 164 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 177 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 192 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 77 kv_set_smc_sram_address(struct amdgpu_device *adev, u32 smc_address, u32 limit) kv_set_smc_sram_address() argument 92 amdgpu_kv_read_smc_sram_dword(struct amdgpu_device *adev, u32 smc_address, u32 *value, u32 limit) amdgpu_kv_read_smc_sram_dword() argument 121 amdgpu_kv_copy_bytes_to_smc(struct amdgpu_device *adev, u32 smc_start_address, const u8 *src, u32 byte_count, u32 limit) amdgpu_kv_copy_bytes_to_smc() argument [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/radeon/ |
H A D | kv_smc.c | 75 u32 smc_address, u32 limit) in kv_set_smc_sram_address() 79 if ((smc_address + 3) > limit) in kv_set_smc_sram_address() 89 u32 *value, u32 limit) in kv_read_smc_sram_dword() 93 ret = kv_set_smc_sram_address(rdev, smc_address, limit); in kv_read_smc_sram_dword() 119 const u8 *src, u32 byte_count, u32 limit) in kv_copy_bytes_to_smc() 124 if ((smc_start_address + byte_count) > limit) in kv_copy_bytes_to_smc() 134 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 160 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 173 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 188 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 74 kv_set_smc_sram_address(struct radeon_device *rdev, u32 smc_address, u32 limit) kv_set_smc_sram_address() argument 88 kv_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, u32 *value, u32 limit) kv_read_smc_sram_dword() argument 117 kv_copy_bytes_to_smc(struct radeon_device *rdev, u32 smc_start_address, const u8 *src, u32 byte_count, u32 limit) kv_copy_bytes_to_smc() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/amd/pm/legacy-dpm/ |
H A D | kv_smc.c | 78 u32 smc_address, u32 limit) in kv_set_smc_sram_address() 82 if ((smc_address + 3) > limit) in kv_set_smc_sram_address() 93 u32 *value, u32 limit) in amdgpu_kv_read_smc_sram_dword() 97 ret = kv_set_smc_sram_address(adev, smc_address, limit); in amdgpu_kv_read_smc_sram_dword() 123 const u8 *src, u32 byte_count, u32 limit) in amdgpu_kv_copy_bytes_to_smc() 128 if ((smc_start_address + byte_count) > limit) in amdgpu_kv_copy_bytes_to_smc() 138 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 164 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 177 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 192 ret = kv_set_smc_sram_address(adev, addr, limit); in amdgpu_kv_copy_bytes_to_smc() 77 kv_set_smc_sram_address(struct amdgpu_device *adev, u32 smc_address, u32 limit) kv_set_smc_sram_address() argument 92 amdgpu_kv_read_smc_sram_dword(struct amdgpu_device *adev, u32 smc_address, u32 *value, u32 limit) amdgpu_kv_read_smc_sram_dword() argument 121 amdgpu_kv_copy_bytes_to_smc(struct amdgpu_device *adev, u32 smc_start_address, const u8 *src, u32 byte_count, u32 limit) amdgpu_kv_copy_bytes_to_smc() argument [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/radeon/ |
H A D | kv_smc.c | 75 u32 smc_address, u32 limit) in kv_set_smc_sram_address() 79 if ((smc_address + 3) > limit) in kv_set_smc_sram_address() 89 u32 *value, u32 limit) in kv_read_smc_sram_dword() 93 ret = kv_set_smc_sram_address(rdev, smc_address, limit); in kv_read_smc_sram_dword() 119 const u8 *src, u32 byte_count, u32 limit) in kv_copy_bytes_to_smc() 124 if ((smc_start_address + byte_count) > limit) in kv_copy_bytes_to_smc() 134 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 160 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 173 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 188 ret = kv_set_smc_sram_address(rdev, addr, limit); in kv_copy_bytes_to_smc() 74 kv_set_smc_sram_address(struct radeon_device *rdev, u32 smc_address, u32 limit) kv_set_smc_sram_address() argument 88 kv_read_smc_sram_dword(struct radeon_device *rdev, u32 smc_address, u32 *value, u32 limit) kv_read_smc_sram_dword() argument 117 kv_copy_bytes_to_smc(struct radeon_device *rdev, u32 smc_start_address, const u8 *src, u32 byte_count, u32 limit) kv_copy_bytes_to_smc() argument [all...] |
/kernel/linux/linux-6.6/arch/powerpc/kernel/ |
H A D | paca.c | 27 unsigned long limit, int cpu) in alloc_paca_data() 45 limit, nid); in alloc_paca_data() 59 static void *__init alloc_shared_lppaca(unsigned long size, unsigned long limit, in alloc_shared_lppaca() argument 80 limit, NUMA_NO_NODE); in alloc_shared_lppaca() 121 static struct lppaca * __init new_lppaca(int cpu, unsigned long limit) in new_lppaca() argument 131 lp = alloc_shared_lppaca(LPPACA_SIZE, limit, cpu); in new_lppaca() 133 lp = alloc_paca_data(LPPACA_SIZE, 0x400, limit, cpu); in new_lppaca() 149 static struct slb_shadow * __init new_slb_shadow(int cpu, unsigned long limit) in new_slb_shadow() argument 163 s = alloc_paca_data(sizeof(*s), L1_CACHE_BYTES, limit, cpu); in new_slb_shadow() 258 u64 limit; in allocate_paca() local 26 alloc_paca_data(unsigned long size, unsigned long align, unsigned long limit, int cpu) alloc_paca_data() argument [all...] |
/kernel/linux/linux-5.10/net/sched/ |
H A D | sch_fifo.c | 22 if (likely(sch->qstats.backlog + qdisc_pkt_len(skb) <= sch->limit)) in bfifo_enqueue() 31 if (likely(sch->q.qlen < sch->limit)) in pfifo_enqueue() 42 if (likely(sch->q.qlen < sch->limit)) in pfifo_tail_enqueue() 46 /* queue full, remove one skb to fulfill the limit */ in pfifo_tail_enqueue() 103 u32 limit = qdisc_dev(sch)->tx_queue_len; in __fifo_init() local 106 limit *= psched_mtu(qdisc_dev(sch)); in __fifo_init() 108 sch->limit = limit; in __fifo_init() 115 sch->limit = ctl->limit; in __fifo_init() 227 fifo_set_limit(struct Qdisc *q, unsigned int limit) fifo_set_limit() argument 252 fifo_create_dflt(struct Qdisc *sch, struct Qdisc_ops *ops, unsigned int limit, struct netlink_ext_ack *extack) fifo_create_dflt() argument [all...] |