Lines Matching defs:offset
2357 * offset/count pair it will no longer be used. In the worst case TC0
2368 if (tc->offset + tc->count > txq) {
2379 if (tc->offset + tc->count > txq) {
2395 if ((txq - tc->offset) < tc->count)
2445 int cpu, u16 offset, u16 count)
2454 for (i = count, j = offset; i--; j++) {
2479 u16 offset, u16 count)
2490 active |= remove_xps_queue_cpu(dev, dev_maps, j, offset, count);
2495 for (i = offset + (count - 1); count--; i--)
2501 static void netif_reset_xps_queues(struct net_device *dev, u16 offset,
2511 clean_xps_maps(dev, XPS_RXQS, offset, count);
2513 clean_xps_maps(dev, XPS_CPUS, offset, count);
2827 int netdev_set_tc_queue(struct net_device *dev, u8 tc, u16 count, u16 offset)
2833 netif_reset_xps_queues(dev, offset, count);
2836 dev->tc_to_txq[tc].offset = offset;
2876 u8 tc, u16 count, u16 offset)
2883 if ((offset + count) > dev->real_num_tx_queues)
2888 sb_dev->tc_to_txq[tc].offset = offset;
2894 netdev_get_tx_queue(dev, count + offset)->sb_dev = sb_dev;
3248 qoffset = sb_dev->tc_to_txq[tc].offset;
3299 int ret = 0, offset;
3318 offset = skb_checksum_start_offset(skb);
3320 if (unlikely(offset >= skb_headlen(skb))) {
3322 WARN_ONCE(true, "offset (%d) >= skb_headlen() (%u)\n",
3323 offset, skb_headlen(skb));
3326 csum = skb_checksum(skb, offset, skb->len - offset, 0);
3328 offset += skb->csum_offset;
3329 if (unlikely(offset + sizeof(__sum16) > skb_headlen(skb))) {
3331 WARN_ONCE(true, "offset+2 (%zu) > skb_headlen() (%u)\n",
3332 offset + sizeof(__sum16), skb_headlen(skb));
3335 ret = skb_ensure_writable(skb, offset + sizeof(__sum16));
3339 *(__sum16 *)(skb->data + offset) = csum_fold(csum) ?: CSUM_MANGLED_0;
3350 int ret = 0, offset, start;
3367 offset = start + offsetof(struct sctphdr, checksum);
3368 if (WARN_ON_ONCE(offset >= skb_headlen(skb))) {
3373 ret = skb_ensure_writable(skb, offset + sizeof(__le32));
3380 *(__le32 *)(skb->data + offset) = crc32c_csum;