/kernel/linux/linux-5.10/fs/erofs/ |
H A D | xattr.c | 19 static inline void xattr_iter_end(struct xattr_iter *it, bool atomic) in xattr_iter_end() argument 23 kunmap(it->page); in xattr_iter_end() 25 kunmap_atomic(it->kaddr); in xattr_iter_end() 27 unlock_page(it->page); in xattr_iter_end() 28 put_page(it->page); in xattr_iter_end() 31 static inline void xattr_iter_end_final(struct xattr_iter *it) in xattr_iter_end_final() argument 33 if (!it->page) in xattr_iter_end_final() 36 xattr_iter_end(it, true); in xattr_iter_end_final() 42 struct xattr_iter it; in init_inode_xattrs() local 70 * 1) it i in init_inode_xattrs() 171 xattr_iter_fixup(struct xattr_iter *it) xattr_iter_fixup() argument 193 inline_xattr_iter_begin(struct xattr_iter *it, struct inode *inode) inline_xattr_iter_begin() argument 223 xattr_foreach(struct xattr_iter *it, const struct xattr_iter_handlers *op, unsigned int *tlimit) xattr_foreach() argument 325 struct xattr_iter it; global() member 335 struct getxattr_iter *it = container_of(_it, struct getxattr_iter, it); xattr_entrymatch() local 344 struct getxattr_iter *it = container_of(_it, struct getxattr_iter, it); xattr_namematch() local 352 struct getxattr_iter *it = container_of(_it, struct getxattr_iter, it); xattr_checkbuffer() local 363 struct getxattr_iter *it = container_of(_it, struct getxattr_iter, it); xattr_copyvalue() local 375 inline_getxattr(struct inode *inode, struct getxattr_iter *it) inline_getxattr() argument 395 shared_getxattr(struct inode *inode, struct getxattr_iter *it) shared_getxattr() argument 446 struct getxattr_iter it; erofs_getxattr() local 530 struct xattr_iter it; global() member 540 struct listxattr_iter *it = xattr_entrylist() local 571 struct listxattr_iter *it = xattr_namelist() local 582 struct listxattr_iter *it = xattr_skipvalue() local 596 inline_listxattr(struct listxattr_iter *it) inline_listxattr() argument 615 shared_listxattr(struct listxattr_iter *it) shared_listxattr() argument 655 struct listxattr_iter it; erofs_listxattr() local [all...] |
/kernel/linux/linux-6.6/fs/erofs/ |
H A D | xattr.c | 31 struct erofs_xattr_iter it; in erofs_init_inode_xattrs() local 57 * 1) it is not enough to contain erofs_xattr_ibody_header then in erofs_init_inode_xattrs() 58 * ->xattr_isize should be 0 (it means no xattr); in erofs_init_inode_xattrs() 59 * 2) it is just to contain erofs_xattr_ibody_header, which is on-disk in erofs_init_inode_xattrs() 79 it.buf = __EROFS_BUF_INITIALIZER; in erofs_init_inode_xattrs() 80 erofs_init_metabuf(&it.buf, sb); in erofs_init_inode_xattrs() 81 it.pos = erofs_iloc(inode) + vi->inode_isize; in erofs_init_inode_xattrs() 84 it.kaddr = erofs_bread(&it.buf, erofs_blknr(sb, it in erofs_init_inode_xattrs() 180 erofs_xattr_copy_to_buffer(struct erofs_xattr_iter *it, unsigned int len) erofs_xattr_copy_to_buffer() argument 203 erofs_listxattr_foreach(struct erofs_xattr_iter *it) erofs_listxattr_foreach() argument 255 erofs_getxattr_foreach(struct erofs_xattr_iter *it) erofs_getxattr_foreach() argument 320 erofs_xattr_iter_inline(struct erofs_xattr_iter *it, struct inode *inode, bool getxattr) erofs_xattr_iter_inline() argument 366 erofs_xattr_iter_shared(struct erofs_xattr_iter *it, struct inode *inode, bool getxattr) erofs_xattr_iter_shared() argument 398 struct erofs_xattr_iter it; erofs_getxattr() local 440 struct erofs_xattr_iter it; erofs_listxattr() local [all...] |
/kernel/linux/linux-6.6/tools/testing/selftests/bpf/progs/ |
H A D | iters.c | 34 struct bpf_iter_num it; in iter_err_unsafe_c_loop() local 39 bpf_iter_num_new(&it, 0, 1000); in iter_err_unsafe_c_loop() 40 while ((v = bpf_iter_num_next(&it))) { in iter_err_unsafe_c_loop() 43 bpf_iter_num_destroy(&it); in iter_err_unsafe_c_loop() 54 struct bpf_iter_num it; in iter_err_unsafe_asm_loop() local 60 "r1 = %[it];" /* iterator state */ in iter_err_unsafe_asm_loop() 66 "r1 = %[it];" in iter_err_unsafe_asm_loop() 72 "r1 = %[it];" in iter_err_unsafe_asm_loop() 80 : [it]"r"(&it), in iter_err_unsafe_asm_loop() 96 struct bpf_iter_num it; iter_while_loop() local 114 __attribute__((cleanup(bpf_iter_num_destroy))) struct bpf_iter_num it; iter_while_loop_auto_cleanup() local 132 struct bpf_iter_num it; iter_for_loop() local 180 struct bpf_iter_num it; iter_pragma_unroll_loop() local 200 struct bpf_iter_num it; iter_manual_unroll_loop() local 223 struct bpf_iter_num it; iter_multiple_sequential_loops() local 266 struct bpf_iter_num it; iter_limit_cond_break_loop() local 291 struct bpf_iter_num it; iter_obfuscate_counter() local 335 struct bpf_iter_num it; iter_search_loop() local 661 fill(struct bpf_iter_num *it, int *arr, __u32 n, int mul) fill() argument 673 sum(struct bpf_iter_num *it, int *arr, __u32 n) sum() argument 692 struct bpf_iter_num it; iter_pass_iter_ptr_to_subprog() local [all...] |
H A D | iters_num.c | 142 struct bpf_iter_num it; in num_invalid_range() local 144 res_invalid_range = bpf_iter_num_new(&it, 1, 0); in num_invalid_range() 145 bpf_iter_num_destroy(&it); in num_invalid_range() 156 struct bpf_iter_num it; in num_max_range() local 158 res_max_range = 10 + bpf_iter_num_new(&it, 0, BPF_MAX_LOOPS); in num_max_range() 159 bpf_iter_num_destroy(&it); in num_max_range() 170 struct bpf_iter_num it; in num_e2big_range() local 172 res_e2big_range = bpf_iter_num_new(&it, -1, BPF_MAX_LOOPS); in num_e2big_range() 173 bpf_iter_num_destroy(&it); in num_e2big_range() 184 struct bpf_iter_num it; in num_succ_elem_cnt() local 204 struct bpf_iter_num it; num_overfetched_elem_cnt() local 226 struct bpf_iter_num it; num_fail_elem_cnt() local [all...] |
/kernel/linux/linux-5.10/fs/cifs/ |
H A D | dfs_cache.h | 40 const struct dfs_cache_tgt_iterator *it); 43 const struct dfs_cache_tgt_iterator *it); 45 const struct dfs_cache_tgt_iterator *it, 52 extern int dfs_cache_get_tgt_share(char *path, const struct dfs_cache_tgt_iterator *it, 57 struct dfs_cache_tgt_iterator *it) in dfs_cache_get_next_tgt() 59 if (!tl || list_empty(&tl->tl_list) || !it || in dfs_cache_get_next_tgt() 60 list_is_last(&it->it_list, &tl->tl_list)) in dfs_cache_get_next_tgt() 62 return list_next_entry(it, it_list); in dfs_cache_get_next_tgt() 77 struct dfs_cache_tgt_iterator *it, *nit; in dfs_cache_free_tgts() local 81 list_for_each_entry_safe(it, ni in dfs_cache_free_tgts() 56 dfs_cache_get_next_tgt(struct dfs_cache_tgt_list *tl, struct dfs_cache_tgt_iterator *it) dfs_cache_get_next_tgt() argument 90 dfs_cache_get_tgt_name(const struct dfs_cache_tgt_iterator *it) dfs_cache_get_tgt_name() argument [all...] |
/kernel/linux/linux-6.6/fs/smb/client/ |
H A D | dfs_cache.h | 45 void dfs_cache_noreq_update_tgthint(const char *path, const struct dfs_cache_tgt_iterator *it); 46 int dfs_cache_get_tgt_referral(const char *path, const struct dfs_cache_tgt_iterator *it, 48 int dfs_cache_get_tgt_share(char *path, const struct dfs_cache_tgt_iterator *it, char **share, 56 struct dfs_cache_tgt_iterator *it) in dfs_cache_get_next_tgt() 59 !it || list_is_last(&it->it_list, &tl->tl_list)) in dfs_cache_get_next_tgt() 61 return list_next_entry(it, it_list); in dfs_cache_get_next_tgt() 76 struct dfs_cache_tgt_iterator *it, *nit; in dfs_cache_free_tgts() local 80 list_for_each_entry_safe(it, nit, &tl->tl_list, it_list) { in dfs_cache_free_tgts() 81 list_del(&it in dfs_cache_free_tgts() 55 dfs_cache_get_next_tgt(struct dfs_cache_tgt_list *tl, struct dfs_cache_tgt_iterator *it) dfs_cache_get_next_tgt() argument 89 dfs_cache_get_tgt_name(const struct dfs_cache_tgt_iterator *it) dfs_cache_get_tgt_name() argument [all...] |
/kernel/linux/linux-6.6/rust/macros/ |
H A D | helpers.rs | 5 pub(crate) fn try_ident(it: &mut token_stream::IntoIter) -> Option<String> { 6 if let Some(TokenTree::Ident(ident)) = it.next() { 13 pub(crate) fn try_literal(it: &mut token_stream::IntoIter) -> Option<String> { 14 if let Some(TokenTree::Literal(literal)) = it.next() { 21 pub(crate) fn try_string(it: &mut token_stream::IntoIter) -> Option<String> { 22 try_literal(it).and_then(|string| { 37 pub(crate) fn expect_ident(it: &mut token_stream::IntoIter) -> String { 38 try_ident(it).expect("Expected Ident") 41 pub(crate) fn expect_punct(it: &mut token_stream::IntoIter) -> char { 42 if let TokenTree::Punct(punct) = it [all...] |
H A D | module.rs | 7 fn expect_string_array(it: &mut token_stream::IntoIter) -> Vec<String> { in expect_string_array() 8 let group = expect_group(it); in expect_string_array() 11 let mut it = group.stream().into_iter(); in expect_string_array() variables 13 while let Some(val) = try_string(&mut it) { in expect_string_array() 16 match it.next() { in expect_string_array() 103 fn parse(it: &mut token_stream::IntoIter) -> Self { in parse() 112 let key = match it.next() { in parse() 125 assert_eq!(expect_punct(it), ':'); in parse() 128 "type" => info.type_ = expect_ident(it), in parse() 129 "name" => info.name = expect_string_ascii(it), in parse() 172 let mut it = ts.into_iter(); global() variables [all...] |
H A D | concat_idents.rs | 7 fn expect_ident(it: &mut token_stream::IntoIter) -> Ident { in expect_ident() 8 if let Some(TokenTree::Ident(ident)) = it.next() { in expect_ident() 16 let mut it = ts.into_iter(); variables 17 let a = expect_ident(&mut it); 18 assert_eq!(expect_punct(&mut it), ','); 19 let b = expect_ident(&mut it); 20 assert!(it.next().is_none(), "only two idents can be concatenated");
|
/kernel/linux/linux-5.10/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 98 nvkm_vmm_trace(struct nvkm_vmm_iter *it, char *buf) in nvkm_vmm_trace() argument 101 for (lvl = it->max; lvl >= 0; lvl--) { in nvkm_vmm_trace() 102 if (lvl >= it->lvl) in nvkm_vmm_trace() 103 buf += sprintf(buf, "%05x:", it->pte[lvl]); in nvkm_vmm_trace() 120 nvkm_vmm_flush_mark(struct nvkm_vmm_iter *it) in nvkm_vmm_flush_mark() argument 122 it->flush = min(it->flush, it->max - it->lvl); in nvkm_vmm_flush_mark() 126 nvkm_vmm_flush(struct nvkm_vmm_iter *it) in nvkm_vmm_flush() argument 138 nvkm_vmm_unref_pdes(struct nvkm_vmm_iter *it) nvkm_vmm_unref_pdes() argument 197 nvkm_vmm_unref_sptes(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgt, const struct nvkm_vmm_desc *desc, u32 ptei, u32 ptes) nvkm_vmm_unref_sptes() argument 258 nvkm_vmm_unref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_unref_ptes() argument 296 nvkm_vmm_ref_sptes(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgt, const struct nvkm_vmm_desc *desc, u32 ptei, u32 ptes) nvkm_vmm_ref_sptes() argument 364 nvkm_vmm_ref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_ref_ptes() argument 394 nvkm_vmm_sparse_unref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_sparse_unref_ptes() argument 406 nvkm_vmm_sparse_ref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_sparse_ref_ptes() argument 413 nvkm_vmm_ref_hwpt(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgd, u32 pdei) nvkm_vmm_ref_hwpt() argument 484 nvkm_vmm_ref_swpt(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgd, u32 pdei) nvkm_vmm_ref_swpt() argument 508 struct nvkm_vmm_iter it; nvkm_vmm_iter() local 1805 nvkm_vmm_boot_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_boot_ptes() argument [all...] |
/kernel/linux/linux-5.10/include/linux/ceph/ |
H A D | messenger.h | 92 #define __ceph_bio_iter_advance_step(it, n, STEP) do { \ 96 BUG_ON(!(it)->iter.bi_size); \ 97 __cur_n = min((it)->iter.bi_size, __n); \ 99 bio_advance_iter((it)->bio, &(it)->iter, __cur_n); \ 100 if (!(it)->iter.bi_size && (it)->bio->bi_next) { \ 102 (it)->bio = (it)->bio->bi_next; \ 103 (it) [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/nouveau/nvkm/subdev/mmu/ |
H A D | vmm.c | 98 nvkm_vmm_trace(struct nvkm_vmm_iter *it, char *buf) in nvkm_vmm_trace() argument 101 for (lvl = it->max; lvl >= 0; lvl--) { in nvkm_vmm_trace() 102 if (lvl >= it->lvl) in nvkm_vmm_trace() 103 buf += sprintf(buf, "%05x:", it->pte[lvl]); in nvkm_vmm_trace() 120 nvkm_vmm_flush_mark(struct nvkm_vmm_iter *it) in nvkm_vmm_flush_mark() argument 122 it->flush = min(it->flush, it->max - it->lvl); in nvkm_vmm_flush_mark() 126 nvkm_vmm_flush(struct nvkm_vmm_iter *it) in nvkm_vmm_flush() argument 138 nvkm_vmm_unref_pdes(struct nvkm_vmm_iter *it) nvkm_vmm_unref_pdes() argument 197 nvkm_vmm_unref_sptes(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgt, const struct nvkm_vmm_desc *desc, u32 ptei, u32 ptes) nvkm_vmm_unref_sptes() argument 258 nvkm_vmm_unref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_unref_ptes() argument 296 nvkm_vmm_ref_sptes(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgt, const struct nvkm_vmm_desc *desc, u32 ptei, u32 ptes) nvkm_vmm_ref_sptes() argument 364 nvkm_vmm_ref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_ref_ptes() argument 394 nvkm_vmm_sparse_unref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_sparse_unref_ptes() argument 406 nvkm_vmm_sparse_ref_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_sparse_ref_ptes() argument 413 nvkm_vmm_ref_hwpt(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgd, u32 pdei) nvkm_vmm_ref_hwpt() argument 484 nvkm_vmm_ref_swpt(struct nvkm_vmm_iter *it, struct nvkm_vmm_pt *pgd, u32 pdei) nvkm_vmm_ref_swpt() argument 508 struct nvkm_vmm_iter it; nvkm_vmm_iter() local 1908 nvkm_vmm_boot_ptes(struct nvkm_vmm_iter *it, bool pfn, u32 ptei, u32 ptes) nvkm_vmm_boot_ptes() argument [all...] |
/kernel/linux/linux-6.6/include/linux/ceph/ |
H A D | messenger.h | 89 * the provided buffer. Otherwise, it should be read into the cursor. 136 #define __ceph_bio_iter_advance_step(it, n, STEP) do { \ 140 BUG_ON(!(it)->iter.bi_size); \ 141 __cur_n = min((it)->iter.bi_size, __n); \ 143 bio_advance_iter((it)->bio, &(it)->iter, __cur_n); \ 144 if (!(it)->iter.bi_size && (it)->bio->bi_next) { \ 146 (it)->bio = (it) [all...] |
/kernel/linux/linux-6.6/drivers/of/ |
H A D | base.c | 246 * of_node_put() will be called on it 249 * of_node_put() on it when done. 517 * of_node_put() on it when done. 538 * This is like of_get_parent() except that it drops the 539 * refcount on the passed node, making it suitable for iterating 543 * of_node_put() on it when done. 584 * it when done. Returns NULL when prev is the last child. Decrements the 605 * This function is like of_get_next_child(), except that it 639 * on it when done. Returns NULL when prev is the last child. Decrements 681 * Return: a node pointer with refcount incremented, use of_node_put() on it 1106 of_phandle_iterator_init(struct of_phandle_iterator *it, const struct device_node *np, const char *list_name, const char *cells_name, int cell_count) of_phandle_iterator_init() argument 1139 of_phandle_iterator_next(struct of_phandle_iterator *it) of_phandle_iterator_next() argument 1224 of_phandle_iterator_args(struct of_phandle_iterator *it, uint32_t *args, int size) of_phandle_iterator_args() argument 1247 struct of_phandle_iterator it; __of_parse_phandle_with_args() local 1500 struct of_phandle_iterator it; of_count_phandle_with_args() local [all...] |
/kernel/linux/linux-5.10/drivers/of/ |
H A D | base.c | 245 * of_node_put() will be called on it 248 * of_node_put() on it when done. 311 * NULL, local thread number within the core is returned in it. 343 * core is returned in it. 349 * for thread ids on PowerPC. If it doesn't exist fallback to in arch_find_n_match_cpu_physical_id() 378 * of_node_put() on it when done. Returns NULL if not found. 431 * for it, so call of_node_put() on it when done. Returns NULL if not found. 658 * of_node_put() on it when done. 679 * This is like of_get_parent() except that it drop 1240 of_phandle_iterator_init(struct of_phandle_iterator *it, const struct device_node *np, const char *list_name, const char *cells_name, int cell_count) of_phandle_iterator_init() argument 1273 of_phandle_iterator_next(struct of_phandle_iterator *it) of_phandle_iterator_next() argument 1358 of_phandle_iterator_args(struct of_phandle_iterator *it, uint32_t *args, int size) of_phandle_iterator_args() argument 1381 struct of_phandle_iterator it; __of_parse_phandle_with_args() local 1747 struct of_phandle_iterator it; of_count_phandle_with_args() local [all...] |
/kernel/linux/linux-5.10/drivers/iio/imu/inv_icm42600/ |
H A D | inv_icm42600_timestamp.c | 115 struct inv_icm42600_timestamp_interval *it; in inv_icm42600_timestamp_interrupt() local 126 it = &ts->it; in inv_icm42600_timestamp_interrupt() 127 it->lo = it->up; in inv_icm42600_timestamp_interrupt() 128 it->up = timestamp; in inv_icm42600_timestamp_interrupt() 129 delta = it->up - it->lo; in inv_icm42600_timestamp_interrupt() 130 if (it->lo != 0) { in inv_icm42600_timestamp_interrupt() 143 ts->timestamp = it in inv_icm42600_timestamp_interrupt() [all...] |
/kernel/linux/linux-5.10/drivers/gpu/drm/i915/ |
H A D | i915_active.c | 133 struct active_node *it, *n; in __active_retire() local 171 /* ... except if you wait on it, you must manage your own references! */ in __active_retire() 175 rbtree_postorder_for_each_entry_safe(it, n, &root, node) { in __active_retire() 176 GEM_BUG_ON(i915_active_fence_isset(&it->base)); in __active_retire() 177 kmem_cache_free(global.slab_cache, it); in __active_retire() 239 struct active_node *it; in __active_lookup() local 246 * at all. We can reuse the last slot if it is empty, that is in __active_lookup() 247 * after the previous activity has been retired, or if it matches the in __active_lookup() 250 it = READ_ONCE(ref->cache); in __active_lookup() 251 if (it) { in __active_lookup() 493 struct active_node *it; __active_fence() local 600 flush_barrier(struct active_node *it) flush_barrier() argument 617 struct active_node *it, *n; flush_lazy_signals() local 740 struct active_node *it, *n; await_active() local 993 struct active_node *it; i915_active_acquire_barrier() local [all...] |
/kernel/linux/linux-5.10/arch/s390/kernel/ |
H A D | relocate_kernel.S | 36 tml %r5,0x1 # is it a destination page? 39 nill %r6,0xf000 # mask it out and... 42 tml %r5,0x2 # is it a indirection page? 45 lgr %r2,%r5 # move it into the right register, 48 tml %r5,0x4 # is it the done indicator? 49 je .source_test # NO! Well, then it should be the source indicator... 50 j .done # ok, lets finish it here... 52 tml %r5,0x8 # it should be a source indicator... 53 je .base # NO, ignore it...
|
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/ |
H A D | i915_active.c | 129 struct active_node *it, *n; in __active_retire() local 166 /* ... except if you wait on it, you must manage your own references! */ in __active_retire() 170 rbtree_postorder_for_each_entry_safe(it, n, &root, node) { in __active_retire() 171 GEM_BUG_ON(i915_active_fence_isset(&it->base)); in __active_retire() 172 kmem_cache_free(slab_cache, it); in __active_retire() 234 struct active_node *it; in __active_lookup() local 241 * at all. We can reuse the last slot if it is empty, that is in __active_lookup() 242 * after the previous activity has been retired, or if it matches the in __active_lookup() 245 it = READ_ONCE(ref->cache); in __active_lookup() 246 if (it) { in __active_lookup() 569 flush_barrier(struct active_node *it) flush_barrier() argument 586 struct active_node *it, *n; flush_lazy_signals() local 709 struct active_node *it, *n; await_active() local 962 struct active_node *it; i915_active_acquire_barrier() local [all...] |
/kernel/linux/linux-6.6/drivers/iio/common/inv_sensors/ |
H A D | inv_sensors_timestamp.c | 108 delta = ts->it.lo - ts->timestamp; in inv_align_timestamp_it() 126 struct inv_sensors_timestamp_interval *it; in inv_sensors_timestamp_interrupt() local 136 it = &ts->it; in inv_sensors_timestamp_interrupt() 137 it->lo = it->up; in inv_sensors_timestamp_interrupt() 138 it->up = timestamp; in inv_sensors_timestamp_interrupt() 139 delta = it->up - it->lo; in inv_sensors_timestamp_interrupt() 140 if (it in inv_sensors_timestamp_interrupt() [all...] |
/kernel/linux/linux-6.6/drivers/gpu/drm/i915/gt/ |
H A D | intel_migrate.c | 95 * it is kept pinned so that it can be used at any time. As we need in migrate_vm() 134 * For the PTE window it's also quite different, since each PTE must in migrate_vm() 135 * point to some 64K page, one for each PT(since it's in lmem), and yet in migrate_vm() 363 struct sgt_dma *it, in emit_pte() 449 GEM_BUG_ON(!IS_ALIGNED(it->dma, page_size)); in emit_pte() 451 *cs++ = lower_32_bits(encode | it->dma); in emit_pte() 452 *cs++ = upper_32_bits(encode | it->dma); in emit_pte() 457 it->dma += page_size; in emit_pte() 458 if (it in emit_pte() 362 emit_pte(struct i915_request *rq, struct sgt_dma *it, unsigned int pat_index, bool is_lmem, u64 offset, int length) emit_pte() argument 657 get_ccs_sg_sgt(struct sgt_dma *it, u64 bytes_to_cpy) get_ccs_sg_sgt() argument 993 struct sgt_dma it = sg_sgt(sg); intel_context_migrate_clear() local [all...] |
/kernel/linux/linux-6.6/kernel/time/ |
H A D | posix-cpu-timers.c | 115 return pid_task(timer->it.cpu.pid, clock_pid_type(timer->it_clock)); in cpu_timer_task_rcu() 124 u64 delta, incr, expires = timer->it.cpu.node.expires; in bump_cpu_timer() 144 timer->it.cpu.node.expires += incr; in bump_cpu_timer() 148 return timer->it.cpu.node.expires; in bump_cpu_timer() 171 * exported, but it is much more than 1s/HZ. in posix_cpu_clock_getres() 289 * CPU timers armed. Before starting a timer it's required to check whether 309 * to synchronize the timer to the clock every time we start it. in thread_group_start_cputime() 316 * only gets written to in one operation. We set it after in thread_group_start_cputime() 410 timerqueue_init(&new_timer->it.cpu.node); in posix_cpu_timer_create() 411 new_timer->it in posix_cpu_timer_create() 947 check_cpu_itimer(struct task_struct *tsk, struct cpu_itimer *it, u64 *expires, u64 cur_time, int signo) check_cpu_itimer() argument 1504 struct itimerspec64 it; do_cpu_nanosleep() local [all...] |
/kernel/linux/linux-5.10/kernel/time/ |
H A D | posix-cpu-timers.c | 108 return pid_task(timer->it.cpu.pid, clock_pid_type(timer->it_clock)); in cpu_timer_task_rcu() 117 u64 delta, incr, expires = timer->it.cpu.node.expires; in bump_cpu_timer() 137 timer->it.cpu.node.expires += incr; in bump_cpu_timer() 141 return timer->it.cpu.node.expires; in bump_cpu_timer() 164 * exported, but it is much more than 1s/HZ. in posix_cpu_clock_getres() 283 * CPU timers armed. Before starting a timer it's required to check whether 301 * to synchronize the timer to the clock every time we start it. in thread_group_start_cputime() 308 * only gets written to in one operation. We set it after in thread_group_start_cputime() 402 timerqueue_init(&new_timer->it.cpu.node); in posix_cpu_timer_create() 403 new_timer->it in posix_cpu_timer_create() 883 check_cpu_itimer(struct task_struct *tsk, struct cpu_itimer *it, u64 *expires, u64 cur_time, int signo) check_cpu_itimer() argument 1441 struct itimerspec64 it; do_cpu_nanosleep() local [all...] |
/kernel/linux/linux-5.10/net/ipv4/ |
H A D | ipmr_base.c | 148 struct mr_mfc_iter *it, loff_t pos) in mr_mfc_seq_idx() 150 struct mr_table *mrt = it->mrt; in mr_mfc_seq_idx() 154 it->cache = &mrt->mfc_cache_list; in mr_mfc_seq_idx() 160 spin_lock_bh(it->lock); in mr_mfc_seq_idx() 161 it->cache = &mrt->mfc_unres_queue; in mr_mfc_seq_idx() 162 list_for_each_entry(mfc, it->cache, list) in mr_mfc_seq_idx() 165 spin_unlock_bh(it->lock); in mr_mfc_seq_idx() 167 it->cache = NULL; in mr_mfc_seq_idx() 175 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_next() local 177 struct mr_table *mrt = it in mr_mfc_seq_next() 147 mr_mfc_seq_idx(struct net *net, struct mr_mfc_iter *it, loff_t pos) mr_mfc_seq_idx() argument [all...] |
/kernel/linux/linux-6.6/net/ipv4/ |
H A D | ipmr_base.c | 148 struct mr_mfc_iter *it, loff_t pos) in mr_mfc_seq_idx() 150 struct mr_table *mrt = it->mrt; in mr_mfc_seq_idx() 154 it->cache = &mrt->mfc_cache_list; in mr_mfc_seq_idx() 160 spin_lock_bh(it->lock); in mr_mfc_seq_idx() 161 it->cache = &mrt->mfc_unres_queue; in mr_mfc_seq_idx() 162 list_for_each_entry(mfc, it->cache, list) in mr_mfc_seq_idx() 165 spin_unlock_bh(it->lock); in mr_mfc_seq_idx() 167 it->cache = NULL; in mr_mfc_seq_idx() 175 struct mr_mfc_iter *it = seq->private; in mr_mfc_seq_next() local 177 struct mr_table *mrt = it in mr_mfc_seq_next() 147 mr_mfc_seq_idx(struct net *net, struct mr_mfc_iter *it, loff_t pos) mr_mfc_seq_idx() argument [all...] |