/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/ARM/ |
H A D | ARMISelLowering.cpp | 138 "arm-promote-constant-max-size", cl::Hidden, 139 cl::desc("Maximum size of constant to promote into a constant pool"), 143 cl::desc("Maximum size of ALL constants to promote into a constant pool"), 1743 // source/dest is aligned and the copy size is large enough. We therefore want 1972 for (unsigned i = 0; i != RVLocs.size(); ++i) { in LowerCallResult() 2110 // If we're optimizing for minimum size and the function is called three or in LowerCall() 2164 for (unsigned i = 0, realArgIdx = 0, e = ArgLocs.size(); in LowerCall() 2255 // If parameter size outsides register area, "offset" value in LowerCall() 2293 for (unsigned i = 0, e = RegsToPass.size(); i != e; ++i) { in LowerCall() 2419 // Emit regular call when code size i in LowerCall() [all...] |
/third_party/vixl/src/aarch32/ |
H A D | macro-assembler-aarch32.h | 55 size_t size, 262 virtual void EnsureEmitPoolsFor(size_t size) VIXL_OVERRIDE; 308 explicit MacroAssembler(size_t size, InstructionSet isa = kDefaultISA) = delete; in MacroAssembler() 310 explicit MacroAssembler(size_t size, InstructionSet isa = kDefaultISA) in MacroAssembler() argument 311 : Assembler(size, isa), in MacroAssembler() 327 MacroAssembler(byte* buffer, size_t size, InstructionSet isa = kDefaultISA) in MacroAssembler() argument 328 : Assembler(buffer, size, isa), in MacroAssembler() 337 MacroAssembler(PandaAllocator* allocator, byte* buffer, size_t size, InstructionSet isa = kDefaultISA) in MacroAssembler() 338 : Assembler(buffer, size, isa), in MacroAssembler() 480 int size in MacroAssembler() local 539 EnsureEmitFor(uint32_t size) MacroAssembler() argument 964 Claim(int32_t size) MacroAssembler() argument 971 Drop(int32_t size) MacroAssembler() argument 1467 EncodingSize size = Best; MacroAssembler() local [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/CodeGen/SelectionDAG/ |
H A D | DAGCombiner.cpp | 230 // We use the minimum store size here, since that's all we can guarantee in DAGCombiner() 257 if (WorklistMap.insert(std::make_pair(N, Worklist.size())).second) in AddToWorklist() 335 // Scalars have size 0 to distinguish from singleton vectors. 760 return ((DAGCombiner*)DC)->CombineTo(N, &To[0], To.size(), AddTo); in CombineTo() 802 // APInts must be the same size for most operations, this helper 1743 for (unsigned i = 0; i < TFs.size(); ++i) { in visitTokenFactor() 1747 if (Ops.size() > TokenFactorInlineLimit) { in visitTokenFactor() 1748 for (unsigned j = i; j < TFs.size(); j++) in visitTokenFactor() 1788 for (unsigned i = 1, e = TFs.size(); i < e; i++) in visitTokenFactor() 1816 while (OrigOpNumber < Ops.size() in visitTokenFactor() [all...] |
H A D | LegalizeDAG.cpp | 281 assert(NewMask.size() == NumDestElts && "Non-integer NumEltsGrowth?"); in ShuffleWithNarrowerEltType() 529 "Can only promote stores to same size type"); in LegalizeStoreOps() 568 "Store size not an integral number of bytes!"); in LegalizeStoreOps() 700 "Can only promote loads to same size type"); in LegalizeLoadOps() 780 "Load size not an integral number of bytes!"); in LegalizeLoadOps() 1450 // Convert to an integer of the same size. 1837 while (IntermedVals.size() > 2) { 1839 for (unsigned i = 0, e = (IntermedVals.size() & ~1u); i < e; i += 2) { 1845 FinalIndices.reserve(IntermedVals[i].second.size() + 1846 IntermedVals[i+1].second.size()); [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_main.c | 129 * @size: size of memory requested 133 u64 size, u32 alignment) in i40e_allocate_dma_mem_d() 137 mem->size = ALIGN(size, alignment); in i40e_allocate_dma_mem_d() 138 mem->va = dma_alloc_coherent(&pf->pdev->dev, mem->size, &mem->pa, in i40e_allocate_dma_mem_d() 155 dma_free_coherent(&pf->pdev->dev, mem->size, mem->va, mem->pa); in i40e_free_dma_mem_d() 158 mem->size = 0; in i40e_free_dma_mem_d() 167 * @size: size o 132 i40e_allocate_dma_mem_d(struct i40e_hw *hw, struct i40e_dma_mem *mem, u64 size, u32 alignment) i40e_allocate_dma_mem_d() argument 169 i40e_allocate_virt_mem_d(struct i40e_hw *hw, struct i40e_virt_mem *mem, u32 size) i40e_allocate_virt_mem_d() argument 10785 int size; i40e_vsi_alloc_arrays() local 11408 ssize_t size; i40e_init_interrupt_scheme() local 12087 int size; i40e_sw_init() local [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/intel/i40e/ |
H A D | i40e_main.c | 132 * @size: size of memory requested 136 u64 size, u32 alignment) in i40e_allocate_dma_mem_d() 140 mem->size = ALIGN(size, alignment); in i40e_allocate_dma_mem_d() 141 mem->va = dma_alloc_coherent(&pf->pdev->dev, mem->size, &mem->pa, in i40e_allocate_dma_mem_d() 158 dma_free_coherent(&pf->pdev->dev, mem->size, mem->va, mem->pa); in i40e_free_dma_mem_d() 161 mem->size = 0; in i40e_free_dma_mem_d() 170 * @size: size o 135 i40e_allocate_dma_mem_d(struct i40e_hw *hw, struct i40e_dma_mem *mem, u64 size, u32 alignment) i40e_allocate_dma_mem_d() argument 172 i40e_allocate_virt_mem_d(struct i40e_hw *hw, struct i40e_virt_mem *mem, u32 size) i40e_allocate_virt_mem_d() argument 11422 int size; i40e_vsi_alloc_arrays() local 12044 ssize_t size; i40e_init_interrupt_scheme() local 12725 int size; i40e_sw_init() local [all...] |
/third_party/vulkan-headers/include/vulkan/ |
H A D | vulkan.hpp | 299 : m_count( static_cast<uint32_t>( list.size() ) ) 306 : m_count( static_cast<uint32_t>( list.size() ) ) 315 // Any type with a .data() return type implicitly convertible to T*, and a .size() return type implicitly 319 std::is_convertible<decltype( std::declval<V>().size() ), std::size_t>::value>::type * = nullptr> 321 : m_count( static_cast<uint32_t>( v.size() ) ) 353 uint32_t size() const VULKAN_HPP_NOEXCEPT 437 : m_count( static_cast<uint32_t>( list.size() ) ) 446 : m_count( static_cast<uint32_t>( list.size() ) ) 455 : m_count( static_cast<uint32_t>( list.size() ) ) 464 : m_count( static_cast<uint32_t>( list.size() ) ) [all...] |
/foundation/arkui/ace_engine/frameworks/core/components_ng/render/adapter/ |
H A D | rosen_render_context.cpp | 247 float RosenRenderContext::ConvertDimensionToScaleBySize(const Dimension& dimension, float size) in ConvertDimensionToScaleBySize() argument 252 return size > 0.0f ? static_cast<float>(dimension.ConvertToPx() / size) : 0.5f; in ConvertDimensionToScaleBySize() 315 // has set size before, trigger transition directly. in OnNodeAppear() 1005 if (bgBlurStyle->blurOption.grayscale.size() > 1) { in UpdateBackBlurStyle() 1051 if (effectOption->blurOption.grayscale.size() > 1) { in UpdateBackgroundEffect() 1067 if (fgBlurStyle->blurOption.grayscale.size() > 1) { in UpdateFrontBlurStyle() 1425 float RosenRenderContext::ConvertDimensionToPx(Dimension& src, float size) in ConvertDimensionToPx() argument 1428 return src.ConvertToPxWithSize(size); in ConvertDimensionToPx() 1782 // size i in MarshallTranslate() 6109 auto size = rect.GetSize(); global() local [all...] |
/foundation/multimedia/media_library/frameworks/js/src/ |
H A D | file_asset_napi.cpp | 172 DECLARE_NAPI_GETTER("size", JSGetSize), in Init() 768 int64_t size; in JSGetSize() local 780 size = obj->fileAssetPtr->GetSize(); in JSGetSize() 781 napi_create_int64(env, size, &jsResult); in JSGetSize() 1733 context->pixelmap = ThumbnailManager::QueryThumbnail(context->objectPtr->GetUri(), context->size, path); in JSGetThumbnailExecute() 1832 asyncContext->size.width = DEFAULT_THUMB_SIZE; in GetJSArgsForGetThumbnail() 1833 asyncContext->size.height = DEFAULT_THUMB_SIZE; in GetJSArgsForGetThumbnail() 1848 GetInt32InfoFromNapiObject(env, argv[PARAM0], "width", asyncContext->size.width); in GetJSArgsForGetThumbnail() 1849 GetInt32InfoFromNapiObject(env, argv[PARAM0], "height", asyncContext->size.height); in GetJSArgsForGetThumbnail() 1892 if (GetNapiObjectFromNapiObject(env, object, "size", in GetPhotoRequestOption() [all...] |
/kernel/linux/linux-5.10/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 471 /* maximum mtu size */ 2219 u32 size = skb_headlen(skb); in nv_start_xmit() local 2220 u32 entries = (size >> NV_TX2_TSO_MAX_SHIFT) + ((size & (NV_TX2_TSO_MAX_SIZE-1)) ? 1 : 0); in nv_start_xmit() 2257 bcnt = (size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : size; in nv_start_xmit() 2280 size -= bcnt; in nv_start_xmit() 2285 } while (size); in nv_start_xmit() 2393 u32 size = skb_headlen(skb); in nv_start_xmit_optimized() local 2394 u32 entries = (size >> NV_TX2_TSO_MAX_SHIF in nv_start_xmit_optimized() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/nvidia/ |
H A D | forcedeth.c | 471 /* maximum mtu size */ 2214 u32 size = skb_headlen(skb); in nv_start_xmit() local 2215 u32 entries = (size >> NV_TX2_TSO_MAX_SHIFT) + ((size & (NV_TX2_TSO_MAX_SIZE-1)) ? 1 : 0); in nv_start_xmit() 2252 bcnt = (size > NV_TX2_TSO_MAX_SIZE) ? NV_TX2_TSO_MAX_SIZE : size; in nv_start_xmit() 2275 size -= bcnt; in nv_start_xmit() 2280 } while (size); in nv_start_xmit() 2388 u32 size = skb_headlen(skb); in nv_start_xmit_optimized() local 2389 u32 entries = (size >> NV_TX2_TSO_MAX_SHIF in nv_start_xmit_optimized() [all...] |
/kernel/linux/linux-6.6/drivers/net/ethernet/stmicro/stmmac/ |
H A D | stmmac_main.c | 107 MODULE_PARM_DESC(buf_sz, "DMA buffer size"); 1628 /* struct stmmac_xdp_buff is using cb field (maximum size of 24 bytes) in stmmac_alloc_rx_buffers_zc() 1630 * use this macro to make sure no size violations. in stmmac_alloc_rx_buffers_zc() 1975 size_t size; in __free_dma_tx_desc_resources() local 1982 size = sizeof(struct dma_extended_desc); in __free_dma_tx_desc_resources() 1985 size = sizeof(struct dma_edesc); in __free_dma_tx_desc_resources() 1988 size = sizeof(struct dma_desc); in __free_dma_tx_desc_resources() 1992 size *= dma_conf->dma_tx_size; in __free_dma_tx_desc_resources() 1994 dma_free_coherent(priv->device, size, addr, tx_q->dma_tx_phy); in __free_dma_tx_desc_resources() 2132 size_t size; in __alloc_dma_tx_desc_resources() local 6157 sysfs_display_ring(void *head, int size, int extend_desc, struct seq_file *seq, dma_addr_t dma_phy_addr) sysfs_display_ring() argument [all...] |
/kernel/linux/linux-6.6/drivers/scsi/ |
H A D | ipr.c | 888 * In the case of sis64, calculate the ioarcb size required. Then or in the 900 /* The default size is 256 bytes */ in ipr_send_command() 903 /* If the number of ioadls * size of ioadl > 128 bytes, in ipr_send_command() 1431 * @i: size of buffer 3362 * @count: buffer size 3390 .size = 0, 3458 * @count: buffer size 3491 * @count: buffer size 3579 * @count: buffer size 3631 * @count: buffer size [all...] |
/third_party/icu/ohos_icu4j/src/main/tests/ohos/global/icu/dev/test/format/ |
H A D | NumberFormatTest.java | 3201 if (errors.size() != 0) { in TestJB5358() 3203 for (int i = 0; i < errors.size(); i++) { in TestJB5358() 4685 assertEquals("Negative US Results: " + strUS, 5, resultUS.size()); in TestFormatToCharacterIteratorIssue11805() 4693 assertEquals("Negative DE Results: " + strDE, 5, resultDE.size()); in TestFormatToCharacterIteratorIssue11805() 4698 assertEquals("Negative IN Results: " + strIN, 5, resultIN.size()); in TestFormatToCharacterIteratorIssue11805() 4703 assertEquals("Negative JA Results: " + strJP, 3, resultJP.size()); in TestFormatToCharacterIteratorIssue11805() 4708 assertEquals("Negative GB Results: " + strGB , 5, resultGB.size()); in TestFormatToCharacterIteratorIssue11805() 4714 assertEquals("Negative GB Results: " + strGB , 5, resultGB.size()); in TestFormatToCharacterIteratorIssue11805() 4718 assertEquals("Negative GB Results: " + strGB , 4, resultGB.size()); in TestFormatToCharacterIteratorIssue11805() 4726 assertEquals("Unit AU Result: " + strAU , 4, resultAU.size()); in TestFormatToCharacterIteratorIssue11805() [all...] |
/third_party/icu/icu4j/main/tests/core/src/com/ibm/icu/dev/test/format/ |
H A D | NumberFormatTest.java | 3198 if (errors.size() != 0) { in TestJB5358() 3200 for (int i = 0; i < errors.size(); i++) { in TestJB5358() 4685 assertEquals("Negative US Results: " + strUS, 5, resultUS.size()); in TestFormatToCharacterIteratorIssue11805() 4693 assertEquals("Negative DE Results: " + strDE, 5, resultDE.size()); in TestFormatToCharacterIteratorIssue11805() 4698 assertEquals("Negative IN Results: " + strIN, 5, resultIN.size()); in TestFormatToCharacterIteratorIssue11805() 4703 assertEquals("Negative JA Results: " + strJP, 3, resultJP.size()); in TestFormatToCharacterIteratorIssue11805() 4708 assertEquals("Negative GB Results: " + strGB , 5, resultGB.size()); in TestFormatToCharacterIteratorIssue11805() 4714 assertEquals("Negative GB Results: " + strGB , 5, resultGB.size()); in TestFormatToCharacterIteratorIssue11805() 4718 assertEquals("Negative GB Results: " + strGB , 4, resultGB.size()); in TestFormatToCharacterIteratorIssue11805() 4726 assertEquals("Unit AU Result: " + strAU , 4, resultAU.size()); in TestFormatToCharacterIteratorIssue11805() [all...] |
/third_party/node/deps/v8/src/codegen/riscv64/ |
H A D | macro-assembler-riscv64.cc | 1771 uint16_t size, bool sign_extend) { in ExtractBits() 1772 DCHECK(pos < 64 && 0 < size && size <= 64 && 0 < pos + size && in ExtractBits() 1773 pos + size <= 64); in ExtractBits() 1774 slli(rt, rs, 64 - (pos + size)); in ExtractBits() 1776 srai(rt, rt, 64 - size); in ExtractBits() 1778 srli(rt, rt, 64 - size); in ExtractBits() 1783 int size) { 1784 DCHECK_LT(size, 6 1770 ExtractBits(Register rt, Register rs, uint16_t pos, uint16_t size, bool sign_extend) ExtractBits() argument 3558 PushArray(Register array, Register size, PushArrayOrder order) PushArray() argument [all...] |
/third_party/node/deps/v8/src/objects/ |
H A D | objects.cc | 1320 Handle<TemplateList> TemplateList::New(Isolate* isolate, int size) { in New() argument 1322 isolate->factory()->NewFixedArray(kLengthIndex + size); in New() 1996 os << "<FreeSpace[" << FreeSpace::cast(*this).size(kRelaxedLoad) << "]>"; in HeapObjectShortPrint() 2222 return FreeSpace::unchecked_cast(*this).size(kRelaxedLoad); in SizeFromMap() 3879 int size = enumeration_index; in CopyUpToAddAttributes() local 3881 DescriptorArray::Allocate(isolate, size, slack); in CopyUpToAddAttributes() 3888 for (InternalIndex i : InternalIndex::Range(size)) { in CopyUpToAddAttributes() 3908 for (InternalIndex i : InternalIndex::Range(size)) { in CopyUpToAddAttributes() 3927 int size = enumeration_index; in CopyForFastObjectClone() local 3929 DescriptorArray::Allocate(isolate, size, slac in CopyForFastObjectClone() 4087 New(Isolate* isolate, int size) New() argument [all...] |
/third_party/node/deps/v8/src/wasm/baseline/arm/ |
H A D | liftoff-assembler-arm.h | 71 "Slot size should be twice the size of the 32 bit pointer."); 502 // The frame size is bigger than 4KB, so we might overflow the available stack in PatchPrepareStackFrame() 611 int offset, int size) { in LoadFromInstance() 614 switch (size) { in LoadFromInstance() 1462 void LiftoffAssembler::FillStackSlotsWithZero(int start, int size) { in FillStackSlotsWithZero() argument 1463 DCHECK_LT(0, size); in FillStackSlotsWithZero() 1464 DCHECK_EQ(0, size % 4); in FillStackSlotsWithZero() 1465 RecordUsedSpillOffset(start + size); in FillStackSlotsWithZero() 1472 if (size < in FillStackSlotsWithZero() 610 LoadFromInstance(Register dst, Register instance, int offset, int size) LoadFromInstance() argument [all...] |
/third_party/node/deps/v8/src/wasm/baseline/x64/ |
H A D | liftoff-assembler-x64.h | 234 // The frame size is bigger than 4KB, so we might overflow the available stack in PatchPrepareStackFrame() 336 int offset, int size) { in LoadFromInstance() 339 switch (size) { in LoadFromInstance() 977 void LiftoffAssembler::FillStackSlotsWithZero(int start, int size) { in FillStackSlotsWithZero() argument 978 DCHECK_LT(0, size); in FillStackSlotsWithZero() 979 RecordUsedSpillOffset(start + size); in FillStackSlotsWithZero() 981 if (size <= 3 * kStackSlotSize) { in FillStackSlotsWithZero() 984 // And a movd (6-9 byte) when size % 8 != 0; in FillStackSlotsWithZero() 985 uint32_t remainder = size; in FillStackSlotsWithZero() 1000 leaq(rdi, liftoff::GetStackSlot(start + size)); in FillStackSlotsWithZero() 335 LoadFromInstance(Register dst, Register instance, int offset, int size) LoadFromInstance() argument 4157 AllocateStackSlot(Register addr, uint32_t size) AllocateStackSlot() argument 4162 DeallocateStackSlot(uint32_t size) DeallocateStackSlot() argument [all...] |
/third_party/skia/tests/ |
H A D | PathTest.cpp | 840 SkScalar size = rand.nextUScalar1() * 30; in test_arb_round_rect_is_convex() local 841 if (size < SK_Scalar1) { in test_arb_round_rect_is_convex() 846 r.fRight = r.fLeft + 2 * size; in test_arb_round_rect_is_convex() 847 r.fBottom = r.fTop + 2 * size; in test_arb_round_rect_is_convex() 867 SkScalar size = rand.nextUScalar1() * 30; in test_arb_zero_rad_round_rect_is_rect() local 868 if (size < SK_Scalar1) { in test_arb_zero_rad_round_rect_is_rect() 873 r.fRight = r.fLeft + 2 * size; in test_arb_zero_rad_round_rect_is_rect() 874 r.fBottom = r.fTop + 2 * size; in test_arb_zero_rad_round_rect_is_rect() 2584 size_t size = writer.bytesWritten(); in write_and_read_back() local 2585 SkAutoMalloc storage(size); in write_and_read_back() [all...] |
/third_party/skia/third_party/externals/swiftshader/third_party/llvm-10.0/llvm/lib/Target/Mips/ |
H A D | MipsISelLowering.cpp | 97 // If I is a shifted mask, set the size (Size) and the first bit of the 807 // $dst = and ((sra or srl) $src , pos), (2**size - 1) in performANDCombine() 808 // => ext $dst, $src, pos, size in performANDCombine() 816 // Return if the shifted mask does not start at bit 0 or the sum of its size in performANDCombine() 817 // and Pos exceeds the word's size. in performANDCombine() 826 // => cins $dst, $src, pos, size in performANDCombine() 828 // size = population count. in performANDCombine() 841 // SMSize is 'location' (position) in this case, not size. in performANDCombine() 846 // $dst = and $src, (2**size - 1) , if size > 1 in performANDCombine() [all...] |
/applications/standard/contacts_data/jstest/ |
H A D | ContactCard_test.js | 81 async function queryBatch(map, tag, size) 92 expect(resultSet.rowCount == size).assertEqual(true); 2751 async function query(size) 2761 expect(resultSet.rowCount == size).assertEqual(true); 4775 async function query(size) 4787 'contactCard_insertContactBlocklist_test_6800 : query result is = ' + resultSet.rowCount == size); 4788 expect(resultSet.rowCount == size).assertEqual(true); 4822 async function query(size) 4831 expect(resultSet.rowCount == size).assertEqual(true);
|
/foundation/bundlemanager/bundle_framework/services/bundlemgr/src/ |
H A D | inner_bundle_info.cpp | 1961 for (unsigned long i = 0; i < innerModuleInfoVector.size(); i++) { in InsertInnerSharedModuleInfo() 2180 if (info.second.hnpPackages.size() > 0) { in GetApplicationInfo() 2196 if (!isModuleJson && info.second.metaData.customizeData.size() > 0) { in GetApplicationInfo() 2199 if (isModuleJson && info.second.metadata.size() > 0) { in GetApplicationInfo() 2238 if (info.second.hnpPackages.size() > 0) { in GetApplicationInfoV9() 2257 if (!isModuleJson && info.second.metaData.customizeData.size() > 0) { in GetApplicationInfoV9() 2260 if (isModuleJson && info.second.metadata.size() > 0) { in GetApplicationInfoV9() 3807 if (modulesIt->second.size() == SINGLE_HSP_VERSION && in DeleteHspModuleByVersion() 4174 for (size_t i = 0; i < querySchemes.size(); i++) { in GetQuerySchemes()
|
/kernel/linux/linux-5.10/crypto/ |
H A D | testmgr.c | 73 * Need slab memory for testing (size in number of pages). 833 static void flip_random_bit(u8 *buf, size_t size) in flip_random_bit() argument 837 bitpos = prandom_u32() % (size * 8); in flip_random_bit() 842 static void flip_random_byte(u8 *buf, size_t size) in flip_random_byte() argument 844 buf[prandom_u32() % size] ^= 0xff; in flip_random_byte() 848 static void mutate_buffer(u8 *buf, size_t size) in mutate_buffer() argument 855 num_flips = min_t(size_t, 1 << (prandom_u32() % 8), size * 8); in mutate_buffer() 857 flip_random_bit(buf, size); in mutate_buffer() 862 num_flips = min_t(size_t, 1 << (prandom_u32() % 8), size); in mutate_buffer() 864 flip_random_byte(buf, size); in mutate_buffer() [all...] |
/kernel/linux/linux-5.10/drivers/infiniband/hw/mlx5/ |
H A D | main.c | 859 /* IB_WR_REG_MR always requires changing the entity size with UMR */ in mlx5_ib_query_device() 1584 /* Large page with non 4k uar support might limit the dynamic size */ in calc_dynamic_bfregs() 2109 mdm->size); in mlx5_ib_mmap_free() 2243 mdm->size, in add_dm_mmap_entry() 2385 dm->size = roundup(attr->length, MLX5_MEMIC_BASE_SIZE); in handle_alloc_dm_memic() 2388 dm->size, attr->alignment); in handle_alloc_dm_memic() 2417 mlx5_cmd_dealloc_memic(dm_db, dm->dev_addr, dm->size); in handle_alloc_dm_memic() 2432 /* Allocation size must a multiple of the basic block size in handle_alloc_dm_sw_icm() 2438 dm->size in handle_alloc_dm_sw_icm() [all...] |