| /linux/drivers/staging/media/atomisp/pci/ |
| A D | sh_css_defs.h | 331 #define __ISP_MIN_INTERNAL_WIDTH(num_chunks, pipelining, mode) \ argument 332 ((num_chunks) * (pipelining) * (1 << _ISP_LOG_VECTOR_STEP(mode)) * \ 341 #define __ISP_CHUNK_STRIDE_DDR(c_subsampling, num_chunks) \ argument 342 ((c_subsampling) * (num_chunks) * HIVE_ISP_DDR_WORD_BYTES) 348 num_chunks, \ argument 353 __ISP_MIN_INTERNAL_WIDTH(num_chunks, \ 359 __ISP_CHUNK_STRIDE_DDR(c_subsampling, num_chunks) \
|
| A D | ia_css_acc_types.h | 119 u32 num_chunks; member
|
| /linux/drivers/crypto/qat/qat_common/ |
| A D | icp_qat_uclo.h | 251 unsigned short num_chunks; member 266 short num_chunks; member 506 unsigned short num_chunks; member 532 unsigned short num_chunks; member 548 unsigned short num_chunks; member
|
| A D | qat_uclo.c | 112 if (suof_hdr->num_chunks <= 0x1) { in qat_uclo_check_suof_format() 453 for (i = 0; i < obj_hdr->num_chunks; i++) { in qat_uclo_find_chunk() 500 for (i = 0; i < file_hdr->num_chunks; i++) { in qat_uclo_map_chunk() 1175 suof_handle->img_table.num_simgs = suof_ptr->num_chunks - 1; in qat_uclo_map_suof() 1692 uobj_chunk_num = uobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof() 1694 sobj_chunk_num = sobj_hdr->num_chunks; in qat_uclo_map_objs_from_mof() 1774 if (mof_hdr->num_chunks <= 0x1) { in qat_uclo_check_mof_format() 1818 chunks_num = mof_ptr->num_chunks; in qat_uclo_map_mof_obj()
|
| /linux/drivers/misc/habanalabs/common/ |
| A D | command_submission.c | 1129 u32 cs_type_flags, num_chunks; in hl_cs_sanity_checks() local 1165 if (!num_chunks) { in hl_cs_sanity_checks() 1171 } else if (num_chunks != 1) { in hl_cs_sanity_checks() 1183 void __user *chunks, u32 num_chunks, in hl_cs_copy_chunk_array() argument 1188 if (num_chunks > HL_MAX_JOBS_PER_CS) { in hl_cs_copy_chunk_array() 1316 for (i = 0 ; i < num_chunks ; i++) { in cs_ioctl_default() 1459 u32 num_chunks, tmp; in hl_cs_ctx_switch() local 1494 if (!num_chunks) { in hl_cs_ctx_switch() 1513 if (num_chunks) { in hl_cs_ctx_switch() 1950 void __user *chunks, u32 num_chunks, in cs_ioctl_signal_wait() argument [all …]
|
| /linux/drivers/gpu/drm/lima/ |
| A D | lima_sched.c | 334 dt->num_chunks++; in lima_sched_build_error_task_list() 341 dt->num_chunks++; in lima_sched_build_error_task_list() 347 dt->num_chunks++; in lima_sched_build_error_task_list() 386 dt->num_chunks++; in lima_sched_build_error_task_list()
|
| A D | lima_dump.h | 47 __u32 num_chunks; member
|
| /linux/drivers/infiniband/sw/siw/ |
| A D | siw_mem.c | 372 int num_pages, num_chunks, i, rv = 0; in siw_umem_get() local 382 num_chunks = (num_pages >> CHUNK_SHIFT) + 1; in siw_umem_get() 408 kcalloc(num_chunks, sizeof(struct siw_page_chunk), GFP_KERNEL); in siw_umem_get()
|
| /linux/drivers/crypto/marvell/octeontx/ |
| A D | otx_cptvf.h | 37 u32 num_chunks; /* Number of command chunks */ member
|
| A D | otx_cptvf_main.c | 181 queue->num_chunks = 0; in free_command_queues() 229 i, queue->num_chunks); in alloc_command_queues() 234 if (queue->num_chunks == 0) { in alloc_command_queues() 241 queue->num_chunks++; in alloc_command_queues()
|
| /linux/drivers/gpu/drm/radeon/ |
| A D | radeon_cs.c | 279 if (!cs->num_chunks) { in radeon_cs_parser_init() 291 p->chunks_array = kvmalloc_array(cs->num_chunks, sizeof(uint64_t), GFP_KERNEL); in radeon_cs_parser_init() 297 sizeof(uint64_t)*cs->num_chunks)) { in radeon_cs_parser_init() 301 p->nchunks = cs->num_chunks; in radeon_cs_parser_init()
|
| /linux/drivers/net/dsa/sja1105/ |
| A D | sja1105_spi.c | 44 int num_chunks; in sja1105_xfer() local 47 num_chunks = DIV_ROUND_UP(len, priv->max_xfer_len); in sja1105_xfer() 56 for (i = 0; i < num_chunks; i++) { in sja1105_xfer()
|
| /linux/drivers/gpu/drm/amd/amdgpu/ |
| A D | amdgpu_cs.c | 117 if (cs->in.num_chunks == 0) in amdgpu_cs_parser_init() 120 chunk_array = kvmalloc_array(cs->in.num_chunks, sizeof(uint64_t), GFP_KERNEL); in amdgpu_cs_parser_init() 141 sizeof(uint64_t)*cs->in.num_chunks)) { in amdgpu_cs_parser_init() 146 p->nchunks = cs->in.num_chunks; in amdgpu_cs_parser_init()
|
| /linux/drivers/mtd/nand/raw/ |
| A D | mxc_nand.c | 253 u16 num_chunks = mtd->writesize / 512; in copy_spare() local 260 oob_chunk_size = (host->used_oobsize / num_chunks) & ~1; in copy_spare() 263 for (i = 0; i < num_chunks - 1; i++) in copy_spare() 273 for (i = 0; i < num_chunks - 1; i++) in copy_spare()
|
| /linux/net/sctp/ |
| A D | socket.c | 6956 u32 num_chunks = 0; in sctp_getsockopt_peer_auth_chunks() local 6978 num_chunks = ntohs(ch->param_hdr.length) - sizeof(struct sctp_paramhdr); in sctp_getsockopt_peer_auth_chunks() 6979 if (len < num_chunks) in sctp_getsockopt_peer_auth_chunks() 6982 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_peer_auth_chunks() 6985 len = sizeof(struct sctp_authchunks) + num_chunks; in sctp_getsockopt_peer_auth_chunks() 6988 if (put_user(num_chunks, &p->gauth_number_of_chunks)) in sctp_getsockopt_peer_auth_chunks() 7001 u32 num_chunks = 0; in sctp_getsockopt_local_auth_chunks() local 7029 if (len < sizeof(struct sctp_authchunks) + num_chunks) in sctp_getsockopt_local_auth_chunks() 7032 if (copy_to_user(to, ch->chunks, num_chunks)) in sctp_getsockopt_local_auth_chunks() 7035 len = sizeof(struct sctp_authchunks) + num_chunks; in sctp_getsockopt_local_auth_chunks() [all …]
|
| /linux/include/uapi/drm/ |
| A D | amdgpu_drm.h | 568 __u32 num_chunks; member
|
| A D | radeon_drm.h | 984 __u32 num_chunks; member
|
| /linux/drivers/net/wireless/intel/ipw2x00/ |
| A D | ipw2200.c | 3811 if (le32_to_cpu(bd->u.data.num_chunks) > NUM_TFD_CHUNKS) { in ipw_queue_tx_free_tfd() 3813 le32_to_cpu(bd->u.data.num_chunks)); in ipw_queue_tx_free_tfd() 3819 for (i = 0; i < le32_to_cpu(bd->u.data.num_chunks); i++) { in ipw_queue_tx_free_tfd() 10207 tfd->u.data.num_chunks = cpu_to_le32(min((u8) (NUM_TFD_CHUNKS - 2), in ipw_tx_skb() 10210 txb->nr_frags, le32_to_cpu(tfd->u.data.num_chunks)); in ipw_tx_skb() 10211 for (i = 0; i < le32_to_cpu(tfd->u.data.num_chunks); i++) { in ipw_tx_skb() 10213 i, le32_to_cpu(tfd->u.data.num_chunks), in ipw_tx_skb() 10216 i, tfd->u.data.num_chunks, in ipw_tx_skb() 10260 le32_add_cpu(&tfd->u.data.num_chunks, 1); in ipw_tx_skb()
|
| A D | ipw2200.h | 488 __le32 num_chunks; member
|
| /linux/fs/btrfs/ |
| A D | inode.c | 464 atomic_t num_chunks; member 1391 if (atomic_dec_and_test(&async_cow->num_chunks)) in async_cow_free() 1407 u64 num_chunks = DIV_ROUND_UP(end - start, SZ_512K); in cow_file_range_async() local 1417 num_chunks = 1; in cow_file_range_async() 1424 ctx = kvmalloc(struct_size(ctx, chunks, num_chunks), GFP_KERNEL); in cow_file_range_async() 1440 atomic_set(&ctx->num_chunks, num_chunks); in cow_file_range_async() 1442 for (i = 0; i < num_chunks; i++) { in cow_file_range_async()
|
| /linux/drivers/staging/media/ipu3/ |
| A D | ipu3-abi.h | 1559 u32 num_chunks; member
|
| /linux/drivers/staging/media/atomisp/pci/runtime/binary/src/ |
| A D | binary.c | 105 info->output.num_chunks, info->pipeline.pipelining); in ia_css_binary_internal_res()
|