/linux/fs/xfs/libxfs/ |
A D | xfs_iext_tree.c | 454 int *nr_entries) in xfs_iext_split_node() argument 466 *nr_entries = 0; in xfs_iext_split_node() 482 *nr_entries = nr_move; in xfs_iext_split_node() 542 int *nr_entries) in xfs_iext_split_leaf() argument 554 *nr_entries = 0; in xfs_iext_split_leaf() 634 int nr_entries, i; in xfs_iext_insert() local 676 int nr_entries) in xfs_iext_rebalance_node() argument 683 if (nr_entries == 0) in xfs_iext_rebalance_node() 793 int nr_entries) in xfs_iext_rebalance_leaf() argument 800 if (nr_entries == 0) in xfs_iext_rebalance_leaf() [all …]
|
/linux/lib/ |
A D | stackdepot.c | 233 unsigned int nr_entries; in stack_depot_snprint() local 235 nr_entries = stack_depot_fetch(handle, &entries); in stack_depot_snprint() 236 return nr_entries ? stack_trace_snprint(buf, size, entries, nr_entries, in stack_depot_snprint() 251 unsigned int nr_entries; in stack_depot_print() local 253 nr_entries = stack_depot_fetch(stack, &entries); in stack_depot_print() 254 if (nr_entries > 0) in stack_depot_print() 255 stack_trace_print(entries, nr_entries, 0); in stack_depot_print() 316 unsigned int nr_entries, in __stack_depot_save() argument 329 hash = hash_stack(entries, nr_entries); in __stack_depot_save() 338 nr_entries, hash); in __stack_depot_save() [all …]
|
/linux/drivers/dma/dw-edma/ |
A D | dw-edma-v0-debugfs.c | 97 int nr_entries, struct dentry *dir) in dw_edma_debugfs_create_x32() argument 101 for (i = 0; i < nr_entries; i++) { in dw_edma_debugfs_create_x32() 111 int nr_entries; in dw_edma_debugfs_regs_ch() local 124 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_ch() 166 int nr_entries, i; in dw_edma_debugfs_regs_wr() local 173 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_wr() 177 nr_entries = ARRAY_SIZE(debugfs_unroll_regs); in dw_edma_debugfs_regs_wr() 235 int nr_entries, i; in dw_edma_debugfs_regs_rd() local 242 nr_entries = ARRAY_SIZE(debugfs_regs); in dw_edma_debugfs_regs_rd() 272 int nr_entries; in dw_edma_debugfs_regs() local [all …]
|
/linux/drivers/md/persistent-data/ |
A D | dm-btree-remove.c | 58 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in node_shift() local 63 BUG_ON(shift > nr_entries); in node_shift() 67 (nr_entries - shift) * sizeof(__le64)); in node_shift() 70 (nr_entries - shift) * value_size); in node_shift() 75 nr_entries * sizeof(__le64)); in node_shift() 78 nr_entries * value_size); in node_shift() 113 unsigned nr_entries = le32_to_cpu(n->header.nr_entries); in delete_at() local 114 unsigned nr_to_copy = nr_entries - (index + 1); in delete_at() 116 BUG_ON(index >= nr_entries); in delete_at() 128 n->header.nr_entries = cpu_to_le32(nr_entries - 1); in delete_at() [all …]
|
A D | dm-btree.c | 74 uint32_t nr_entries = le32_to_cpu(n->header.nr_entries); in inc_children() local 87 uint32_t nr_entries = le32_to_cpu(node->header.nr_entries); in insert_at() local 90 if (index > nr_entries || in insert_at() 101 node->header.nr_entries = cpu_to_le32(nr_entries + 1); in insert_at() 343 uint32_t flags, nr_entries; in btree_lookup_raw() local 353 nr_entries = le32_to_cpu(ro_node(s)->header.nr_entries); in btree_lookup_raw() 418 uint32_t flags, nr_entries; in dm_btree_lookup_next_single() local 428 nr_entries = le32_to_cpu(n->header.nr_entries); in dm_btree_lookup_next_single() 439 if (i >= nr_entries) { in dm_btree_lookup_next_single() 965 unsigned nr_entries; in get_node_free_space() local [all …]
|
A D | dm-array.c | 26 __le32 nr_entries; member 113 unsigned nr_entries = le32_to_cpu(ab->nr_entries); in on_entries() local 162 (*ab)->nr_entries = cpu_to_le32(0); in alloc_ablock() 176 uint32_t nr_entries, delta, i; in fill_ablock() local 182 nr_entries = le32_to_cpu(ab->nr_entries); in fill_ablock() 183 delta = new_nr - nr_entries; in fill_ablock() 199 uint32_t nr_entries, delta; in trim_ablock() local 205 nr_entries = le32_to_cpu(ab->nr_entries); in trim_ablock() 206 delta = nr_entries - new_nr; in trim_ablock() 866 unsigned nr_entries, max_entries; in walk_ablock() local [all …]
|
/linux/kernel/ |
A D | stacktrace.c | 24 void stack_trace_print(const unsigned long *entries, unsigned int nr_entries, in stack_trace_print() argument 32 for (i = 0; i < nr_entries; i++) in stack_trace_print() 48 unsigned int nr_entries, int spaces) in stack_trace_snprint() argument 55 for (i = 0; i < nr_entries && size; i++) { in stack_trace_snprint() 280 return trace.nr_entries; in stack_trace_save() 305 return trace.nr_entries; in stack_trace_save_tsk() 327 return trace.nr_entries; in stack_trace_save_regs() 352 return ret ? ret : trace.nr_entries; in stack_trace_save_tsk_reliable() 372 return trace.nr_entries; in stack_trace_save_user() 397 for (i = 0; i < nr_entries; i++) { in filter_irq_stacks() [all …]
|
/linux/arch/x86/kernel/ |
A D | e820.c | 82 for (i = 0; i < table->nr_entries; i++) { in _e820__mapped_any() 168 int x = table->nr_entries; in __e820__range_add() 179 table->nr_entries++; in __e820__range_add() 332 if (table->nr_entries < 2) in e820__update_table() 417 table->nr_entries = new_nr_entries; in e820__update_table() 426 while (nr_entries) { in __append_e820_table() 439 nr_entries--; in __append_e820_table() 456 if (nr_entries < 2) in append_e820_table() 624 int i = e820_table->nr_entries; in e820_search_gap() 927 e820_table->nr_entries = 0; in parse_memmap_one() [all …]
|
A D | ldt.c | 81 set_ldt(ldt_slot_va(ldt->slot), ldt->nr_entries); in load_mm_ldt() 83 set_ldt(ldt->entries, ldt->nr_entries); in load_mm_ldt() 183 new_ldt->nr_entries = num_entries; in alloc_ldt_struct() 309 nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); in map_ldt_struct() 416 paravirt_alloc_ldt(ldt->entries, ldt->nr_entries); in finalize_ldt_struct() 437 paravirt_free_ldt(ldt->entries, ldt->nr_entries); in free_ldt_struct() 438 if (ldt->nr_entries * LDT_ENTRY_SIZE > PAGE_SIZE) in free_ldt_struct() 461 new_ldt = alloc_ldt_struct(old_mm->context.ldt->nr_entries); in ldt_dup_context() 468 new_ldt->nr_entries * LDT_ENTRY_SIZE); in ldt_dup_context() 516 entries_size = mm->context.ldt->nr_entries * LDT_ENTRY_SIZE; in read_ldt() [all …]
|
/linux/tools/perf/util/ |
A D | syscalltbl.c | 65 int nr_entries = 0, i, j; in syscalltbl__init_native() local 70 ++nr_entries; in syscalltbl__init_native() 72 entries = tbl->syscalls.entries = malloc(sizeof(struct syscall) * nr_entries); in syscalltbl__init_native() 84 qsort(tbl->syscalls.entries, nr_entries, sizeof(struct syscall), syscallcmp); in syscalltbl__init_native() 85 tbl->syscalls.nr_entries = nr_entries; in syscalltbl__init_native() 116 tbl->syscalls.nr_entries, sizeof(*sc), in syscalltbl__id() 127 for (i = *idx + 1; i < tbl->syscalls.nr_entries; ++i) { in syscalltbl__strglobmatch_next()
|
/linux/tools/perf/ui/ |
A D | browser.c | 108 if (browser->nr_entries == 0) in ui_browser__list_head_seek() 316 if (browser->nr_entries > 1) { in ui_browser__scrollbar_set() 318 (browser->nr_entries - 1)); in ui_browser__scrollbar_set() 370 off_t offset = nr_entries - browser->nr_entries; in ui_browser__update_nr_entries() 372 browser->nr_entries = nr_entries; in ui_browser__update_nr_entries() 424 if (browser->index == browser->nr_entries - 1) in ui_browser__run() 460 offset = browser->nr_entries - 1 - browser->index; in ui_browser__run() 483 if (offset >= browser->nr_entries) in ui_browser__run() 484 offset = browser->nr_entries - 1; in ui_browser__run() 486 browser->index = browser->nr_entries - 1; in ui_browser__run() [all …]
|
/linux/tools/perf/trace/beauty/ |
A D | ioctl.c | 41 if (nr < strarray__ioctl_tty_cmd.nr_entries && strarray__ioctl_tty_cmd.entries[nr] != NULL) in ioctl__scnprintf_tty_cmd() 52 if (nr < strarray__drm_ioctl_cmds.nr_entries && strarray__drm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_drm_cmd() 63 …if (nr < strarray__sndrv_pcm_ioctl_cmds.nr_entries && strarray__sndrv_pcm_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_pcm_cmd() 74 …if (nr < strarray__sndrv_ctl_ioctl_cmds.nr_entries && strarray__sndrv_ctl_ioctl_cmds.entries[nr] !… in ioctl__scnprintf_sndrv_ctl_cmd() 85 if (nr < strarray__kvm_ioctl_cmds.nr_entries && strarray__kvm_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_kvm_cmd() 98 if (nr < s->nr_entries && s->entries[nr] != NULL) in ioctl__scnprintf_vhost_virtio_cmd() 109 if (nr < strarray__perf_ioctl_cmds.nr_entries && strarray__perf_ioctl_cmds.entries[nr] != NULL) in ioctl__scnprintf_perf_cmd() 120 …if (nr < strarray__usbdevfs_ioctl_cmds.nr_entries && strarray__usbdevfs_ioctl_cmds.entries[nr] != … in ioctl__scnprintf_usbdevfs_cmd()
|
/linux/arch/mips/kernel/ |
A D | stacktrace.c | 30 trace->entries[trace->nr_entries++] = addr; in save_raw_context_stack() 31 if (trace->nr_entries >= trace->max_entries) in save_raw_context_stack() 58 trace->entries[trace->nr_entries++] = pc; in save_context_stack() 59 if (trace->nr_entries >= trace->max_entries) in save_context_stack() 83 WARN_ON(trace->nr_entries || !trace->max_entries); in save_stack_trace_tsk()
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
A D | spectrum1_kvdl.c | 122 unsigned int entry_index, nr_entries; in mlxsw_sp1_kvdl_part_alloc() local 124 nr_entries = (info->end_index - info->start_index + 1) / in mlxsw_sp1_kvdl_part_alloc() 126 entry_index = find_first_zero_bit(part->usage, nr_entries); in mlxsw_sp1_kvdl_part_alloc() 127 if (entry_index == nr_entries) in mlxsw_sp1_kvdl_part_alloc() 215 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_init() local 227 nr_entries = div_u64(resource_size, info->alloc_size); in mlxsw_sp1_kvdl_part_init() 228 usage_size = BITS_TO_LONGS(nr_entries) * sizeof(unsigned long); in mlxsw_sp1_kvdl_part_init() 281 unsigned int nr_entries; in mlxsw_sp1_kvdl_part_occ() local 285 nr_entries = (info->end_index - in mlxsw_sp1_kvdl_part_occ() 288 while ((bit = find_next_bit(part->usage, nr_entries, bit + 1)) in mlxsw_sp1_kvdl_part_occ() [all …]
|
/linux/drivers/md/ |
A D | dm-cache-policy-internal.h | 115 static inline size_t bitset_size_in_bytes(unsigned nr_entries) in bitset_size_in_bytes() argument 117 return sizeof(unsigned long) * dm_div_up(nr_entries, BITS_PER_LONG); in bitset_size_in_bytes() 120 static inline unsigned long *alloc_bitset(unsigned nr_entries) in alloc_bitset() argument 122 size_t s = bitset_size_in_bytes(nr_entries); in alloc_bitset() 126 static inline void clear_bitset(void *bitset, unsigned nr_entries) in clear_bitset() argument 128 size_t s = bitset_size_in_bytes(nr_entries); in clear_bitset()
|
/linux/arch/sh/kernel/ |
A D | stacktrace.c | 33 if (trace->nr_entries < trace->max_entries) in save_stack_address() 34 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 65 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 66 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
/linux/arch/openrisc/kernel/ |
A D | stacktrace.c | 38 if (trace->nr_entries < trace->max_entries) in save_stack_address() 39 trace->entries[trace->nr_entries++] = addr; in save_stack_address() 64 if (trace->nr_entries < trace->max_entries) in save_stack_address_nosched() 65 trace->entries[trace->nr_entries++] = addr; in save_stack_address_nosched()
|
/linux/include/linux/ |
A D | stacktrace.h | 12 void stack_trace_print(const unsigned long *trace, unsigned int nr_entries, 15 unsigned int nr_entries, int spaces); 24 unsigned int filter_irq_stacks(unsigned long *entries, unsigned int nr_entries); 82 unsigned int nr_entries, max_entries; member
|
/linux/tools/testing/selftests/bpf/benchs/ |
A D | bench_bloom_filter_map.c | 39 __u32 nr_entries; member 43 .nr_entries = 1000, 68 args.nr_entries = strtol(arg, NULL, 10); in parse_arg() 69 if (args.nr_entries == 0) { in parse_arg() 140 if (i > args.nr_entries) in map_prepare_thread() 241 if (args.nr_entries > nr_unique_entries) { in check_args() 267 bpf_map__set_max_entries(skel->maps.hashmap, args.nr_entries); in setup_skeleton() 269 bpf_map__set_max_entries(skel->maps.array_map, args.nr_entries); in setup_skeleton() 271 bpf_map__set_max_entries(skel->maps.bloom_map, args.nr_entries); in setup_skeleton()
|
/linux/tools/perf/tests/ |
A D | parse-events.c | 53 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_tracepoint() 66 TEST_ASSERT_VAL("wrong number of entries", evlist->core.nr_entries > 1); in test__checkevent_tracepoint_multi() 84 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_raw() 94 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_numeric() 104 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_symbolic_name() 115 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_symbolic_name_config() 136 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_symbolic_alias() 147 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_genhw() 157 TEST_ASSERT_VAL("wrong number of entries", 1 == evlist->core.nr_entries); in test__checkevent_breakpoint() 242 TEST_ASSERT_VAL("wrong number of entries", evlist->core.nr_entries > 1); in test__checkevent_tracepoint_multi_modifier() [all …]
|
A D | hists_filter.c | 155 hists->nr_entries == 9); in test__hists_filter() 162 hists->nr_entries == hists->nr_non_filtered_entries); in test__hists_filter() 180 hists->nr_entries == 9); in test__hists_filter() 209 hists->nr_entries == 9); in test__hists_filter() 244 hists->nr_entries == 9); in test__hists_filter() 273 hists->nr_entries == 9); in test__hists_filter() 304 hists->nr_entries == 9); in test__hists_filter()
|
/linux/arch/sparc/kernel/ |
A D | stacktrace.c | 58 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 66 if (trace->nr_entries < in __save_stack_trace() 68 trace->entries[trace->nr_entries++] = pc; in __save_stack_trace() 74 } while (trace->nr_entries < trace->max_entries); in __save_stack_trace()
|
/linux/tools/perf/arch/x86/util/ |
A D | iostat.c | 70 int nr_entries; member 102 for (idx = 0; idx < list->nr_entries; idx++) in iio_root_ports_list_free() 116 for (idx = 0; idx < list->nr_entries; idx++) { in iio_root_port_find_by_notation() 131 rp->idx = list->nr_entries++; in iio_root_ports_list_insert() 133 list->nr_entries * sizeof(*list->rps)); in iio_root_ports_list_insert() 282 if (tmp_list->nr_entries == 0) { in iio_root_ports_list_filter() 315 for (idx = 0; idx < list->nr_entries; idx++) { in iostat_event_group() 327 list->nr_entries = 0; in iostat_event_group() 336 if (evlist->core.nr_entries > 0) { in iostat_prepare()
|
/linux/drivers/pci/pcie/ |
A D | portdrv_core.c | 103 int nr_entries, nvec, pcie_irq; in pcie_port_enable_irq_vec() local 107 nr_entries = pci_alloc_irq_vectors(dev, 1, PCIE_PORT_MAX_MSI_ENTRIES, in pcie_port_enable_irq_vec() 109 if (nr_entries < 0) in pcie_port_enable_irq_vec() 110 return nr_entries; in pcie_port_enable_irq_vec() 114 if (nvec > nr_entries) { in pcie_port_enable_irq_vec() 130 if (nvec != nr_entries) { in pcie_port_enable_irq_vec() 133 nr_entries = pci_alloc_irq_vectors(dev, nvec, nvec, in pcie_port_enable_irq_vec() 135 if (nr_entries < 0) in pcie_port_enable_irq_vec() 136 return nr_entries; in pcie_port_enable_irq_vec()
|
/linux/arch/ia64/kernel/ |
A D | stacktrace.c | 19 trace->nr_entries = 0; in ia64_do_save_stack() 25 trace->entries[trace->nr_entries++] = ip; in ia64_do_save_stack() 26 if (trace->nr_entries == trace->max_entries) in ia64_do_save_stack()
|