/linux-master/tools/perf/scripts/python/Perf-Trace-Util/lib/Perf/Trace/ |
H A D | Core.py | 35 for idx in sorted(flag_fields[event_name][field_name]['values']): 52 for idx in sorted(symbolic_fields[event_name][field_name]['values']):
|
/linux-master/tools/perf/scripts/python/ |
H A D | event_analyzing_sample.py | 192 print (' '.join(['%s=%s'%(k,str(v))for k,v in sorted(event_fields_dict.items())]))
|
H A D | failed-syscalls-by-pid.py | 78 for ret, val in sorted(syscalls[comm][pid][id].items(), key = lambda kv: (kv[1], kv[0]), reverse = True):
|
H A D | mem-phys-addr.py | 49 for mem_type, count in sorted(load_mem_type_cnt.most_common(), \ 63 #/proc/iomem is sorted
|
H A D | sctop.py | 81 for id, val in sorted(syscalls.items(),
|
H A D | stackcollapse.py | 10 # The file is sorted according to the first field. 125 list = sorted(lines)
|
H A D | syscall-counts-by-pid.py | 73 for id, val in sorted(syscalls[comm][pid].items(),
|
H A D | syscall-counts.py | 63 for id, val in sorted(syscalls.items(),
|
/linux-master/Documentation/sphinx/ |
H A D | automarkup.py | 107 sorted_matches = sorted(chain(*match_iterators), key=lambda m: m.start())
|
H A D | cdomain.py | 114 matches = sorted(chain(*match_iterators), key=lambda m: m.start())
|
/linux-master/drivers/comedi/drivers/ni_routing/tools/ |
H A D | convert_csv_to_c.py | 40 D0 = sorted(D0.items(), key=lambda i: eval(i[0], comedi_h.__dict__, Locals)) 43 D1 = sorted(D1_D.items(), key=lambda i: eval(i[0], comedi_h.__dict__, Locals)) 89 D0 = sorted(D0.items(), key=lambda i: eval(i[0], comedi_h.__dict__, Locals)) 233 sheets = sorted(self.items(), key=lambda i : tuple(i[0].split('-')[::-1]) ) 421 sheets = sorted(self.items(), key=lambda i : i[0] )
|
H A D | convert_py_to_csv.py | 23 fieldnames = [value_to_name[i] for i in sorted(D.keys())] 31 S = sorted(S.items(), key = lambda src_destD : src_destD[0])
|
H A D | make_blank_csv.py | 20 fieldnames = [sig for sig_val, sig in sorted(value_to_name.items())]
|
/linux-master/drivers/gpu/drm/amd/display/dc/dml2/ |
H A D | dml2_dc_resource_mgmt.c | 453 bool sorted, swapped; local 463 sorted = true; 465 sorted = false; 469 while (!sorted) { 484 sorted = false; 486 sorted = true;
|
/linux-master/drivers/gpu/drm/i915/gt/uc/ |
H A D | selftest_guc_multi_lrc.c | 15 struct intel_engine_cs *sorted[MAX_ENGINE_INSTANCE + 1]; local 21 sorted[i] = engines[j]; 26 memcpy(*engines, *sorted,
|
/linux-master/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
H A D | gf100.c | 1921 bool sorted; local 1947 sorted = false; 1949 while (!sorted) { 1950 for (sorted = true, i = 0; i < gr->gpc_nr - 1; i++) { 1956 sorted = false;
|
/linux-master/drivers/md/bcache/ |
H A D | debug.c | 34 struct bset *ondisk, *sorted, *inmemory; local 44 sorted = b->c->verify_data->keys.set->data; 57 bch_bio_map(bio, sorted); 62 memcpy(ondisk, sorted, KEY_SIZE(&v->key) << 9); 65 sorted = v->keys.set->data; 67 if (inmemory->keys != sorted->keys || 69 sorted->start, 81 bch_dump_bset(&v->keys, sorted, 0); 95 if (inmemory->d[j] != sorted->d[j])
|
/linux-master/fs/bcachefs/ |
H A D | btree_io.c | 477 * Returns true if we sorted (i.e. invalidated iterators 980 struct btree_node *sorted; local 1185 sorted = btree_bounce_alloc(c, btree_buf_bytes(b), &used_mempool); 1186 sorted->keys.u64s = 0; 1190 b->nr = bch2_key_sort_fix_overlapping(c, &sorted->keys, iter); 1192 u64s = le16_to_cpu(sorted->keys.u64s); 1193 *sorted = *b->data; 1194 sorted->keys.u64s = cpu_to_le16(u64s); 1195 swap(sorted, b->data); 1201 btree_bounce_free(c, btree_buf_bytes(b), used_mempool, sorted); [all...] |
H A D | btree_iter.c | 1017 btree_path_idx_t idx = trans->sorted[i]; 1552 btree_path_idx_t *sorted = p; local 1553 memcpy(sorted, trans->sorted, trans->nr_sorted * sizeof(btree_path_idx_t)); 1563 rcu_assign_pointer(trans->sorted, sorted); 2610 BUG_ON(trans->sorted[path->sorted_idx] != i); 2614 unsigned idx = trans->sorted[i]; 2654 * mostly sorted. 2662 if (btree_path_cmp(trans->paths + trans->sorted[ [all...] |
H A D | btree_iter.h | 125 ? trans->paths + trans->sorted[idx] 134 ? trans->paths + trans->sorted[idx - 1] 140 (_iter.path_idx = trans->sorted[_iter.sorted_idx], \ 151 (_iter.path_idx = trans->sorted[_iter.sorted_idx], \ 158 ((_path) = (_trans)->paths + trans->sorted[_i]), (_i) >= 0;\
|
H A D | btree_types.h | 90 * Sets of sorted keys - the real btree node - plus a binary search tree 406 btree_path_idx_t *sorted; member in struct:btree_trans
|
H A D | btree_write_buffer.c | 217 darray_resize(&wb->sorted, wb->flushing.keys.size); 219 if (!wb->flushing.keys.nr && wb->sorted.size >= wb->inc.keys.nr) { 225 wb->sorted.size - wb->flushing.keys.nr); 251 BUG_ON(wb->sorted.size < wb->flushing.keys.nr); 272 wb->sorted.data[i].idx = i; 273 wb->sorted.data[i].btree = wb->flushing.keys.data[i].btree; 274 memcpy(&wb->sorted.data[i].pos, &wb->flushing.keys.data[i].k.k.p, sizeof(struct bpos)); 276 wb->sorted.nr = wb->flushing.keys.nr; 280 * then we attempt to flush in sorted btree order, as this is most 292 wb_sort(wb->sorted [all...] |
H A D | btree_write_buffer_types.h | 51 DARRAY(struct wb_key_ref) sorted; member in struct:btree_write_buffer
|
H A D | debug.c | 39 struct bset *sorted, *inmemory = &b->data->keys; local 67 sorted = &n_sorted->keys; 69 if (inmemory->u64s != sorted->u64s || 71 sorted->start, 83 bch2_dump_bset(c, v, sorted, 0); 106 if (inmemory->_data[j] != sorted->_data[j])
|
H A D | disk_groups.c | 27 struct bch_disk_group *g, *sorted = NULL; local 69 sorted = kmalloc_array(nr_groups, sizeof(*sorted), GFP_KERNEL); 70 if (!sorted) 73 memcpy(sorted, groups->entries, nr_groups * sizeof(*sorted)); 74 sort(sorted, nr_groups, sizeof(*sorted), group_cmp, NULL); 76 for (g = sorted; g + 1 < sorted [all...] |