/linux-master/arch/arm/common/ |
H A D | mcpm_entry.c | 3 * arch/arm/common/mcpm_entry.c -- entry point for multi-cluster PM 24 * see Documentation/arch/arm/cluster-pm-race-avoidance.rst. 34 static void __mcpm_cpu_going_down(unsigned int cpu, unsigned int cluster) argument 36 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN; 37 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); 42 * cluster can be torn down without disrupting this CPU. 47 static void __mcpm_cpu_down(unsigned int cpu, unsigned int cluster) argument 50 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN; 51 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu); 56 * __mcpm_outbound_leave_critical: Leave the cluster teardow 63 __mcpm_outbound_leave_critical(unsigned int cluster, int state) argument 82 __mcpm_outbound_enter_critical(unsigned int cpu, unsigned int cluster) argument 135 __mcpm_cluster_state(unsigned int cluster) argument 143 mcpm_set_entry_vector(unsigned cpu, unsigned cluster, void *ptr) argument 152 mcpm_set_early_poke(unsigned cpu, unsigned cluster, unsigned long poke_phys_addr, unsigned long poke_val) argument 187 mcpm_cluster_unused(unsigned int cluster) argument 195 mcpm_cpu_power_up(unsigned int cpu, unsigned int cluster) argument 241 unsigned int mpidr, cpu, cluster; local 308 mcpm_wait_for_cpu_powerdown(unsigned int cpu, unsigned int cluster) argument 332 unsigned int cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); local 342 unsigned int mpidr, cpu, cluster; local 378 unsigned int cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); local [all...] |
H A D | bL_switcher_dummy_if.c | 22 unsigned int cpu, cluster; local 33 /* format: <cpu#>,<cluster#> */ 40 cluster = val[2] - '0'; 41 ret = bL_switch_request(cpu, cluster);
|
H A D | mcpm_head.S | 3 * arch/arm/common/mcpm_head.S -- kernel entry point for multi-cluster PM 8 * Refer to Documentation/arch/arm/cluster-pm-race-avoidance.rst 28 1903: .asciz " cluster" 56 ubfx r10, r0, #8, #8 @ r10 = cluster 88 mla r8, r0, r10, r8 @ r8 = sync cluster base 96 @ At this point, the cluster cannot unexpectedly enter the GOING_DOWN 100 mla r11, r0, r10, r11 @ r11 = cluster first man lock 106 bne mcpm_setup_wait @ wait for cluster setup if so 109 cmp r0, #CLUSTER_UP @ cluster already up? 110 bne mcpm_setup @ if not, set up the cluster [all...] |
H A D | bL_switcher.c | 3 * arch/arm/common/bL_switcher.c -- big.LITTLE cluster switcher core driver 118 * with the cluster number. 141 * bL_switch_to - Switch to a specific cluster for the current CPU 142 * @new_cluster_id: the ID of the cluster to switch to. 273 int cluster; local 288 cluster = t->wanted_cluster; 295 if (cluster != -1) { 296 bL_switch_to(cluster); 321 * bL_switch_request_cb - Switch to a specific cluster for the given CPU, 325 * @new_cluster_id: the ID of the cluster t 422 unsigned int cpu, cluster, mask; local 604 unsigned int cpu, cluster; local [all...] |
/linux-master/arch/arm/include/asm/ |
H A D | mcpm.h | 13 * Maximum number of possible clusters / CPUs per cluster. 39 * This is used to indicate where the given CPU from given cluster should 44 void mcpm_set_entry_vector(unsigned cpu, unsigned cluster, void *ptr); 51 void mcpm_set_early_poke(unsigned cpu, unsigned cluster, 55 * CPU/cluster power operations API for higher subsystems to use. 66 * mcpm_cpu_power_up - make given CPU in given cluster runable 68 * @cpu: CPU number within given cluster 69 * @cluster: cluster number for the CPU 71 * The identified CPU is brought out of reset. If the cluster wa [all...] |
/linux-master/arch/arm/mach-versatile/ |
H A D | dcscb.c | 39 static int dcscb_cpu_powerup(unsigned int cpu, unsigned int cluster) argument 43 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); 44 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster])) 47 rst_hold = readl_relaxed(dcscb_base + RST_HOLD0 + cluster * 4); 49 writel_relaxed(rst_hold, dcscb_base + RST_HOLD0 + cluster * 4); 53 static int dcscb_cluster_powerup(unsigned int cluster) argument 57 pr_debug("%s: cluster %u\n", __func__, cluster); 69 dcscb_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) argument 81 dcscb_cluster_powerdown_prepare(unsigned int cluster) argument [all...] |
H A D | spc.h | 13 void ve_spc_cpu_wakeup_irq(u32 cluster, u32 cpu, bool set); 14 void ve_spc_set_resume_addr(u32 cluster, u32 cpu, u32 addr); 15 void ve_spc_powerdown(u32 cluster, bool enable); 16 int ve_spc_cpu_in_wfi(u32 cpu, u32 cluster);
|
H A D | tc2_pm.c | 46 static int tc2_pm_cpu_powerup(unsigned int cpu, unsigned int cluster) argument 48 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); 49 if (cluster >= TC2_CLUSTERS || cpu >= tc2_nr_cpus[cluster]) 51 ve_spc_set_resume_addr(cluster, cpu, 53 ve_spc_cpu_wakeup_irq(cluster, cpu, true); 57 static int tc2_pm_cluster_powerup(unsigned int cluster) argument 59 pr_debug("%s: cluster %u\n", __func__, cluster); 66 tc2_pm_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) argument 81 tc2_pm_cluster_powerdown_prepare(unsigned int cluster) argument 112 tc2_core_in_reset(unsigned int cpu, unsigned int cluster) argument 124 tc2_pm_wait_for_powerdown(unsigned int cpu, unsigned int cluster) argument 155 tc2_pm_cpu_suspend_prepare(unsigned int cpu, unsigned int cluster) argument 160 tc2_pm_cpu_is_up(unsigned int cpu, unsigned int cluster) argument 168 tc2_pm_cluster_is_up(unsigned int cluster) argument 202 unsigned int mpidr, cpu, cluster; local [all...] |
H A D | spc.c | 50 /* SPC CPU/cluster reset statue */ 71 /* TC2 static dual-cluster configuration */ 97 * A15s cluster identifier 111 static inline bool cluster_is_a15(u32 cluster) argument 113 return cluster == info->a15_clusid; 142 * @cluster: mpidr[15:8] bitfield describing cluster affinity level 150 void ve_spc_cpu_wakeup_irq(u32 cluster, u32 cpu, bool set) argument 154 if (cluster >= MAX_CLUSTERS) 159 if (!cluster_is_a15(cluster)) 179 ve_spc_set_resume_addr(u32 cluster, u32 cpu, u32 addr) argument 204 ve_spc_powerdown(u32 cluster, bool enable) argument 215 standbywfi_cpu_mask(u32 cpu, u32 cluster) argument 234 ve_spc_cpu_in_wfi(u32 cpu, u32 cluster) argument 250 ve_spc_get_performance(int cluster, u32 *freq) argument 269 ve_spc_round_performance(int cluster, u32 freq) argument 292 ve_spc_find_performance_index(int cluster, u32 freq) argument 314 ve_spc_set_performance(int cluster, u32 freq) argument 393 ve_spc_populate_opps(uint32_t cluster) argument 421 int cluster; local 484 int cluster; member in struct:clk_spc 546 int cpu, cluster; local [all...] |
/linux-master/arch/arm/mach-sunxi/ |
H A D | mc_smp.c | 87 static bool sunxi_core_is_cortex_a15(unsigned int core, unsigned int cluster) argument 90 int cpu = cluster * SUNXI_CPUS_PER_CLUSTER + core; 102 * would be mid way in a core or cluster power sequence. 104 pr_err("%s: Couldn't get CPU cluster %u core %u device node\n", 105 __func__, cluster, core); 115 static int sunxi_cpu_power_switch_set(unsigned int cpu, unsigned int cluster, argument 121 reg = readl(prcm_base + PRCM_PWR_SWITCH_REG(cluster, cpu)); 124 pr_debug("power clamp for cluster %u cpu %u already open\n", 125 cluster, cpu); 129 writel(0xff, prcm_base + PRCM_PWR_SWITCH_REG(cluster, cp 158 sunxi_cpu_powerup(unsigned int cpu, unsigned int cluster) argument 253 sunxi_cluster_powerup(unsigned int cluster) argument 374 sunxi_mc_smp_cluster_is_down(unsigned int cluster) argument 393 unsigned int mpidr, cpu, cluster; local 430 unsigned int cluster = MPIDR_AFFINITY_LEVEL(read_cpuid_mpidr(), 1); local 445 unsigned int mpidr, cpu, cluster; local 480 sunxi_cpu_powerdown(unsigned int cpu, unsigned int cluster) argument 504 sunxi_cluster_powerdown(unsigned int cluster) argument 535 unsigned int mpidr, cpu, cluster; local 630 unsigned int mpidr, cpu, cluster; local [all...] |
/linux-master/arch/arm/mach-exynos/ |
H A D | mcpm-exynos.c | 57 static int exynos_cpu_powerup(unsigned int cpu, unsigned int cluster) argument 59 unsigned int cpunr = cpu + (cluster * EXYNOS5420_CPUS_PER_CLUSTER); 62 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster); 64 cluster >= EXYNOS5420_NR_CLUSTERS) 71 * This assumes the cluster number of the big cores(Cortex A15) 76 if (cluster && 77 cluster == MPIDR_AFFINITY_LEVEL(cpu_logical_map(0), 1)) { 92 pr_err("cpu %u cluster %u powerup failed\n", 93 cpu, cluster); 106 exynos_cluster_powerup(unsigned int cluster) argument 116 exynos_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) argument 126 exynos_cluster_powerdown_prepare(unsigned int cluster) argument 163 exynos_wait_for_powerdown(unsigned int cpu, unsigned int cluster) argument 184 exynos_cpu_is_up(unsigned int cpu, unsigned int cluster) argument [all...] |
H A D | common.h | 140 extern void exynos_cluster_power_down(int cluster); 141 extern void exynos_cluster_power_up(int cluster); 142 extern int exynos_cluster_power_state(int cluster);
|
/linux-master/arch/arm/mach-hisi/ |
H A D | platmcpm.c | 71 static bool hip04_cluster_is_down(unsigned int cluster) argument 76 if (hip04_cpu_table[cluster][i]) 81 static void hip04_set_snoop_filter(unsigned int cluster, unsigned int on) argument 89 data |= 1 << cluster; 91 data &= ~(1 << cluster); 100 unsigned int mpidr, cpu, cluster; local 106 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); 110 if (cluster >= HIP04_MAX_CLUSTERS || cpu >= HIP04_MAX_CPUS_PER_CLUSTER) 115 if (hip04_cpu_table[cluster][cpu]) 118 sys_dreq = sysctrl + SC_CPU_RESET_DREQ(cluster); 155 unsigned int mpidr, cpu, cluster; local 193 unsigned int mpidr, cpu, cluster; local 249 unsigned int mpidr, cpu, cluster; local [all...] |
/linux-master/fs/ocfs2/cluster/ |
H A D | nodemanager.c | 18 * cluster active at a time. Changing this will require trickling 19 * cluster references throughout where nodes are looked up */ 49 struct o2nm_cluster *cluster = o2nm_single_cluster; local 51 BUG_ON(bytes < (sizeof(cluster->cl_nodes_bitmap))); 53 if (cluster == NULL) 56 read_lock(&cluster->cl_nodes_lock); 57 bitmap_copy(map, cluster->cl_nodes_bitmap, O2NM_MAX_NODES); 58 read_unlock(&cluster->cl_nodes_lock); 64 static struct o2nm_node *o2nm_node_ip_tree_lookup(struct o2nm_cluster *cluster, argument 69 struct rb_node **p = &cluster 102 struct o2nm_cluster *cluster = o2nm_single_cluster; local 187 struct o2nm_cluster *cluster; local 271 struct o2nm_cluster *cluster; local 326 struct o2nm_cluster *cluster; local 453 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local 490 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local 534 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local 606 struct o2nm_cluster *cluster = to_o2nm_cluster(group->cg_item.ci_parent); local 654 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local 688 struct o2nm_cluster *cluster = NULL; local 734 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local [all...] |
/linux-master/drivers/perf/ |
H A D | qcom_l2_pmu.c | 121 * The cache is made up of one or more clusters, each cluster has its own PMU. 122 * Each cluster is associated with one or more CPUs. 143 /* The CPU that is used for collecting events on this cluster */ 145 /* All the CPUs associated with this cluster */ 243 static void cluster_pmu_set_resr(struct cluster_pmu *cluster, argument 254 spin_lock_irqsave(&cluster->pmu_lock, flags); 262 spin_unlock_irqrestore(&cluster->pmu_lock, flags); 268 * all CPUS, subunits and ID independent events in this cluster. 319 static void l2_cache_cluster_set_period(struct cluster_pmu *cluster, argument 339 static int l2_cache_get_event_idx(struct cluster_pmu *cluster, argument 374 l2_cache_clear_event_idx(struct cluster_pmu *cluster, struct perf_event *event) argument 387 struct cluster_pmu *cluster = data; local 441 struct cluster_pmu *cluster; local 543 struct cluster_pmu *cluster; local 593 struct cluster_pmu *cluster; local 618 struct cluster_pmu *cluster; local 739 struct cluster_pmu *cluster; local 769 struct cluster_pmu *cluster; local 804 struct cluster_pmu *cluster; local 844 struct cluster_pmu *cluster; local [all...] |
/linux-master/drivers/remoteproc/ |
H A D | ti_k3_r5_remoteproc.c | 73 * All cluster mode values are not applicable on all SoCs. The following 105 * @cores: list of R5 cores within the cluster 152 * @cluster: cached pointer to parent cluster structure 162 struct k3_r5_cluster *cluster; member in struct:k3_r5_rproc 282 static int k3_r5_lockstep_reset(struct k3_r5_cluster *cluster) argument 288 list_for_each_entry(core, &cluster->cores, elem) { 299 list_for_each_entry(core, &cluster->cores, elem) { 312 list_for_each_entry_continue_reverse(core, &cluster->cores, elem) { 317 core = list_last_entry(&cluster 327 k3_r5_lockstep_release(struct k3_r5_cluster *cluster) argument 441 struct k3_r5_cluster *cluster = kproc->cluster; local 508 struct k3_r5_cluster *cluster = kproc->cluster; local 543 struct k3_r5_cluster *cluster = kproc->cluster; local 615 struct k3_r5_cluster *cluster = kproc->cluster; local 834 struct k3_r5_cluster *cluster = kproc->cluster; local 1072 struct k3_r5_cluster *cluster = kproc->cluster; local 1111 struct k3_r5_cluster *cluster = kproc->cluster; local 1209 struct k3_r5_cluster *cluster = platform_get_drvdata(pdev); local 1316 struct k3_r5_cluster *cluster = platform_get_drvdata(data); local 1639 struct k3_r5_cluster *cluster = platform_get_drvdata(data); local 1652 struct k3_r5_cluster *cluster = platform_get_drvdata(pdev); local 1694 struct k3_r5_cluster *cluster; local [all...] |
H A D | mtk_scp.c | 71 struct mtk_scp_of_cluster *scp_cluster = scp->cluster; 161 val = readl(scp->cluster->reg_base + MT8183_SW_RSTN); 163 writel(val, scp->cluster->reg_base + MT8183_SW_RSTN); 170 val = readl(scp->cluster->reg_base + MT8183_SW_RSTN); 172 writel(val, scp->cluster->reg_base + MT8183_SW_RSTN); 177 writel(1, scp->cluster->reg_base + MT8192_CORE0_SW_RSTN_SET); 182 writel(1, scp->cluster->reg_base + MT8192_CORE0_SW_RSTN_CLR); 187 writel(1, scp->cluster->reg_base + MT8195_CORE1_SW_RSTN_SET); 192 writel(1, scp->cluster->reg_base + MT8195_CORE1_SW_RSTN_CLR); 199 scp_to_host = readl(scp->cluster [all...] |
H A D | xlnx_r5_remoteproc.c | 29 * settings for RPU cluster mode which 30 * reflects possible values of xlnx,cluster-mode dt-property 87 /* In lockstep mode cluster combines each 64KB TCM and makes 128KB TCM */ 119 * @dev: r5f subsystem cluster device node 120 * @mode: cluster mode of type zynqmp_r5_cluster_mode 121 * @core_count: number of r5 cores used for this cluster mode 306 * set RPU cluster and TCM operation mode 309 * @fw_reg_val: value expected by firmware to configure RPU cluster mode 699 * allocate and add remoteproc carveouts for TCM memory based on cluster mode 705 struct zynqmp_r5_cluster *cluster; local 891 zynqmp_r5_get_tcm_node(struct zynqmp_r5_cluster *cluster) argument 951 zynqmp_r5_core_init(struct zynqmp_r5_cluster *cluster, enum rpu_oper_mode fw_reg_val, enum rpu_tcm_comb tcm_mode) argument 995 zynqmp_r5_cluster_init(struct zynqmp_r5_cluster *cluster) argument 1149 struct zynqmp_r5_cluster *cluster; local 1182 struct zynqmp_r5_cluster *cluster; local [all...] |
/linux-master/fs/dlm/ |
H A D | lockspace.h | 30 int dlm_new_user_lockspace(const char *name, const char *cluster,
|
/linux-master/drivers/cpufreq/ |
H A D | tegra186-cpufreq.c | 75 unsigned int cluster = data->cpus[policy->cpu].bpmp_cluster_id; local 77 policy->freq_table = data->clusters[cluster].table; 100 struct tegra186_cpufreq_cluster *cluster; local 112 cluster = &data->clusters[cluster_id]; 115 return (cluster->ref_clk_khz * ndiv) / cluster->div; 131 struct tegra186_cpufreq_cluster *cluster, unsigned int cluster_id) 187 cluster->ref_clk_khz = data->ref_clk_hz / 1000; 188 cluster->div = data->pdiv * data->mdiv; 207 point->frequency = (cluster 129 init_vhint_table( struct platform_device *pdev, struct tegra_bpmp *bpmp, struct tegra186_cpufreq_cluster *cluster, unsigned int cluster_id) argument 243 struct tegra186_cpufreq_cluster *cluster = &data->clusters[i]; local [all...] |
H A D | vexpress-spc-cpufreq.c | 45 #define ACTUAL_FREQ(cluster, freq) ((cluster == A7_CLUSTER) ? freq << 1 : freq) 46 #define VIRT_FREQ(cluster, freq) ((cluster == A7_CLUSTER) ? freq >> 1 : freq) 71 static unsigned int find_cluster_maxfreq(int cluster) argument 79 if (cluster == per_cpu(physical_cluster, j) && 154 /* Recalc freq for old cluster when switching clusters */ 156 /* Switch cluster */ 161 /* Set freq of old cluster if there are cpus left on it */ 167 pr_err("%s: clk_set_rate failed: %d, old cluster 282 u32 cluster = raw_cpu_to_cluster(cpu_dev->id); local 294 u32 cluster = cpu_to_cluster(cpu_dev->id); local 319 u32 cluster = raw_cpu_to_cluster(cpu_dev->id); local 355 u32 cluster = cpu_to_cluster(cpu_dev->id); local [all...] |
/linux-master/arch/mips/include/asm/ |
H A D | mips-cps.h | 123 * mips_cps_cluster_config - return (GCR|CPC)_CONFIG from a cluster 124 * @cluster: the ID of the cluster whose config we want 126 * Read the value of GCR_CONFIG (or its CPC_CONFIG mirror) from a @cluster. 130 static inline uint64_t mips_cps_cluster_config(unsigned int cluster) argument 138 * within this cluster. 140 WARN_ON(cluster != 0); 148 mips_cm_lock_other(cluster, 0, 0, CM_GCR_Cx_OTHER_BLOCK_GLOBAL); 157 * mips_cps_numcores - return the number of cores present in a cluster 158 * @cluster 163 mips_cps_numcores(unsigned int cluster) argument 180 mips_cps_numiocu(unsigned int cluster) argument 198 mips_cps_numvps(unsigned int cluster, unsigned int core) argument [all...] |
/linux-master/arch/arm/mach-milbeaut/ |
H A D | platsmp.c | 25 unsigned int mpidr, cpu, cluster; local 32 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); 37 pr_info("%s: cpu %u l_cpu %u cluster %u\n", 38 __func__, cpu, l_cpu, cluster); 48 unsigned int mpidr, cpu, cluster; local 61 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); 62 pr_info("MCPM boot on cpu_%u cluster_%u\n", cpu, cluster);
|
/linux-master/fs/fat/ |
H A D | cache.c | 7 * Mar 1999. AV. Changed cache, so that it uses the starting cluster instead 21 int fcluster; /* cluster number in the file. */ 22 int dcluster; /* cluster number on disk. */ 123 /* Find the same part as "new" in cluster-chain. */ 225 int fat_get_cluster(struct inode *inode, int cluster, int *fclus, int *dclus) argument 240 "%s: invalid start cluster (i_pos %lld, start %08x)", 244 if (cluster == 0) 247 if (fat_cache_lookup(inode, cluster, &cid, fclus, dclus) < 0) { 256 while (*fclus < cluster) { 257 /* prevent the infinite loop of cluster chai 291 fat_bmap_cluster(struct inode *inode, int cluster) argument 316 int cluster, offset; local [all...] |
/linux-master/arch/x86/kernel/apic/ |
H A D | x2apic_cluster.c | 56 /* Collapse cpus in a cluster so a single IPI per cluster is sent */ 68 /* Remove cluster CPUs from tmpmask */ 105 static void prefill_clustermask(struct cpumask *cmsk, unsigned int cpu, u32 cluster) argument 113 if (apicid == BAD_APICID || cpu_i == cpu || apic_cluster(apicid) != cluster) 124 static int alloc_clustermask(unsigned int cpu, u32 cluster, int node) argument 130 * At boot time, the CPU present mask is stable. The cluster mask is 131 * allocated for the first CPU in the cluster and propagated to all 132 * present siblings in the cluster. If the cluster mas 179 u32 cluster = apic_cluster(phys_apicid); local [all...] |