Searched refs:cluster (Results 1 - 25 of 131) sorted by relevance

123456

/linux-master/arch/arm/common/
H A Dmcpm_entry.c3 * arch/arm/common/mcpm_entry.c -- entry point for multi-cluster PM
24 * see Documentation/arch/arm/cluster-pm-race-avoidance.rst.
34 static void __mcpm_cpu_going_down(unsigned int cpu, unsigned int cluster) argument
36 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_GOING_DOWN;
37 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu);
42 * cluster can be torn down without disrupting this CPU.
47 static void __mcpm_cpu_down(unsigned int cpu, unsigned int cluster) argument
50 mcpm_sync.clusters[cluster].cpus[cpu].cpu = CPU_DOWN;
51 sync_cache_w(&mcpm_sync.clusters[cluster].cpus[cpu].cpu);
56 * __mcpm_outbound_leave_critical: Leave the cluster teardow
63 __mcpm_outbound_leave_critical(unsigned int cluster, int state) argument
82 __mcpm_outbound_enter_critical(unsigned int cpu, unsigned int cluster) argument
135 __mcpm_cluster_state(unsigned int cluster) argument
143 mcpm_set_entry_vector(unsigned cpu, unsigned cluster, void *ptr) argument
152 mcpm_set_early_poke(unsigned cpu, unsigned cluster, unsigned long poke_phys_addr, unsigned long poke_val) argument
187 mcpm_cluster_unused(unsigned int cluster) argument
195 mcpm_cpu_power_up(unsigned int cpu, unsigned int cluster) argument
241 unsigned int mpidr, cpu, cluster; local
308 mcpm_wait_for_cpu_powerdown(unsigned int cpu, unsigned int cluster) argument
332 unsigned int cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); local
342 unsigned int mpidr, cpu, cluster; local
378 unsigned int cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1); local
[all...]
H A DbL_switcher_dummy_if.c22 unsigned int cpu, cluster; local
33 /* format: <cpu#>,<cluster#> */
40 cluster = val[2] - '0';
41 ret = bL_switch_request(cpu, cluster);
H A Dmcpm_head.S3 * arch/arm/common/mcpm_head.S -- kernel entry point for multi-cluster PM
8 * Refer to Documentation/arch/arm/cluster-pm-race-avoidance.rst
28 1903: .asciz " cluster"
56 ubfx r10, r0, #8, #8 @ r10 = cluster
88 mla r8, r0, r10, r8 @ r8 = sync cluster base
96 @ At this point, the cluster cannot unexpectedly enter the GOING_DOWN
100 mla r11, r0, r10, r11 @ r11 = cluster first man lock
106 bne mcpm_setup_wait @ wait for cluster setup if so
109 cmp r0, #CLUSTER_UP @ cluster already up?
110 bne mcpm_setup @ if not, set up the cluster
[all...]
H A DbL_switcher.c3 * arch/arm/common/bL_switcher.c -- big.LITTLE cluster switcher core driver
118 * with the cluster number.
141 * bL_switch_to - Switch to a specific cluster for the current CPU
142 * @new_cluster_id: the ID of the cluster to switch to.
273 int cluster; local
288 cluster = t->wanted_cluster;
295 if (cluster != -1) {
296 bL_switch_to(cluster);
321 * bL_switch_request_cb - Switch to a specific cluster for the given CPU,
325 * @new_cluster_id: the ID of the cluster t
422 unsigned int cpu, cluster, mask; local
604 unsigned int cpu, cluster; local
[all...]
/linux-master/arch/arm/include/asm/
H A Dmcpm.h13 * Maximum number of possible clusters / CPUs per cluster.
39 * This is used to indicate where the given CPU from given cluster should
44 void mcpm_set_entry_vector(unsigned cpu, unsigned cluster, void *ptr);
51 void mcpm_set_early_poke(unsigned cpu, unsigned cluster,
55 * CPU/cluster power operations API for higher subsystems to use.
66 * mcpm_cpu_power_up - make given CPU in given cluster runable
68 * @cpu: CPU number within given cluster
69 * @cluster: cluster number for the CPU
71 * The identified CPU is brought out of reset. If the cluster wa
[all...]
/linux-master/arch/arm/mach-versatile/
H A Ddcscb.c39 static int dcscb_cpu_powerup(unsigned int cpu, unsigned int cluster) argument
43 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster);
44 if (cluster >= 2 || !(cpumask & dcscb_allcpus_mask[cluster]))
47 rst_hold = readl_relaxed(dcscb_base + RST_HOLD0 + cluster * 4);
49 writel_relaxed(rst_hold, dcscb_base + RST_HOLD0 + cluster * 4);
53 static int dcscb_cluster_powerup(unsigned int cluster) argument
57 pr_debug("%s: cluster %u\n", __func__, cluster);
69 dcscb_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) argument
81 dcscb_cluster_powerdown_prepare(unsigned int cluster) argument
[all...]
H A Dspc.h13 void ve_spc_cpu_wakeup_irq(u32 cluster, u32 cpu, bool set);
14 void ve_spc_set_resume_addr(u32 cluster, u32 cpu, u32 addr);
15 void ve_spc_powerdown(u32 cluster, bool enable);
16 int ve_spc_cpu_in_wfi(u32 cpu, u32 cluster);
H A Dtc2_pm.c46 static int tc2_pm_cpu_powerup(unsigned int cpu, unsigned int cluster) argument
48 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster);
49 if (cluster >= TC2_CLUSTERS || cpu >= tc2_nr_cpus[cluster])
51 ve_spc_set_resume_addr(cluster, cpu,
53 ve_spc_cpu_wakeup_irq(cluster, cpu, true);
57 static int tc2_pm_cluster_powerup(unsigned int cluster) argument
59 pr_debug("%s: cluster %u\n", __func__, cluster);
66 tc2_pm_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) argument
81 tc2_pm_cluster_powerdown_prepare(unsigned int cluster) argument
112 tc2_core_in_reset(unsigned int cpu, unsigned int cluster) argument
124 tc2_pm_wait_for_powerdown(unsigned int cpu, unsigned int cluster) argument
155 tc2_pm_cpu_suspend_prepare(unsigned int cpu, unsigned int cluster) argument
160 tc2_pm_cpu_is_up(unsigned int cpu, unsigned int cluster) argument
168 tc2_pm_cluster_is_up(unsigned int cluster) argument
202 unsigned int mpidr, cpu, cluster; local
[all...]
H A Dspc.c50 /* SPC CPU/cluster reset statue */
71 /* TC2 static dual-cluster configuration */
97 * A15s cluster identifier
111 static inline bool cluster_is_a15(u32 cluster) argument
113 return cluster == info->a15_clusid;
142 * @cluster: mpidr[15:8] bitfield describing cluster affinity level
150 void ve_spc_cpu_wakeup_irq(u32 cluster, u32 cpu, bool set) argument
154 if (cluster >= MAX_CLUSTERS)
159 if (!cluster_is_a15(cluster))
179 ve_spc_set_resume_addr(u32 cluster, u32 cpu, u32 addr) argument
204 ve_spc_powerdown(u32 cluster, bool enable) argument
215 standbywfi_cpu_mask(u32 cpu, u32 cluster) argument
234 ve_spc_cpu_in_wfi(u32 cpu, u32 cluster) argument
250 ve_spc_get_performance(int cluster, u32 *freq) argument
269 ve_spc_round_performance(int cluster, u32 freq) argument
292 ve_spc_find_performance_index(int cluster, u32 freq) argument
314 ve_spc_set_performance(int cluster, u32 freq) argument
393 ve_spc_populate_opps(uint32_t cluster) argument
421 int cluster; local
484 int cluster; member in struct:clk_spc
546 int cpu, cluster; local
[all...]
/linux-master/arch/arm/mach-sunxi/
H A Dmc_smp.c87 static bool sunxi_core_is_cortex_a15(unsigned int core, unsigned int cluster) argument
90 int cpu = cluster * SUNXI_CPUS_PER_CLUSTER + core;
102 * would be mid way in a core or cluster power sequence.
104 pr_err("%s: Couldn't get CPU cluster %u core %u device node\n",
105 __func__, cluster, core);
115 static int sunxi_cpu_power_switch_set(unsigned int cpu, unsigned int cluster, argument
121 reg = readl(prcm_base + PRCM_PWR_SWITCH_REG(cluster, cpu));
124 pr_debug("power clamp for cluster %u cpu %u already open\n",
125 cluster, cpu);
129 writel(0xff, prcm_base + PRCM_PWR_SWITCH_REG(cluster, cp
158 sunxi_cpu_powerup(unsigned int cpu, unsigned int cluster) argument
253 sunxi_cluster_powerup(unsigned int cluster) argument
374 sunxi_mc_smp_cluster_is_down(unsigned int cluster) argument
393 unsigned int mpidr, cpu, cluster; local
430 unsigned int cluster = MPIDR_AFFINITY_LEVEL(read_cpuid_mpidr(), 1); local
445 unsigned int mpidr, cpu, cluster; local
480 sunxi_cpu_powerdown(unsigned int cpu, unsigned int cluster) argument
504 sunxi_cluster_powerdown(unsigned int cluster) argument
535 unsigned int mpidr, cpu, cluster; local
630 unsigned int mpidr, cpu, cluster; local
[all...]
/linux-master/arch/arm/mach-exynos/
H A Dmcpm-exynos.c57 static int exynos_cpu_powerup(unsigned int cpu, unsigned int cluster) argument
59 unsigned int cpunr = cpu + (cluster * EXYNOS5420_CPUS_PER_CLUSTER);
62 pr_debug("%s: cpu %u cluster %u\n", __func__, cpu, cluster);
64 cluster >= EXYNOS5420_NR_CLUSTERS)
71 * This assumes the cluster number of the big cores(Cortex A15)
76 if (cluster &&
77 cluster == MPIDR_AFFINITY_LEVEL(cpu_logical_map(0), 1)) {
92 pr_err("cpu %u cluster %u powerup failed\n",
93 cpu, cluster);
106 exynos_cluster_powerup(unsigned int cluster) argument
116 exynos_cpu_powerdown_prepare(unsigned int cpu, unsigned int cluster) argument
126 exynos_cluster_powerdown_prepare(unsigned int cluster) argument
163 exynos_wait_for_powerdown(unsigned int cpu, unsigned int cluster) argument
184 exynos_cpu_is_up(unsigned int cpu, unsigned int cluster) argument
[all...]
H A Dcommon.h140 extern void exynos_cluster_power_down(int cluster);
141 extern void exynos_cluster_power_up(int cluster);
142 extern int exynos_cluster_power_state(int cluster);
/linux-master/arch/arm/mach-hisi/
H A Dplatmcpm.c71 static bool hip04_cluster_is_down(unsigned int cluster) argument
76 if (hip04_cpu_table[cluster][i])
81 static void hip04_set_snoop_filter(unsigned int cluster, unsigned int on) argument
89 data |= 1 << cluster;
91 data &= ~(1 << cluster);
100 unsigned int mpidr, cpu, cluster; local
106 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1);
110 if (cluster >= HIP04_MAX_CLUSTERS || cpu >= HIP04_MAX_CPUS_PER_CLUSTER)
115 if (hip04_cpu_table[cluster][cpu])
118 sys_dreq = sysctrl + SC_CPU_RESET_DREQ(cluster);
155 unsigned int mpidr, cpu, cluster; local
193 unsigned int mpidr, cpu, cluster; local
249 unsigned int mpidr, cpu, cluster; local
[all...]
/linux-master/fs/ocfs2/cluster/
H A Dnodemanager.c18 * cluster active at a time. Changing this will require trickling
19 * cluster references throughout where nodes are looked up */
49 struct o2nm_cluster *cluster = o2nm_single_cluster; local
51 BUG_ON(bytes < (sizeof(cluster->cl_nodes_bitmap)));
53 if (cluster == NULL)
56 read_lock(&cluster->cl_nodes_lock);
57 bitmap_copy(map, cluster->cl_nodes_bitmap, O2NM_MAX_NODES);
58 read_unlock(&cluster->cl_nodes_lock);
64 static struct o2nm_node *o2nm_node_ip_tree_lookup(struct o2nm_cluster *cluster, argument
69 struct rb_node **p = &cluster
102 struct o2nm_cluster *cluster = o2nm_single_cluster; local
187 struct o2nm_cluster *cluster; local
271 struct o2nm_cluster *cluster; local
326 struct o2nm_cluster *cluster; local
453 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local
490 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local
534 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local
606 struct o2nm_cluster *cluster = to_o2nm_cluster(group->cg_item.ci_parent); local
654 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local
688 struct o2nm_cluster *cluster = NULL; local
734 struct o2nm_cluster *cluster = to_o2nm_cluster(item); local
[all...]
/linux-master/drivers/perf/
H A Dqcom_l2_pmu.c121 * The cache is made up of one or more clusters, each cluster has its own PMU.
122 * Each cluster is associated with one or more CPUs.
143 /* The CPU that is used for collecting events on this cluster */
145 /* All the CPUs associated with this cluster */
243 static void cluster_pmu_set_resr(struct cluster_pmu *cluster, argument
254 spin_lock_irqsave(&cluster->pmu_lock, flags);
262 spin_unlock_irqrestore(&cluster->pmu_lock, flags);
268 * all CPUS, subunits and ID independent events in this cluster.
319 static void l2_cache_cluster_set_period(struct cluster_pmu *cluster, argument
339 static int l2_cache_get_event_idx(struct cluster_pmu *cluster, argument
374 l2_cache_clear_event_idx(struct cluster_pmu *cluster, struct perf_event *event) argument
387 struct cluster_pmu *cluster = data; local
441 struct cluster_pmu *cluster; local
543 struct cluster_pmu *cluster; local
593 struct cluster_pmu *cluster; local
618 struct cluster_pmu *cluster; local
739 struct cluster_pmu *cluster; local
769 struct cluster_pmu *cluster; local
804 struct cluster_pmu *cluster; local
844 struct cluster_pmu *cluster; local
[all...]
/linux-master/drivers/remoteproc/
H A Dti_k3_r5_remoteproc.c73 * All cluster mode values are not applicable on all SoCs. The following
105 * @cores: list of R5 cores within the cluster
152 * @cluster: cached pointer to parent cluster structure
162 struct k3_r5_cluster *cluster; member in struct:k3_r5_rproc
282 static int k3_r5_lockstep_reset(struct k3_r5_cluster *cluster) argument
288 list_for_each_entry(core, &cluster->cores, elem) {
299 list_for_each_entry(core, &cluster->cores, elem) {
312 list_for_each_entry_continue_reverse(core, &cluster->cores, elem) {
317 core = list_last_entry(&cluster
327 k3_r5_lockstep_release(struct k3_r5_cluster *cluster) argument
441 struct k3_r5_cluster *cluster = kproc->cluster; local
508 struct k3_r5_cluster *cluster = kproc->cluster; local
543 struct k3_r5_cluster *cluster = kproc->cluster; local
615 struct k3_r5_cluster *cluster = kproc->cluster; local
834 struct k3_r5_cluster *cluster = kproc->cluster; local
1072 struct k3_r5_cluster *cluster = kproc->cluster; local
1111 struct k3_r5_cluster *cluster = kproc->cluster; local
1209 struct k3_r5_cluster *cluster = platform_get_drvdata(pdev); local
1316 struct k3_r5_cluster *cluster = platform_get_drvdata(data); local
1639 struct k3_r5_cluster *cluster = platform_get_drvdata(data); local
1652 struct k3_r5_cluster *cluster = platform_get_drvdata(pdev); local
1694 struct k3_r5_cluster *cluster; local
[all...]
H A Dmtk_scp.c71 struct mtk_scp_of_cluster *scp_cluster = scp->cluster;
161 val = readl(scp->cluster->reg_base + MT8183_SW_RSTN);
163 writel(val, scp->cluster->reg_base + MT8183_SW_RSTN);
170 val = readl(scp->cluster->reg_base + MT8183_SW_RSTN);
172 writel(val, scp->cluster->reg_base + MT8183_SW_RSTN);
177 writel(1, scp->cluster->reg_base + MT8192_CORE0_SW_RSTN_SET);
182 writel(1, scp->cluster->reg_base + MT8192_CORE0_SW_RSTN_CLR);
187 writel(1, scp->cluster->reg_base + MT8195_CORE1_SW_RSTN_SET);
192 writel(1, scp->cluster->reg_base + MT8195_CORE1_SW_RSTN_CLR);
199 scp_to_host = readl(scp->cluster
[all...]
H A Dxlnx_r5_remoteproc.c29 * settings for RPU cluster mode which
30 * reflects possible values of xlnx,cluster-mode dt-property
87 /* In lockstep mode cluster combines each 64KB TCM and makes 128KB TCM */
119 * @dev: r5f subsystem cluster device node
120 * @mode: cluster mode of type zynqmp_r5_cluster_mode
121 * @core_count: number of r5 cores used for this cluster mode
306 * set RPU cluster and TCM operation mode
309 * @fw_reg_val: value expected by firmware to configure RPU cluster mode
699 * allocate and add remoteproc carveouts for TCM memory based on cluster mode
705 struct zynqmp_r5_cluster *cluster; local
891 zynqmp_r5_get_tcm_node(struct zynqmp_r5_cluster *cluster) argument
951 zynqmp_r5_core_init(struct zynqmp_r5_cluster *cluster, enum rpu_oper_mode fw_reg_val, enum rpu_tcm_comb tcm_mode) argument
995 zynqmp_r5_cluster_init(struct zynqmp_r5_cluster *cluster) argument
1149 struct zynqmp_r5_cluster *cluster; local
1182 struct zynqmp_r5_cluster *cluster; local
[all...]
/linux-master/fs/dlm/
H A Dlockspace.h30 int dlm_new_user_lockspace(const char *name, const char *cluster,
/linux-master/drivers/cpufreq/
H A Dtegra186-cpufreq.c75 unsigned int cluster = data->cpus[policy->cpu].bpmp_cluster_id; local
77 policy->freq_table = data->clusters[cluster].table;
100 struct tegra186_cpufreq_cluster *cluster; local
112 cluster = &data->clusters[cluster_id];
115 return (cluster->ref_clk_khz * ndiv) / cluster->div;
131 struct tegra186_cpufreq_cluster *cluster, unsigned int cluster_id)
187 cluster->ref_clk_khz = data->ref_clk_hz / 1000;
188 cluster->div = data->pdiv * data->mdiv;
207 point->frequency = (cluster
129 init_vhint_table( struct platform_device *pdev, struct tegra_bpmp *bpmp, struct tegra186_cpufreq_cluster *cluster, unsigned int cluster_id) argument
243 struct tegra186_cpufreq_cluster *cluster = &data->clusters[i]; local
[all...]
H A Dvexpress-spc-cpufreq.c45 #define ACTUAL_FREQ(cluster, freq) ((cluster == A7_CLUSTER) ? freq << 1 : freq)
46 #define VIRT_FREQ(cluster, freq) ((cluster == A7_CLUSTER) ? freq >> 1 : freq)
71 static unsigned int find_cluster_maxfreq(int cluster) argument
79 if (cluster == per_cpu(physical_cluster, j) &&
154 /* Recalc freq for old cluster when switching clusters */
156 /* Switch cluster */
161 /* Set freq of old cluster if there are cpus left on it */
167 pr_err("%s: clk_set_rate failed: %d, old cluster
282 u32 cluster = raw_cpu_to_cluster(cpu_dev->id); local
294 u32 cluster = cpu_to_cluster(cpu_dev->id); local
319 u32 cluster = raw_cpu_to_cluster(cpu_dev->id); local
355 u32 cluster = cpu_to_cluster(cpu_dev->id); local
[all...]
/linux-master/arch/mips/include/asm/
H A Dmips-cps.h123 * mips_cps_cluster_config - return (GCR|CPC)_CONFIG from a cluster
124 * @cluster: the ID of the cluster whose config we want
126 * Read the value of GCR_CONFIG (or its CPC_CONFIG mirror) from a @cluster.
130 static inline uint64_t mips_cps_cluster_config(unsigned int cluster) argument
138 * within this cluster.
140 WARN_ON(cluster != 0);
148 mips_cm_lock_other(cluster, 0, 0, CM_GCR_Cx_OTHER_BLOCK_GLOBAL);
157 * mips_cps_numcores - return the number of cores present in a cluster
158 * @cluster
163 mips_cps_numcores(unsigned int cluster) argument
180 mips_cps_numiocu(unsigned int cluster) argument
198 mips_cps_numvps(unsigned int cluster, unsigned int core) argument
[all...]
/linux-master/arch/arm/mach-milbeaut/
H A Dplatsmp.c25 unsigned int mpidr, cpu, cluster; local
32 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1);
37 pr_info("%s: cpu %u l_cpu %u cluster %u\n",
38 __func__, cpu, l_cpu, cluster);
48 unsigned int mpidr, cpu, cluster; local
61 cluster = MPIDR_AFFINITY_LEVEL(mpidr, 1);
62 pr_info("MCPM boot on cpu_%u cluster_%u\n", cpu, cluster);
/linux-master/fs/fat/
H A Dcache.c7 * Mar 1999. AV. Changed cache, so that it uses the starting cluster instead
21 int fcluster; /* cluster number in the file. */
22 int dcluster; /* cluster number on disk. */
123 /* Find the same part as "new" in cluster-chain. */
225 int fat_get_cluster(struct inode *inode, int cluster, int *fclus, int *dclus) argument
240 "%s: invalid start cluster (i_pos %lld, start %08x)",
244 if (cluster == 0)
247 if (fat_cache_lookup(inode, cluster, &cid, fclus, dclus) < 0) {
256 while (*fclus < cluster) {
257 /* prevent the infinite loop of cluster chai
291 fat_bmap_cluster(struct inode *inode, int cluster) argument
316 int cluster, offset; local
[all...]
/linux-master/arch/x86/kernel/apic/
H A Dx2apic_cluster.c56 /* Collapse cpus in a cluster so a single IPI per cluster is sent */
68 /* Remove cluster CPUs from tmpmask */
105 static void prefill_clustermask(struct cpumask *cmsk, unsigned int cpu, u32 cluster) argument
113 if (apicid == BAD_APICID || cpu_i == cpu || apic_cluster(apicid) != cluster)
124 static int alloc_clustermask(unsigned int cpu, u32 cluster, int node) argument
130 * At boot time, the CPU present mask is stable. The cluster mask is
131 * allocated for the first CPU in the cluster and propagated to all
132 * present siblings in the cluster. If the cluster mas
179 u32 cluster = apic_cluster(phys_apicid); local
[all...]

Completed in 200 milliseconds

123456