Lines Matching defs:die

1211 		int die = box->dieid;
1212 struct pci_dev *filter_pdev = uncore_extra_pci_dev[die].dev[idx];
3720 static struct intel_uncore_topology *pmu_topology(struct intel_uncore_pmu *pmu, int die)
3725 if (pmu->type->topology[die][idx].pmu_idx == pmu->pmu_idx)
3726 return &pmu->type->topology[die][idx];
3734 int die, int zero_bus_pmu)
3737 struct intel_uncore_topology *pmut = pmu_topology(pmu, die);
3743 skx_iio_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
3746 return pmu_iio_mapping_visible(kobj, attr, die, 0);
3754 long die = (long)ea->var;
3755 struct intel_uncore_topology *pmut = pmu_topology(pmu, die);
3774 static int die_to_cpu(int die)
3784 if (current_die == die) {
3806 int die, idx;
3816 for (die = 0; die < uncore_max_dies(); die++) {
3817 topology[die] = kcalloc(type->num_boxes, sizeof(**topology), GFP_KERNEL);
3818 if (!topology[die])
3821 topology[die][idx].untyped = kcalloc(type->num_boxes,
3824 if (!topology[die][idx].untyped)
3833 for (; die >= 0; die--) {
3835 kfree(topology[die][idx].untyped);
3836 kfree(topology[die]);
3845 int die, idx;
3848 for (die = 0; die < uncore_max_dies(); die++) {
3850 kfree(type->topology[die][idx].untyped);
3851 kfree(type->topology[die]);
3861 int die, ret = -EPERM;
3864 for (die = 0; die < uncore_max_dies(); die++) {
3865 ret = skx_msr_cpu_bus_read(die_to_cpu(die), &cpu_bus_msr);
3869 ret = uncore_die_to_segment(die);
3873 ret = topology_cb(type, ret, die, cpu_bus_msr);
3882 int die, u64 cpu_bus_msr)
3888 t = &type->topology[die][idx];
3933 long die = -1;
3954 for (die = 0; die < uncore_max_dies(); die++) {
3955 snprintf(buf, sizeof(buf), "die%ld", die);
3956 sysfs_attr_init(&eas[die].attr.attr);
3957 eas[die].attr.attr.name = kstrdup(buf, GFP_KERNEL);
3958 if (!eas[die].attr.attr.name)
3960 eas[die].attr.attr.mode = 0444;
3961 eas[die].attr.show = show;
3962 eas[die].attr.store = NULL;
3963 eas[die].var = (void *)die;
3964 attrs[die] = &eas[die].attr.attr;
3970 for (; die >= 0; die--)
3971 kfree(eas[die].attr.attr.name);
4272 skx_upi_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
4276 return pmu->type->topology[die][pmu->pmu_idx].upi->enabled ? attr->mode : 0;
4284 long die = (long)ea->var;
4285 struct uncore_upi_topology *upi = pmu->type->topology[die][pmu->pmu_idx].upi;
4298 * | 12:8 | 00h | sending_port - The processor die port number of the sending port.
4339 int die, u64 cpu_bus_msr)
4348 upi = &type->topology[die][idx];
4702 snr_iio_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
4705 return pmu_iio_mapping_visible(kobj, attr, die, 1);
4720 int die, stack_id, ret = -EPERM;
4730 die = uncore_pcibus_to_dieid(dev->bus);
4732 if (die < 0 || stack_id >= type->num_boxes) {
4740 type->topology[die][stack_id].iio->segment = pci_domain_nr(dev->bus);
4741 type->topology[die][stack_id].pmu_idx = stack_id;
4742 type->topology[die][stack_id].iio->pci_bus_no = dev->bus->number;
5353 icx_iio_mapping_visible(struct kobject *kobj, struct attribute *attr, int die)
5356 return pmu_iio_mapping_visible(kobj, attr, die, 5);
6519 int die;
6542 die = uncore_device_to_die(dev);
6543 if (die < 0)
6546 ctls[die] = pci_domain_nr(dev->bus) << UNCORE_DISCOVERY_PCI_DOMAIN_OFFSET |