// SPDX-License-Identifier: MIT /* * Copyright © 2019 Intel Corporation */ #include #include #include "i915_drv.h" #include "intel_engine.h" #include "intel_engine_heartbeat.h" #include "sysfs_engines.h" struct kobj_engine { struct kobject base; struct intel_engine_cs *engine; }; static struct intel_engine_cs *kobj_to_engine(struct kobject *kobj) { return container_of(kobj, struct kobj_engine, base)->engine; } static ssize_t name_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return sysfs_emit(buf, "%s\n", kobj_to_engine(kobj)->name); } static const struct kobj_attribute name_attr = __ATTR(name, 0444, name_show, NULL); static ssize_t class_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return sysfs_emit(buf, "%d\n", kobj_to_engine(kobj)->uabi_class); } static const struct kobj_attribute class_attr = __ATTR(class, 0444, class_show, NULL); static ssize_t inst_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return sysfs_emit(buf, "%d\n", kobj_to_engine(kobj)->uabi_instance); } static const struct kobj_attribute inst_attr = __ATTR(instance, 0444, inst_show, NULL); static ssize_t mmio_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return sysfs_emit(buf, "0x%x\n", kobj_to_engine(kobj)->mmio_base); } static const struct kobj_attribute mmio_attr = __ATTR(mmio_base, 0444, mmio_show, NULL); static const char * const vcs_caps[] = { [ilog2(I915_VIDEO_CLASS_CAPABILITY_HEVC)] = "hevc", [ilog2(I915_VIDEO_AND_ENHANCE_CLASS_CAPABILITY_SFC)] = "sfc", }; static const char * const vecs_caps[] = { [ilog2(I915_VIDEO_AND_ENHANCE_CLASS_CAPABILITY_SFC)] = "sfc", }; static ssize_t repr_trim(char *buf, ssize_t len) { /* Trim off the trailing space and replace with a newline */ if (len > PAGE_SIZE) len = PAGE_SIZE; if (len > 0) buf[len - 1] = '\n'; return len; } static ssize_t __caps_show(struct intel_engine_cs *engine, unsigned long caps, char *buf, bool show_unknown) { const char * const *repr; int count, n; ssize_t len; switch (engine->class) { case VIDEO_DECODE_CLASS: repr = vcs_caps; count = ARRAY_SIZE(vcs_caps); break; case VIDEO_ENHANCEMENT_CLASS: repr = vecs_caps; count = ARRAY_SIZE(vecs_caps); break; default: repr = NULL; count = 0; break; } GEM_BUG_ON(count > BITS_PER_LONG); len = 0; for_each_set_bit(n, &caps, show_unknown ? BITS_PER_LONG : count) { if (n >= count || !repr[n]) { if (GEM_WARN_ON(show_unknown)) len += sysfs_emit_at(buf, len, "[%x] ", n); } else { len += sysfs_emit_at(buf, len, "%s ", repr[n]); } if (GEM_WARN_ON(len >= PAGE_SIZE)) break; } return repr_trim(buf, len); } static ssize_t caps_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return __caps_show(engine, engine->uabi_capabilities, buf, true); } static const struct kobj_attribute caps_attr = __ATTR(capabilities, 0444, caps_show, NULL); static ssize_t all_caps_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { return __caps_show(kobj_to_engine(kobj), -1, buf, false); } static const struct kobj_attribute all_caps_attr = __ATTR(known_capabilities, 0444, all_caps_show, NULL); static ssize_t max_spin_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) { struct intel_engine_cs *engine = kobj_to_engine(kobj); unsigned long long duration, clamped; int err; /* * When waiting for a request, if is it currently being executed * on the GPU, we busywait for a short while before sleeping. The * premise is that most requests are short, and if it is already * executing then there is a good chance that it will complete * before we can setup the interrupt handler and go to sleep. * We try to offset the cost of going to sleep, by first spinning * on the request -- if it completed in less time than it would take * to go sleep, process the interrupt and return back to the client, * then we have saved the client some latency, albeit at the cost * of spinning on an expensive CPU core. * * While we try to avoid waiting at all for a request that is unlikely * to complete, deciding how long it is worth spinning is for is an * arbitrary decision: trading off power vs latency. */ err = kstrtoull(buf, 0, &duration); if (err) return err; clamped = intel_clamp_max_busywait_duration_ns(engine, duration); if (duration != clamped) return -EINVAL; WRITE_ONCE(engine->props.max_busywait_duration_ns, duration); return count; } static ssize_t max_spin_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->props.max_busywait_duration_ns); } static const struct kobj_attribute max_spin_attr = __ATTR(max_busywait_duration_ns, 0644, max_spin_show, max_spin_store); static ssize_t max_spin_default(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->defaults.max_busywait_duration_ns); } static const struct kobj_attribute max_spin_def = __ATTR(max_busywait_duration_ns, 0444, max_spin_default, NULL); static ssize_t timeslice_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) { struct intel_engine_cs *engine = kobj_to_engine(kobj); unsigned long long duration, clamped; int err; /* * Execlists uses a scheduling quantum (a timeslice) to alternate * execution between ready-to-run contexts of equal priority. This * ensures that all users (though only if they of equal importance) * have the opportunity to run and prevents livelocks where contexts * may have implicit ordering due to userspace semaphores. */ err = kstrtoull(buf, 0, &duration); if (err) return err; clamped = intel_clamp_timeslice_duration_ms(engine, duration); if (duration != clamped) return -EINVAL; WRITE_ONCE(engine->props.timeslice_duration_ms, duration); if (execlists_active(&engine->execlists)) set_timer_ms(&engine->execlists.timer, duration); return count; } static ssize_t timeslice_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->props.timeslice_duration_ms); } static const struct kobj_attribute timeslice_duration_attr = __ATTR(timeslice_duration_ms, 0644, timeslice_show, timeslice_store); static ssize_t timeslice_default(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->defaults.timeslice_duration_ms); } static const struct kobj_attribute timeslice_duration_def = __ATTR(timeslice_duration_ms, 0444, timeslice_default, NULL); static ssize_t stop_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) { struct intel_engine_cs *engine = kobj_to_engine(kobj); unsigned long long duration, clamped; int err; /* * When we allow ourselves to sleep before a GPU reset after disabling * submission, even for a few milliseconds, gives an innocent context * the opportunity to clear the GPU before the reset occurs. However, * how long to sleep depends on the typical non-preemptible duration * (a similar problem to determining the ideal preempt-reset timeout * or even the heartbeat interval). */ err = kstrtoull(buf, 0, &duration); if (err) return err; clamped = intel_clamp_stop_timeout_ms(engine, duration); if (duration != clamped) return -EINVAL; WRITE_ONCE(engine->props.stop_timeout_ms, duration); return count; } static ssize_t stop_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->props.stop_timeout_ms); } static const struct kobj_attribute stop_timeout_attr = __ATTR(stop_timeout_ms, 0644, stop_show, stop_store); static ssize_t stop_default(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->defaults.stop_timeout_ms); } static const struct kobj_attribute stop_timeout_def = __ATTR(stop_timeout_ms, 0444, stop_default, NULL); static ssize_t preempt_timeout_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) { struct intel_engine_cs *engine = kobj_to_engine(kobj); unsigned long long timeout, clamped; int err; /* * After initialising a preemption request, we give the current * resident a small amount of time to vacate the GPU. The preemption * request is for a higher priority context and should be immediate to * maintain high quality of service (and avoid priority inversion). * However, the preemption granularity of the GPU can be quite coarse * and so we need a compromise. */ err = kstrtoull(buf, 0, &timeout); if (err) return err; clamped = intel_clamp_preempt_timeout_ms(engine, timeout); if (timeout != clamped) return -EINVAL; WRITE_ONCE(engine->props.preempt_timeout_ms, timeout); if (READ_ONCE(engine->execlists.pending[0])) set_timer_ms(&engine->execlists.preempt, timeout); return count; } static ssize_t preempt_timeout_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->props.preempt_timeout_ms); } static const struct kobj_attribute preempt_timeout_attr = __ATTR(preempt_timeout_ms, 0644, preempt_timeout_show, preempt_timeout_store); static ssize_t preempt_timeout_default(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->defaults.preempt_timeout_ms); } static const struct kobj_attribute preempt_timeout_def = __ATTR(preempt_timeout_ms, 0444, preempt_timeout_default, NULL); static ssize_t heartbeat_store(struct kobject *kobj, struct kobj_attribute *attr, const char *buf, size_t count) { struct intel_engine_cs *engine = kobj_to_engine(kobj); unsigned long long delay, clamped; int err; /* * We monitor the health of the system via periodic heartbeat pulses. * The pulses also provide the opportunity to perform garbage * collection. However, we interpret an incomplete pulse (a missed * heartbeat) as an indication that the system is no longer responsive, * i.e. hung, and perform an engine or full GPU reset. Given that the * preemption granularity can be very coarse on a system, the optimal * value for any workload is unknowable! */ err = kstrtoull(buf, 0, &delay); if (err) return err; clamped = intel_clamp_heartbeat_interval_ms(engine, delay); if (delay != clamped) return -EINVAL; err = intel_engine_set_heartbeat(engine, delay); if (err) return err; return count; } static ssize_t heartbeat_show(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->props.heartbeat_interval_ms); } static const struct kobj_attribute heartbeat_interval_attr = __ATTR(heartbeat_interval_ms, 0644, heartbeat_show, heartbeat_store); static ssize_t heartbeat_default(struct kobject *kobj, struct kobj_attribute *attr, char *buf) { struct intel_engine_cs *engine = kobj_to_engine(kobj); return sysfs_emit(buf, "%lu\n", engine->defaults.heartbeat_interval_ms); } static const struct kobj_attribute heartbeat_interval_def = __ATTR(heartbeat_interval_ms, 0444, heartbeat_default, NULL); static void kobj_engine_release(struct kobject *kobj) { kfree(kobj); } static const struct kobj_type kobj_engine_type = { .release = kobj_engine_release, .sysfs_ops = &kobj_sysfs_ops }; static struct kobject * kobj_engine(struct kobject *dir, struct intel_engine_cs *engine) { struct kobj_engine *ke; ke = kzalloc(sizeof(*ke), GFP_KERNEL); if (!ke) return NULL; kobject_init(&ke->base, &kobj_engine_type); ke->engine = engine; if (kobject_add(&ke->base, dir, "%s", engine->name)) { kobject_put(&ke->base); return NULL; } /* xfer ownership to sysfs tree */ return &ke->base; } static void add_defaults(struct kobj_engine *parent) { static const struct attribute * const files[] = { &max_spin_def.attr, &stop_timeout_def.attr, #if CONFIG_DRM_I915_HEARTBEAT_INTERVAL &heartbeat_interval_def.attr, #endif NULL }; struct kobj_engine *ke; ke = kzalloc(sizeof(*ke), GFP_KERNEL); if (!ke) return; kobject_init(&ke->base, &kobj_engine_type); ke->engine = parent->engine; if (kobject_add(&ke->base, &parent->base, "%s", ".defaults")) { kobject_put(&ke->base); return; } if (sysfs_create_files(&ke->base, files)) return; if (intel_engine_has_timeslices(ke->engine) && sysfs_create_file(&ke->base, ×lice_duration_def.attr)) return; if (intel_engine_has_preempt_reset(ke->engine) && sysfs_create_file(&ke->base, &preempt_timeout_def.attr)) return; } void intel_engines_add_sysfs(struct drm_i915_private *i915) { static const struct attribute * const files[] = { &name_attr.attr, &class_attr.attr, &inst_attr.attr, &mmio_attr.attr, &caps_attr.attr, &all_caps_attr.attr, &max_spin_attr.attr, &stop_timeout_attr.attr, #if CONFIG_DRM_I915_HEARTBEAT_INTERVAL &heartbeat_interval_attr.attr, #endif NULL }; struct device *kdev = i915->drm.primary->kdev; struct intel_engine_cs *engine; struct kobject *dir; dir = kobject_create_and_add("engine", &kdev->kobj); if (!dir) return; for_each_uabi_engine(engine, i915) { struct kobject *kobj; kobj = kobj_engine(dir, engine); if (!kobj) goto err_engine; if (sysfs_create_files(kobj, files)) goto err_object; if (intel_engine_has_timeslices(engine) && sysfs_create_file(kobj, ×lice_duration_attr.attr)) goto err_engine; if (intel_engine_has_preempt_reset(engine) && sysfs_create_file(kobj, &preempt_timeout_attr.attr)) goto err_engine; add_defaults(container_of(kobj, struct kobj_engine, base)); if (0) { err_object: kobject_put(kobj); err_engine: dev_err(kdev, "Failed to add sysfs engine '%s'\n", engine->name); break; } } }