gems-kernel/source/THIRDPARTY/xnu/bsd/dev/monotonic.c
2024-06-03 11:29:39 -05:00

460 lines
11 KiB
C

/*
* Copyright (c) 2017 Apple Inc. All rights reserved.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
*
* This file contains Original Code and/or Modifications of Original Code
* as defined in and that are subject to the Apple Public Source License
* Version 2.0 (the 'License'). You may not use this file except in
* compliance with the License. The rights granted to you under the License
* may not be used to create, or enable the creation or redistribution of,
* unlawful or unlicensed copies of an Apple operating system, or to
* circumvent, violate, or enable the circumvention or violation of, any
* terms of an Apple operating system software license agreement.
*
* Please obtain a copy of the License at
* http://www.opensource.apple.com/apsl/ and read it before using this file.
*
* The Original Code and all software distributed under the License are
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
* INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
* Please see the License for the specific language governing rights and
* limitations under the License.
*
* @APPLE_OSREFERENCE_LICENSE_HEADER_END@
*/
#include <kern/monotonic.h>
#include <kern/recount.h>
#include <machine/machine_routines.h>
#include <machine/monotonic.h>
#include <pexpert/pexpert.h>
#include <sys/param.h> /* NULL */
#include <sys/stat.h> /* dev_t */
#include <miscfs/devfs/devfs.h> /* must come after sys/stat.h */
#include <sys/conf.h> /* must come after sys/stat.h */
#include <sys/resource_private.h>
#include <sys/sysctl.h>
#include <sys/sysproto.h>
#include <sys/systm.h>
#include <sys/types.h>
#include <sys/monotonic.h>
#include <kern/cpc.h>
static int mt_cdev_open(dev_t dev, int flags, int devtype, proc_t p);
static int mt_cdev_close(dev_t dev, int flags, int devtype, proc_t p);
static int mt_cdev_ioctl(dev_t dev, unsigned long cmd, char *uptr, int fflag,
proc_t p);
#define MT_NODE "monotonic"
static const struct cdevsw mt_cdevsw = {
.d_open = mt_cdev_open,
.d_close = mt_cdev_close,
.d_ioctl = mt_cdev_ioctl,
.d_read = eno_rdwrt, .d_write = eno_rdwrt, .d_stop = eno_stop,
.d_reset = eno_reset, .d_ttys = NULL, .d_select = eno_select,
.d_mmap = eno_mmap, .d_strategy = eno_strat, .d_type = 0
};
/*
* Written at initialization, read-only thereafter.
*/
LCK_GRP_DECLARE(mt_lock_grp, MT_NODE);
static int mt_dev_major;
static mt_device_t
mt_get_device(dev_t devnum)
{
const int minor_dev = minor(devnum);
if (minor_dev < 0 || minor_dev >= MT_NDEVS) {
return NULL;
}
return &mt_devices[minor_dev];
}
static void
mt_device_lock(mt_device_t dev)
{
lck_mtx_lock(&dev->mtd_lock);
}
static void
mt_device_unlock(mt_device_t dev)
{
lck_mtx_unlock(&dev->mtd_lock);
}
static void
mt_device_assert_lock_held(__assert_only mt_device_t dev)
{
LCK_MTX_ASSERT(&dev->mtd_lock, LCK_MTX_ASSERT_OWNED);
}
static void
mt_device_assert_inuse(__assert_only mt_device_t dev)
{
assert(dev->mtd_inuse == true);
}
int
mt_dev_init(void)
{
mt_dev_major = cdevsw_add(-1 /* allocate a major number */, &mt_cdevsw);
if (mt_dev_major < 0) {
panic("monotonic: cdevsw_add failed: %d", mt_dev_major);
}
for (int i = 0; i < MT_NDEVS; i++) {
if (mt_devices[i].mtd_init(&mt_devices[i])) {
continue;
}
assert(mt_devices[i].mtd_ncounters > 0);
dev_t dev = makedev(mt_dev_major, i);
void *node = devfs_make_node(dev, DEVFS_CHAR, UID_ROOT,
GID_WINDOWSERVER, 0666, MT_NODE "/%s",
mt_devices[i].mtd_name);
if (!node) {
panic("monotonic: devfs_make_node failed for '%s'",
mt_devices[i].mtd_name);
}
lck_mtx_init(&mt_devices[i].mtd_lock, &mt_lock_grp, LCK_ATTR_NULL);
}
return 0;
}
static int
mt_cdev_open(dev_t devnum, __unused int flags, __unused int devtype,
__unused proc_t p)
{
int error = 0;
mt_device_t dev = mt_get_device(devnum);
if (!dev) {
return ENODEV;
}
if (!cpc_hw_acquire(CPC_HW_UPMU, "monotonic")) {
return EBUSY;
}
mt_device_lock(dev);
if (dev->mtd_inuse) {
error = EALREADY;
} else if (!mt_acquire_counters()) {
error = ECONNREFUSED;
} else {
dev->mtd_reset();
dev->mtd_inuse = true;
}
mt_device_unlock(dev);
if (error != 0) {
cpc_hw_release(CPC_HW_UPMU, "monotonic");
}
return error;
}
static int
mt_cdev_close(dev_t devnum, __unused int flags, __unused int devtype,
__unused struct proc *p)
{
mt_device_t dev = mt_get_device(devnum);
if (!dev) {
return ENODEV;
}
cpc_hw_release(CPC_HW_UPMU, "monotonic");
mt_device_lock(dev);
mt_device_assert_inuse(dev);
dev->mtd_inuse = false;
dev->mtd_reset();
mt_release_counters();
mt_device_unlock(dev);
return 0;
}
static int
mt_ctl_add(mt_device_t dev, user_addr_t uptr)
{
int error;
uint32_t ctr;
union monotonic_ctl_add ctl;
mt_device_assert_lock_held(dev);
error = copyin(uptr, &ctl, sizeof(ctl.in));
if (error) {
return error;
}
error = dev->mtd_add(&ctl.in.config, &ctr);
if (error) {
return error;
}
ctl.out.ctr = ctr;
error = copyout(&ctl, uptr, sizeof(ctl.out));
if (error) {
return error;
}
return 0;
}
static int
mt_ctl_counts(mt_device_t dev, user_addr_t uptr)
{
int error;
union monotonic_ctl_counts ctl;
mt_device_assert_lock_held(dev);
error = copyin(uptr, &ctl, sizeof(ctl.in));
if (error) {
return error;
}
if (ctl.in.ctr_mask == 0) {
return EINVAL;
}
{
uint64_t counts[dev->mtd_nmonitors][dev->mtd_ncounters];
memset(counts, 0,
dev->mtd_ncounters * dev->mtd_nmonitors * sizeof(counts[0][0]));
error = dev->mtd_read(ctl.in.ctr_mask, (uint64_t *)counts);
if (error) {
return error;
}
error = copyout(&counts, uptr, sizeof(counts));
if (error) {
return error;
}
}
return 0;
}
static int
mt_ctl_enable(mt_device_t dev, user_addr_t uptr)
{
int error;
union monotonic_ctl_enable ctl;
mt_device_assert_lock_held(dev);
error = copyin(uptr, &ctl, sizeof(ctl));
if (error) {
return error;
}
dev->mtd_enable(ctl.in.enable);
return 0;
}
static int
mt_ctl_reset(mt_device_t dev)
{
mt_device_assert_lock_held(dev);
dev->mtd_reset();
return 0;
}
static int
mt_cdev_ioctl(dev_t devnum, unsigned long cmd, char *arg, __unused int flags,
__unused proc_t p)
{
int error = ENODEV;
user_addr_t uptr = *(user_addr_t *)(void *)arg;
mt_device_t dev = mt_get_device(devnum);
if (!dev) {
return ENODEV;
}
mt_device_lock(dev);
switch (cmd) {
case MT_IOC_RESET:
error = mt_ctl_reset(dev);
break;
case MT_IOC_ADD:
error = mt_ctl_add(dev, uptr);
break;
case MT_IOC_ENABLE:
error = mt_ctl_enable(dev, uptr);
break;
case MT_IOC_COUNTS:
error = mt_ctl_counts(dev, uptr);
break;
case MT_IOC_GET_INFO: {
union monotonic_ctl_info info = {
.out = {
.nmonitors = dev->mtd_nmonitors,
.ncounters = dev->mtd_ncounters,
},
};
error = copyout(&info, uptr, sizeof(info));
break;
}
default:
error = ENODEV;
break;
}
mt_device_unlock(dev);
return error;
}
static void
_convert_usage_to_counts(struct recount_usage *usage, uint64_t *counts)
{
#if CONFIG_PERVASIVE_CPI
counts[MT_CORE_INSTRS] = usage->ru_metrics[RCT_LVL_KERNEL].rm_instructions;
counts[MT_CORE_CYCLES] = usage->ru_metrics[RCT_LVL_KERNEL].rm_cycles;
#else /* CONFIG_PERVASIVE_CPI */
#pragma unused(usage, counts)
#endif /* !CONFIG_PERVASIVE_CPI */
}
enum mt_sysctl {
MT_SUPPORTED,
MT_PMIS,
MT_RETROGRADE,
MT_TASK_THREAD,
MT_DEBUG,
MT_KDBG_TEST,
MT_FIX_CPU_PERF,
MT_FIX_THREAD_PERF,
MT_FIX_TASK_PERF,
};
static int
mt_sysctl SYSCTL_HANDLER_ARGS
{
#pragma unused(oidp, arg2)
uint64_t start[MT_CORE_NFIXED] = { 0 }, end[MT_CORE_NFIXED] = { 0 };
uint64_t counts[2] = { 0 };
switch ((enum mt_sysctl)arg1) {
case MT_SUPPORTED:
return sysctl_io_number(req, (int)mt_core_supported, sizeof(int), NULL, NULL);
case MT_PMIS:
return sysctl_io_number(req, mt_count_pmis(), sizeof(uint64_t), NULL, NULL);
case MT_RETROGRADE: {
uint64_t value = os_atomic_load_wide(&mt_retrograde, relaxed);
return sysctl_io_number(req, value, sizeof(mt_retrograde), NULL, NULL);
}
case MT_TASK_THREAD:
return sysctl_io_number(req, (int)mt_core_supported, sizeof(int), NULL, NULL);
case MT_DEBUG: {
int value = mt_debug;
int r = sysctl_io_number(req, value, sizeof(value), &value, NULL);
if (r) {
return r;
}
mt_debug = value;
return 0;
}
case MT_KDBG_TEST: {
if (req->newptr == USER_ADDR_NULL) {
return EINVAL;
}
int intrs_en = ml_set_interrupts_enabled(FALSE);
MT_KDBG_TMPCPU_START(0x3fff);
MT_KDBG_TMPCPU_END(0x3fff);
ml_set_interrupts_enabled(intrs_en);
return 0;
}
case MT_FIX_CPU_PERF: {
int intrs_en = ml_set_interrupts_enabled(FALSE);
mt_fixed_counts(start);
mt_fixed_counts(end);
ml_set_interrupts_enabled(intrs_en);
goto copyout_counts;
}
case MT_FIX_THREAD_PERF: {
int intrs_en = ml_set_interrupts_enabled(FALSE);
struct recount_usage start_usage = { 0 };
struct recount_usage end_usage = { 0 };
recount_current_thread_usage(&start_usage);
recount_current_thread_usage(&end_usage);
ml_set_interrupts_enabled(intrs_en);
_convert_usage_to_counts(&start_usage, start);
_convert_usage_to_counts(&end_usage, end);
goto copyout_counts;
}
case MT_FIX_TASK_PERF: {
int intrs_en = ml_set_interrupts_enabled(FALSE);
struct recount_usage start_usage = { 0 };
struct recount_usage end_usage = { 0 };
recount_current_task_usage(&start_usage);
recount_current_task_usage(&end_usage);
ml_set_interrupts_enabled(intrs_en);
_convert_usage_to_counts(&start_usage, start);
_convert_usage_to_counts(&end_usage, end);
goto copyout_counts;
}
default:
return ENOENT;
}
copyout_counts:
counts[0] = end[MT_CORE_INSTRS] - start[MT_CORE_INSTRS];
counts[1] = end[MT_CORE_CYCLES] - start[MT_CORE_CYCLES];
return copyout(counts, req->oldptr, MIN(req->oldlen, sizeof(counts)));
}
SYSCTL_DECL(_kern_monotonic);
SYSCTL_NODE(_kern, OID_AUTO, monotonic, CTLFLAG_RW | CTLFLAG_LOCKED, 0,
"monotonic");
#define MT_SYSCTL(NAME, ARG, FLAGS, SIZE, SIZESTR, DESC) \
SYSCTL_PROC(_kern_monotonic, OID_AUTO, NAME, \
CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_LOCKED | (FLAGS), \
(void *)(ARG), SIZE, mt_sysctl, SIZESTR, DESC)
MT_SYSCTL(supported, MT_SUPPORTED, 0, sizeof(int), "I",
"whether monotonic is supported");
MT_SYSCTL(debug, MT_DEBUG, CTLFLAG_MASKED, sizeof(int), "I",
"whether monotonic is printing debug messages");
MT_SYSCTL(pmis, MT_PMIS, 0, sizeof(uint64_t), "Q",
"number of PMIs seen");
MT_SYSCTL(retrograde_updates, MT_RETROGRADE, 0, sizeof(uint64_t), "Q",
"number of times a counter appeared to go backwards");
MT_SYSCTL(task_thread_counting, MT_TASK_THREAD, 0, sizeof(int), "I",
"whether task and thread counting is enabled");
MT_SYSCTL(kdebug_test, MT_KDBG_TEST, CTLFLAG_MASKED, sizeof(int), "O",
"whether task and thread counting is enabled");
MT_SYSCTL(fixed_cpu_perf, MT_FIX_CPU_PERF, CTLFLAG_MASKED,
sizeof(uint64_t) * 2, "O",
"overhead of accessing the current CPU's counters");
MT_SYSCTL(fixed_thread_perf, MT_FIX_THREAD_PERF, CTLFLAG_MASKED,
sizeof(uint64_t) * 2, "O",
"overhead of accessing the current thread's counters");
MT_SYSCTL(fixed_task_perf, MT_FIX_TASK_PERF, CTLFLAG_MASKED,
sizeof(uint64_t) * 2, "O",
"overhead of accessing the current task's counters");