mirror of
https://github.com/apple-oss-distributions/xnu.git
synced 2026-01-16 23:12:24 +00:00
596 lines
17 KiB
C
596 lines
17 KiB
C
/*
|
|
* Copyright (c) 2019 Apple Computer, Inc. All rights reserved.
|
|
*
|
|
* @APPLE_OSREFERENCE_LICENSE_HEADER_START@
|
|
*
|
|
* This file contains Original Code and/or Modifications of Original Code
|
|
* as defined in and that are subject to the Apple Public Source License
|
|
* Version 2.0 (the 'License'). You may not use this file except in
|
|
* compliance with the License. The rights granted to you under the License
|
|
* may not be used to create, or enable the creation or redistribution of,
|
|
* unlawful or unlicensed copies of an Apple operating system, or to
|
|
* circumvent, violate, or enable the circumvention or violation of, any
|
|
* terms of an Apple operating system software license agreement.
|
|
*
|
|
* Please obtain a copy of the License at
|
|
* http://www.opensource.apple.com/apsl/ and read it before using this file.
|
|
*
|
|
* The Original Code and all software distributed under the License are
|
|
* distributed on an 'AS IS' basis, WITHOUT WARRANTY OF ANY KIND, EITHER
|
|
* EXPRESS OR IMPLIED, AND APPLE HEREBY DISCLAIMS ALL SUCH WARRANTIES,
|
|
* INCLUDING WITHOUT LIMITATION, ANY WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE, QUIET ENJOYMENT OR NON-INFRINGEMENT.
|
|
* Please see the License for the specific language governing rights and
|
|
* limitations under the License.
|
|
*
|
|
* @APPLE_OSREFERENCE_LICENSE_HEADER_END@
|
|
*/
|
|
/**
|
|
* On devices that support it, this test ensures that a mach exception is
|
|
* generated when a matrix-math exception is triggered, and that the
|
|
* matrix register file is correctly preserved or zeroed on context switch.
|
|
*/
|
|
|
|
/*
|
|
* IMPLEMENTATION NOTE:
|
|
*
|
|
* This test code goes to some unusual lengths to avoid calling out to libc or
|
|
* libdarwintest while the CPU is in streaming SVE mode (i.e., between
|
|
* ops->start() and ops->stop()). Both of these libraries are built with SIMD
|
|
* instructions that will cause the test executable to crash while in streaming
|
|
* SVE mode.
|
|
*
|
|
* Ordinarily this is the wrong way to solve this problem. Functions that use
|
|
* streaming SVE mode should have annotations telling the compiler so, and the
|
|
* compiler will automatically generate appropriate interworking code. However
|
|
* this interworking code will stash SME state to memory and temporarily exit
|
|
* streaming SVE mode. We're specifically testing how xnu manages live SME
|
|
* register state, so we can't let the compiler stash and disable this state
|
|
* behind our backs.
|
|
*/
|
|
|
|
#ifdef __arm64__
|
|
#include <mach/error.h>
|
|
#endif /* __arm64__ */
|
|
|
|
#include <darwintest.h>
|
|
#include <pthread.h>
|
|
#include <stdlib.h>
|
|
#include <mach/mach.h>
|
|
#include <mach/thread_act.h>
|
|
#include <mach/thread_status.h>
|
|
#include <mach/exception.h>
|
|
#include <machine/cpu_capabilities.h>
|
|
#include <sys/types.h>
|
|
#include <sys/sysctl.h>
|
|
|
|
#include "arm_matrix.h"
|
|
#include "exc_helpers.h"
|
|
#include "test_utils.h"
|
|
|
|
T_GLOBAL_META(
|
|
T_META_NAMESPACE("xnu.arm"),
|
|
T_META_RADAR_COMPONENT_NAME("xnu"),
|
|
T_META_RADAR_COMPONENT_VERSION("arm"),
|
|
T_META_OWNER("ghackmann"),
|
|
T_META_RUN_CONCURRENTLY(true)
|
|
);
|
|
|
|
#ifdef __arm64__
|
|
|
|
#ifndef EXC_ARM_SME_DISALLOWED
|
|
#define EXC_ARM_SME_DISALLOWED 2
|
|
#endif
|
|
|
|
/* Whether we caught the EXC_BAD_INSTRUCTION mach exception or not. */
|
|
static volatile bool mach_exc_caught = false;
|
|
|
|
static size_t
|
|
bad_instruction_exception_handler(
|
|
__unused mach_port_t task,
|
|
__unused mach_port_t thread,
|
|
exception_type_t type,
|
|
mach_exception_data_t codes,
|
|
__unused uint64_t exception_pc)
|
|
{
|
|
T_QUIET; T_ASSERT_EQ(type, EXC_BAD_INSTRUCTION, "Caught an EXC_BAD_INSTRUCTION exception");
|
|
T_QUIET; T_ASSERT_EQ(codes[0], (uint64_t)EXC_ARM_UNDEFINED, "The subcode is EXC_ARM_UNDEFINED");
|
|
|
|
mach_exc_caught = true;
|
|
return 4;
|
|
}
|
|
#endif
|
|
|
|
|
|
#ifdef __arm64__
|
|
static void
|
|
test_matrix_not_started(const struct arm_matrix_operations *ops)
|
|
{
|
|
if (!ops->is_available()) {
|
|
T_SKIP("Running on non-%s target, skipping...", ops->name);
|
|
}
|
|
|
|
mach_port_t exc_port = create_exception_port(EXC_MASK_BAD_INSTRUCTION);
|
|
|
|
size_t size = ops->data_size();
|
|
uint8_t *d = ops->alloc_data();
|
|
bzero(d, size);
|
|
|
|
ops->start();
|
|
ops->load_one_vector(d);
|
|
ops->stop();
|
|
T_PASS("%s instruction after start instruction should not cause an exception", ops->name);
|
|
|
|
mach_exc_caught = false;
|
|
run_exception_handler(exc_port, bad_instruction_exception_handler);
|
|
ops->load_one_vector(d);
|
|
T_EXPECT_TRUE(mach_exc_caught, "%s instruction before start instruction should cause an exception", ops->name);
|
|
|
|
free(d);
|
|
}
|
|
#endif
|
|
|
|
|
|
T_DECL(sme_not_started,
|
|
"Test that SME instructions before smstart generate mach exceptions.", T_META_TAG_VM_NOT_ELIGIBLE)
|
|
{
|
|
#ifndef __arm64__
|
|
T_SKIP("Running on non-arm64 target, skipping...");
|
|
#else
|
|
test_matrix_not_started(&sme_operations);
|
|
#endif
|
|
}
|
|
|
|
#ifdef __arm64__
|
|
struct test_thread;
|
|
typedef bool (*thread_fn_t)(struct test_thread const* thread);
|
|
|
|
struct test_thread {
|
|
pthread_t thread;
|
|
pthread_t companion_thread;
|
|
thread_fn_t thread_fn;
|
|
uint32_t cpuid;
|
|
uint32_t thread_id;
|
|
const struct arm_matrix_operations *ops;
|
|
};
|
|
|
|
static uint32_t barrier;
|
|
static pthread_cond_t barrier_cond = PTHREAD_COND_INITIALIZER;
|
|
static pthread_mutex_t barrier_lock = PTHREAD_MUTEX_INITIALIZER;
|
|
|
|
static uint32_t end_barrier;
|
|
static pthread_cond_t end_barrier_cond = PTHREAD_COND_INITIALIZER;
|
|
static pthread_mutex_t end_barrier_lock = PTHREAD_MUTEX_INITIALIZER;
|
|
|
|
static void
|
|
test_thread_barrier(void)
|
|
{
|
|
/* Wait for all threads to reach this barrier */
|
|
pthread_mutex_lock(&barrier_lock);
|
|
barrier--;
|
|
if (barrier) {
|
|
while (barrier) {
|
|
pthread_cond_wait(&barrier_cond, &barrier_lock);
|
|
}
|
|
} else {
|
|
pthread_cond_broadcast(&barrier_cond);
|
|
}
|
|
pthread_mutex_unlock(&barrier_lock);
|
|
}
|
|
|
|
static void
|
|
test_thread_notify_exited(void)
|
|
{
|
|
pthread_mutex_lock(&end_barrier_lock);
|
|
if (0 == --end_barrier) {
|
|
pthread_cond_signal(&end_barrier_cond);
|
|
}
|
|
pthread_mutex_unlock(&end_barrier_lock);
|
|
}
|
|
|
|
static void
|
|
wait_for_test_threads(void)
|
|
{
|
|
pthread_mutex_lock(&end_barrier_lock);
|
|
while (end_barrier) {
|
|
pthread_cond_wait(&end_barrier_cond, &end_barrier_lock);
|
|
}
|
|
pthread_mutex_unlock(&end_barrier_lock);
|
|
}
|
|
|
|
static uint32_t
|
|
ncpus(void)
|
|
{
|
|
uint32_t ncpu;
|
|
size_t ncpu_size = sizeof(ncpu);
|
|
int err = sysctlbyname("hw.ncpu", &ncpu, &ncpu_size, NULL, 0);
|
|
T_QUIET; T_ASSERT_POSIX_ZERO(err, "Retrieved CPU count");
|
|
|
|
return ncpu;
|
|
}
|
|
|
|
static int
|
|
thread_bind_cpu_unchecked(uint32_t cpuid)
|
|
{
|
|
/*
|
|
* libc's sysctl() implementation calls strlen(name), which is
|
|
* SIMD-accelerated. Avoid this by directly invoking the libsyscall
|
|
* wrapper with namelen computed at compile time.
|
|
*/
|
|
#define THREAD_BIND_CPU "kern.sched_thread_bind_cpu"
|
|
extern int __sysctlbyname(const char *name, size_t namelen, void *oldp, size_t *oldlenp, void *newp, size_t newlen);
|
|
const char *name = THREAD_BIND_CPU;
|
|
size_t namelen = sizeof(THREAD_BIND_CPU) - 1;
|
|
return __sysctlbyname(name, namelen, NULL, 0, &cpuid, sizeof(cpuid));
|
|
}
|
|
|
|
static void
|
|
thread_bind_cpu(uint32_t cpuid)
|
|
{
|
|
int err = thread_bind_cpu_unchecked(cpuid);
|
|
T_QUIET; T_ASSERT_POSIX_ZERO(err, "Bound thread to CPU %u", cpuid);
|
|
}
|
|
|
|
static void *
|
|
test_thread_shim(void *arg)
|
|
{
|
|
struct test_thread const *thread = arg;
|
|
|
|
thread_bind_cpu(thread->cpuid);
|
|
bool const ret = thread->thread_fn(thread);
|
|
test_thread_notify_exited();
|
|
return (void *)(uintptr_t)ret;
|
|
}
|
|
|
|
static void
|
|
test_on_each_cpu(thread_fn_t thread_fn, const struct arm_matrix_operations *ops, const char *desc)
|
|
{
|
|
uint32_t ncpu = ncpus();
|
|
uint32_t nthreads = ncpu * 2;
|
|
barrier = 1 /* This thread */ + nthreads;
|
|
end_barrier = nthreads;
|
|
struct test_thread *threads = calloc(nthreads, sizeof(threads[0]));
|
|
|
|
for (uint32_t i = 0; i < nthreads; i++) {
|
|
threads[i].thread_fn = thread_fn;
|
|
threads[i].cpuid = i % ncpu;
|
|
threads[i].thread_id = i;
|
|
threads[i].ops = ops;
|
|
|
|
int const err = pthread_create(&threads[i].thread, NULL, test_thread_shim, &threads[i]);
|
|
T_QUIET; T_ASSERT_EQ(err, 0, "%s: created thread #%u", desc, i);
|
|
|
|
// The other of two threads under test pinned to the same CPU.
|
|
threads[(ncpu + i) % nthreads].companion_thread = threads[i].thread;
|
|
}
|
|
|
|
// Wait for all companion_threads to be set.
|
|
test_thread_barrier();
|
|
|
|
// like pthread_join()ing all threads, but without the priority boosting shenanigans.
|
|
wait_for_test_threads();
|
|
|
|
for (uint32_t i = 0; i < nthreads; i++) {
|
|
void *thread_ret_ptr;
|
|
int err = pthread_join(threads[i].thread, &thread_ret_ptr);
|
|
T_QUIET; T_ASSERT_EQ(err, 0, "%s: joined thread #%u", desc, i);
|
|
|
|
bool thread_ret = (uintptr_t)thread_ret_ptr;
|
|
if (thread_ret) {
|
|
T_PASS("%s: thread #%u passed", desc, i);
|
|
} else {
|
|
T_FAIL("%s: thread #%u failed", desc, i);
|
|
}
|
|
}
|
|
|
|
free(threads);
|
|
}
|
|
|
|
static bool
|
|
active_context_switch_thread(struct test_thread const* thread)
|
|
{
|
|
const struct arm_matrix_operations *ops = thread->ops;
|
|
const uint32_t thread_id = thread->thread_id;
|
|
size_t size = ops->data_size();
|
|
uint8_t *d1 = ops->alloc_data();
|
|
memset(d1, (char)thread_id, size);
|
|
|
|
uint8_t *d2 = ops->alloc_data();
|
|
|
|
test_thread_barrier();
|
|
|
|
// companion_thread will be valid only after the barrier.
|
|
thread_t const companion_thread = pthread_mach_thread_np(thread->companion_thread);
|
|
T_QUIET; T_ASSERT_NE(companion_thread, THREAD_NULL, "pthread_mach_thread_np");
|
|
|
|
bool ok = true;
|
|
for (unsigned int i = 0; i < 100000 && ok; i++) {
|
|
ops->start();
|
|
ops->load_data(d1);
|
|
|
|
/*
|
|
* Rescheduling with the matrix registers active must preserve
|
|
* state, even after a context switch.
|
|
*/
|
|
thread_switch(companion_thread, SWITCH_OPTION_NONE, 0);
|
|
|
|
ops->store_data(d2);
|
|
ops->stop();
|
|
|
|
if (memcmp(d1, d2, size)) {
|
|
ok = false;
|
|
}
|
|
}
|
|
|
|
free(d2);
|
|
free(d1);
|
|
return ok;
|
|
}
|
|
|
|
static bool
|
|
inactive_context_switch_thread(struct test_thread const* thread)
|
|
{
|
|
const struct arm_matrix_operations *ops = thread->ops;
|
|
const uint32_t thread_id = thread->thread_id;
|
|
size_t size = ops->data_size();
|
|
uint8_t *d1 = ops->alloc_data();
|
|
memset(d1, (char)thread_id, size);
|
|
|
|
uint8_t *d2 = ops->alloc_data();
|
|
|
|
test_thread_barrier();
|
|
|
|
// companion_thread will be valid only after the barrier.
|
|
thread_t const companion_thread = pthread_mach_thread_np(thread->companion_thread);
|
|
T_QUIET; T_ASSERT_NE(companion_thread, THREAD_NULL, "pthread_mach_thread_np");
|
|
|
|
bool ok = true;
|
|
for (unsigned int i = 0; i < 100000 && ok; i++) {
|
|
ops->start();
|
|
ops->load_data(d1);
|
|
ops->stop();
|
|
|
|
/*
|
|
* Rescheduling with the matrix registers inactive may preserve
|
|
* state or may zero it out.
|
|
*/
|
|
thread_switch(companion_thread, SWITCH_OPTION_NONE, 0);
|
|
|
|
ops->start();
|
|
ops->store_data(d2);
|
|
ops->stop();
|
|
|
|
for (size_t j = 0; j < size; j++) {
|
|
if (d1[j] != d2[j] && d2[j] != 0) {
|
|
ok = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
free(d2);
|
|
free(d1);
|
|
return ok;
|
|
}
|
|
|
|
static void
|
|
test_thread_migration(const struct arm_matrix_operations *ops)
|
|
{
|
|
size_t size = ops->data_size();
|
|
uint8_t *d = ops->alloc_data();
|
|
arc4random_buf(d, size);
|
|
|
|
uint32_t ncpu = ncpus();
|
|
uint8_t *cpu_d[ncpu];
|
|
for (uint32_t cpuid = 0; cpuid < ncpu; cpuid++) {
|
|
cpu_d[cpuid] = ops->alloc_data();
|
|
memset(cpu_d[cpuid], 0, size);
|
|
}
|
|
|
|
ops->start();
|
|
ops->load_data(d);
|
|
for (uint32_t cpuid = 0; cpuid < ncpu; cpuid++) {
|
|
int err = thread_bind_cpu_unchecked(cpuid);
|
|
if (err) {
|
|
ops->stop();
|
|
T_ASSERT_POSIX_ZERO(err, "Bound thread to CPU %u", cpuid);
|
|
}
|
|
ops->store_data(cpu_d[cpuid]);
|
|
}
|
|
ops->stop();
|
|
|
|
for (uint32_t cpuid = 0; cpuid < ncpu; cpuid++) {
|
|
int cmp = memcmp(d, cpu_d[cpuid], size);
|
|
T_EXPECT_EQ(cmp, 0, "Matrix state migrated to CPU %u", cpuid);
|
|
free(cpu_d[cpuid]);
|
|
}
|
|
free(d);
|
|
}
|
|
#endif
|
|
|
|
|
|
T_DECL(sme_context_switch,
|
|
"Test that SME contexts are migrated during context switch and do not leak between process contexts.",
|
|
T_META_BOOTARGS_SET("enable_skstb=1"),
|
|
T_META_REQUIRES_SYSCTL_EQ("hw.optional.arm.FEAT_SME2", 1),
|
|
XNU_T_META_SOC_SPECIFIC, T_META_TAG_VM_NOT_ELIGIBLE)
|
|
{
|
|
#ifndef __arm64__
|
|
T_SKIP("Running on non-arm64 target, skipping...");
|
|
#else
|
|
if (!sme_operations.is_available()) {
|
|
T_SKIP("Running on non-SME target, skipping...");
|
|
}
|
|
|
|
test_thread_migration(&sme_operations);
|
|
test_on_each_cpu(active_context_switch_thread, &sme_operations, "SME context migrates when active");
|
|
test_on_each_cpu(inactive_context_switch_thread, &sme_operations, "SME context does not leak across processes");
|
|
#endif
|
|
}
|
|
|
|
|
|
#if __arm64__
|
|
/*
|
|
* Sequence of events in thread_{get,set}_state test:
|
|
*
|
|
* 1. Parent creates child thread.
|
|
* 2. Child thread signals parent thread to proceed.
|
|
* 3. Parent populates child's matrix state registers via thread_set_state(),
|
|
* and signals child thread to proceed.
|
|
* 4. Child arbitrarily updates each byte in its local matrix register state
|
|
* by adding 1, and signals parent thread to proceed.
|
|
* 5. Parent reads back the child's updated matrix state with
|
|
* thread_get_state(), and confirms that every byte has been modified as
|
|
* expected.
|
|
*/
|
|
static enum thread_state_test_state {
|
|
INIT,
|
|
CHILD_READY,
|
|
PARENT_POPULATED_MATRIX_STATE,
|
|
CHILD_UPDATED_MATRIX_STATE,
|
|
DONE
|
|
} thread_state_test_state;
|
|
|
|
static pthread_cond_t thread_state_test_cond = PTHREAD_COND_INITIALIZER;
|
|
static pthread_mutex_t thread_state_test_lock = PTHREAD_MUTEX_INITIALIZER;
|
|
|
|
static void
|
|
wait_for_thread_state_test_state(enum thread_state_test_state state)
|
|
{
|
|
pthread_mutex_lock(&thread_state_test_lock);
|
|
while (thread_state_test_state != state) {
|
|
pthread_cond_wait(&thread_state_test_cond, &thread_state_test_lock);
|
|
}
|
|
pthread_mutex_unlock(&thread_state_test_lock);
|
|
}
|
|
|
|
static void
|
|
thread_set_state_test_state(enum thread_state_test_state state)
|
|
{
|
|
pthread_mutex_lock(&thread_state_test_lock);
|
|
thread_state_test_state = state;
|
|
pthread_cond_broadcast(&thread_state_test_cond);
|
|
pthread_mutex_unlock(&thread_state_test_lock);
|
|
}
|
|
|
|
static void *
|
|
test_matrix_thread_state_child(void *arg __unused)
|
|
{
|
|
const struct arm_matrix_operations *ops = arg;
|
|
|
|
size_t size = ops->data_size();
|
|
uint8_t *d = ops->alloc_data();
|
|
|
|
|
|
thread_set_state_test_state(CHILD_READY);
|
|
wait_for_thread_state_test_state(PARENT_POPULATED_MATRIX_STATE);
|
|
ops->store_data(d);
|
|
for (size_t i = 0; i < size; i++) {
|
|
d[i]++;
|
|
}
|
|
ops->load_data(d);
|
|
thread_set_state_test_state(CHILD_UPDATED_MATRIX_STATE);
|
|
|
|
wait_for_thread_state_test_state(DONE);
|
|
ops->stop();
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
test_matrix_thread_state(const struct arm_matrix_operations *ops)
|
|
{
|
|
if (!ops->is_available()) {
|
|
T_SKIP("Running on non-%s target, skipping...", ops->name);
|
|
}
|
|
|
|
size_t size = ops->data_size();
|
|
uint8_t *d = ops->alloc_data();
|
|
arc4random_buf(d, size);
|
|
|
|
thread_state_test_state = INIT;
|
|
|
|
pthread_t thread;
|
|
#pragma clang diagnostic push
|
|
#pragma clang diagnostic ignored "-Wincompatible-pointer-types-discards-qualifiers"
|
|
void *arg = ops;
|
|
#pragma clang diagnostic pop
|
|
int err = pthread_create(&thread, NULL, test_matrix_thread_state_child, arg);
|
|
T_QUIET; T_ASSERT_EQ(err, 0, "pthread_create()");
|
|
|
|
mach_port_t mach_thread = pthread_mach_thread_np(thread);
|
|
T_QUIET; T_ASSERT_NE(mach_thread, MACH_PORT_NULL, "pthread_mach_thread_np()");
|
|
|
|
wait_for_thread_state_test_state(CHILD_READY);
|
|
kern_return_t kr = ops->thread_set_state(mach_thread, d);
|
|
T_QUIET; T_ASSERT_EQ(kr, KERN_SUCCESS, "%s thread_set_state()", ops->name);
|
|
thread_set_state_test_state(PARENT_POPULATED_MATRIX_STATE);
|
|
|
|
wait_for_thread_state_test_state(CHILD_UPDATED_MATRIX_STATE);
|
|
uint8_t *thread_d = ops->alloc_data();
|
|
kr = ops->thread_get_state(mach_thread, thread_d);
|
|
T_QUIET; T_ASSERT_EQ(kr, KERN_SUCCESS, "%s thread_get_state()", ops->name);
|
|
for (size_t i = 0; i < size; i++) {
|
|
d[i]++;
|
|
}
|
|
T_EXPECT_EQ(memcmp(d, thread_d, size), 0, "thread_get_state() read expected %s data from child thread", ops->name);
|
|
|
|
thread_set_state_test_state(DONE);
|
|
free(thread_d);
|
|
free(d);
|
|
pthread_join(thread, NULL);
|
|
}
|
|
|
|
#endif
|
|
|
|
#ifdef __arm64__
|
|
|
|
T_DECL(sme_thread_state,
|
|
"Test thread_{get,set}_state with SME thread state.",
|
|
XNU_T_META_SOC_SPECIFIC)
|
|
{
|
|
test_matrix_thread_state(&sme_operations);
|
|
}
|
|
|
|
T_DECL(sme_exception_ports,
|
|
"Test that thread_set_exception_ports rejects SME thread-state flavors.",
|
|
XNU_T_META_SOC_SPECIFIC)
|
|
{
|
|
mach_port_t exc_port;
|
|
mach_port_t task = mach_task_self();
|
|
mach_port_t thread = mach_thread_self();
|
|
|
|
kern_return_t kr = mach_port_allocate(task, MACH_PORT_RIGHT_RECEIVE, &exc_port);
|
|
T_QUIET; T_ASSERT_MACH_SUCCESS(kr, "Allocated mach exception port");
|
|
kr = mach_port_insert_right(task, exc_port, exc_port, MACH_MSG_TYPE_MAKE_SEND);
|
|
T_QUIET; T_ASSERT_MACH_SUCCESS(kr, "Inserted a SEND right into the exception port");
|
|
|
|
kr = thread_set_exception_ports(thread, EXC_MASK_ALL, exc_port, EXCEPTION_STATE, ARM_THREAD_STATE64);
|
|
T_EXPECT_MACH_SUCCESS(kr, "thread_set_exception_ports accepts flavor %u", (unsigned int)ARM_THREAD_STATE64);
|
|
|
|
for (thread_state_flavor_t flavor = ARM_SME_STATE; flavor <= ARM_SME2_STATE; flavor++) {
|
|
kr = thread_set_exception_ports(thread, EXC_MASK_ALL, exc_port, EXCEPTION_STATE, flavor);
|
|
T_EXPECT_MACH_ERROR(kr, KERN_INVALID_ARGUMENT, "thread_set_exception_ports rejects flavor %u", (unsigned int)flavor);
|
|
}
|
|
}
|
|
|
|
T_DECL(sme_max_svl_b_sysctl,
|
|
"Test the hw.optional.arm.sme_max_svl_b sysctl",
|
|
XNU_T_META_SOC_SPECIFIC)
|
|
{
|
|
unsigned int max_svl_b;
|
|
size_t max_svl_b_size = sizeof(max_svl_b);
|
|
|
|
int err = sysctlbyname("hw.optional.arm.sme_max_svl_b", &max_svl_b, &max_svl_b_size, NULL, 0);
|
|
T_QUIET; T_ASSERT_POSIX_SUCCESS(err, "sysctlbyname(hw.optional.arm.sme_max_svl_b)");
|
|
if (sme_operations.is_available()) {
|
|
/* Architecturally SVL must be a power-of-two between 128 and 2048 bits */
|
|
const unsigned int ARCH_MIN_SVL_B = 128 / 8;
|
|
const unsigned int ARCH_MAX_SVL_B = 2048 / 8;
|
|
|
|
T_EXPECT_EQ(__builtin_popcount(max_svl_b), 1, "Maximum SVL_B is a power of 2");
|
|
T_EXPECT_GE(max_svl_b, ARCH_MIN_SVL_B, "Maximum SVL_B >= architectural minimum");
|
|
T_EXPECT_LE(max_svl_b, ARCH_MAX_SVL_B, "Maximum SVL_B <= architectural maximum");
|
|
} else {
|
|
T_EXPECT_EQ(max_svl_b, 0, "Maximum SVL_B is 0 when SME is unavailable");
|
|
}
|
|
}
|
|
#endif /* __arm64__ */
|