| /* |
| * User interface for Resource Alloction in Resource Director Technology(RDT) |
| * |
| * Copyright (C) 2016 Intel Corporation |
| * |
| * Author: Fenghua Yu <fenghua.yu@intel.com> |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| * |
| * More information about RDT be found in the Intel (R) x86 Architecture |
| * Software Developer Manual. |
| */ |
| |
| #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt |
| |
| #include <linux/cpu.h> |
| #include <linux/fs.h> |
| #include <linux/sysfs.h> |
| #include <linux/kernfs.h> |
| #include <linux/seq_file.h> |
| #include <linux/sched/signal.h> |
| #include <linux/sched/task.h> |
| #include <linux/slab.h> |
| #include <linux/cpu.h> |
| #include <linux/task_work.h> |
| |
| #include <uapi/linux/magic.h> |
| |
| #include <asm/intel_rdt.h> |
| #include <asm/intel_rdt_common.h> |
| |
| DEFINE_STATIC_KEY_FALSE(rdt_enable_key); |
| struct kernfs_root *rdt_root; |
| struct rdtgroup rdtgroup_default; |
| LIST_HEAD(rdt_all_groups); |
| |
| /* Kernel fs node for "info" directory under root */ |
| static struct kernfs_node *kn_info; |
| |
| /* |
| * Trivial allocator for CLOSIDs. Since h/w only supports a small number, |
| * we can keep a bitmap of free CLOSIDs in a single integer. |
| * |
| * Using a global CLOSID across all resources has some advantages and |
| * some drawbacks: |
| * + We can simply set "current->closid" to assign a task to a resource |
| * group. |
| * + Context switch code can avoid extra memory references deciding which |
| * CLOSID to load into the PQR_ASSOC MSR |
| * - We give up some options in configuring resource groups across multi-socket |
| * systems. |
| * - Our choices on how to configure each resource become progressively more |
| * limited as the number of resources grows. |
| */ |
| static int closid_free_map; |
| |
| static void closid_init(void) |
| { |
| struct rdt_resource *r; |
| int rdt_min_closid = 32; |
| |
| /* Compute rdt_min_closid across all resources */ |
| for_each_enabled_rdt_resource(r) |
| rdt_min_closid = min(rdt_min_closid, r->num_closid); |
| |
| closid_free_map = BIT_MASK(rdt_min_closid) - 1; |
| |
| /* CLOSID 0 is always reserved for the default group */ |
| closid_free_map &= ~1; |
| } |
| |
| int closid_alloc(void) |
| { |
| int closid = ffs(closid_free_map); |
| |
| if (closid == 0) |
| return -ENOSPC; |
| closid--; |
| closid_free_map &= ~(1 << closid); |
| |
| return closid; |
| } |
| |
| static void closid_free(int closid) |
| { |
| closid_free_map |= 1 << closid; |
| } |
| |
| /* set uid and gid of rdtgroup dirs and files to that of the creator */ |
| static int rdtgroup_kn_set_ugid(struct kernfs_node *kn) |
| { |
| struct iattr iattr = { .ia_valid = ATTR_UID | ATTR_GID, |
| .ia_uid = current_fsuid(), |
| .ia_gid = current_fsgid(), }; |
| |
| if (uid_eq(iattr.ia_uid, GLOBAL_ROOT_UID) && |
| gid_eq(iattr.ia_gid, GLOBAL_ROOT_GID)) |
| return 0; |
| |
| return kernfs_setattr(kn, &iattr); |
| } |
| |
| static int rdtgroup_add_file(struct kernfs_node *parent_kn, struct rftype *rft) |
| { |
| struct kernfs_node *kn; |
| int ret; |
| |
| kn = __kernfs_create_file(parent_kn, rft->name, rft->mode, |
| 0, rft->kf_ops, rft, NULL, NULL); |
| if (IS_ERR(kn)) |
| return PTR_ERR(kn); |
| |
| ret = rdtgroup_kn_set_ugid(kn); |
| if (ret) { |
| kernfs_remove(kn); |
| return ret; |
| } |
| |
| return 0; |
| } |
| |
| static int rdtgroup_add_files(struct kernfs_node *kn, struct rftype *rfts, |
| int len) |
| { |
| struct rftype *rft; |
| int ret; |
| |
| lockdep_assert_held(&rdtgroup_mutex); |
| |
| for (rft = rfts; rft < rfts + len; rft++) { |
| ret = rdtgroup_add_file(kn, rft); |
| if (ret) |
| goto error; |
| } |
| |
| return 0; |
| error: |
| pr_warn("Failed to add %s, err=%d\n", rft->name, ret); |
| while (--rft >= rfts) |
| kernfs_remove_by_name(kn, rft->name); |
| return ret; |
| } |
| |
| static int rdtgroup_seqfile_show(struct seq_file *m, void *arg) |
| { |
| struct kernfs_open_file *of = m->private; |
| struct rftype *rft = of->kn->priv; |
| |
| if (rft->seq_show) |
| return rft->seq_show(of, m, arg); |
| return 0; |
| } |
| |
| static ssize_t rdtgroup_file_write(struct kernfs_open_file *of, char *buf, |
| size_t nbytes, loff_t off) |
| { |
| struct rftype *rft = of->kn->priv; |
| |
| if (rft->write) |
| return rft->write(of, buf, nbytes, off); |
| |
| return -EINVAL; |
| } |
| |
| static struct kernfs_ops rdtgroup_kf_single_ops = { |
| .atomic_write_len = PAGE_SIZE, |
| .write = rdtgroup_file_write, |
| .seq_show = rdtgroup_seqfile_show, |
| }; |
| |
| static int rdtgroup_cpus_show(struct kernfs_open_file *of, |
| struct seq_file *s, void *v) |
| { |
| struct rdtgroup *rdtgrp; |
| int ret = 0; |
| |
| rdtgrp = rdtgroup_kn_lock_live(of->kn); |
| |
| if (rdtgrp) |
| seq_printf(s, "%*pb\n", cpumask_pr_args(&rdtgrp->cpu_mask)); |
| else |
| ret = -ENOENT; |
| rdtgroup_kn_unlock(of->kn); |
| |
| return ret; |
| } |
| |
| /* |
| * This is safe against intel_rdt_sched_in() called from __switch_to() |
| * because __switch_to() is executed with interrupts disabled. A local call |
| * from rdt_update_closid() is proteced against __switch_to() because |
| * preemption is disabled. |
| */ |
| static void rdt_update_cpu_closid(void *closid) |
| { |
| if (closid) |
| this_cpu_write(cpu_closid, *(int *)closid); |
| /* |
| * We cannot unconditionally write the MSR because the current |
| * executing task might have its own closid selected. Just reuse |
| * the context switch code. |
| */ |
| intel_rdt_sched_in(); |
| } |
| |
| /* |
| * Update the PGR_ASSOC MSR on all cpus in @cpu_mask, |
| * |
| * Per task closids must have been set up before calling this function. |
| * |
| * The per cpu closids are updated with the smp function call, when @closid |
| * is not NULL. If @closid is NULL then all affected percpu closids must |
| * have been set up before calling this function. |
| */ |
| static void |
| rdt_update_closid(const struct cpumask *cpu_mask, int *closid) |
| { |
| int cpu = get_cpu(); |
| |
| if (cpumask_test_cpu(cpu, cpu_mask)) |
| rdt_update_cpu_closid(closid); |
| smp_call_function_many(cpu_mask, rdt_update_cpu_closid, closid, 1); |
| put_cpu(); |
| } |
| |
| static ssize_t rdtgroup_cpus_write(struct kernfs_open_file *of, |
| char *buf, size_t nbytes, loff_t off) |
| { |
| cpumask_var_t tmpmask, newmask; |
| struct rdtgroup *rdtgrp, *r; |
| int ret; |
| |
| if (!buf) |
| return -EINVAL; |
| |
| if (!zalloc_cpumask_var(&tmpmask, GFP_KERNEL)) |
| return -ENOMEM; |
| if (!zalloc_cpumask_var(&newmask, GFP_KERNEL)) { |
| free_cpumask_var(tmpmask); |
| return -ENOMEM; |
| } |
| |
| rdtgrp = rdtgroup_kn_lock_live(of->kn); |
| if (!rdtgrp) { |
| ret = -ENOENT; |
| goto unlock; |
| } |
| |
| ret = cpumask_parse(buf, newmask); |
| if (ret) |
| goto unlock; |
| |
| /* check that user didn't specify any offline cpus */ |
| cpumask_andnot(tmpmask, newmask, cpu_online_mask); |
| if (cpumask_weight(tmpmask)) { |
| ret = -EINVAL; |
| goto unlock; |
| } |
| |
| /* Check whether cpus are dropped from this group */ |
| cpumask_andnot(tmpmask, &rdtgrp->cpu_mask, newmask); |
| if (cpumask_weight(tmpmask)) { |
| /* Can't drop from default group */ |
| if (rdtgrp == &rdtgroup_default) { |
| ret = -EINVAL; |
| goto unlock; |
| } |
| /* Give any dropped cpus to rdtgroup_default */ |
| cpumask_or(&rdtgroup_default.cpu_mask, |
| &rdtgroup_default.cpu_mask, tmpmask); |
| rdt_update_closid(tmpmask, &rdtgroup_default.closid); |
| } |
| |
| /* |
| * If we added cpus, remove them from previous group that owned them |
| * and update per-cpu closid |
| */ |
| cpumask_andnot(tmpmask, newmask, &rdtgrp->cpu_mask); |
| if (cpumask_weight(tmpmask)) { |
| list_for_each_entry(r, &rdt_all_groups, rdtgroup_list) { |
| if (r == rdtgrp) |
| continue; |
| cpumask_andnot(&r->cpu_mask, &r->cpu_mask, tmpmask); |
| } |
| rdt_update_closid(tmpmask, &rdtgrp->closid); |
| } |
| |
| /* Done pushing/pulling - update this group with new mask */ |
| cpumask_copy(&rdtgrp->cpu_mask, newmask); |
| |
| unlock: |
| rdtgroup_kn_unlock(of->kn); |
| free_cpumask_var(tmpmask); |
| free_cpumask_var(newmask); |
| |
| return ret ?: nbytes; |
| } |
| |
| struct task_move_callback { |
| struct callback_head work; |
| struct rdtgroup *rdtgrp; |
| }; |
| |
| static void move_myself(struct callback_head *head) |
| { |
| struct task_move_callback *callback; |
| struct rdtgroup *rdtgrp; |
| |
| callback = container_of(head, struct task_move_callback, work); |
| rdtgrp = callback->rdtgrp; |
| |
| /* |
| * If resource group was deleted before this task work callback |
| * was invoked, then assign the task to root group and free the |
| * resource group. |
| */ |
| if (atomic_dec_and_test(&rdtgrp->waitcount) && |
| (rdtgrp->flags & RDT_DELETED)) { |
| current->closid = 0; |
| kfree(rdtgrp); |
| } |
| |
| preempt_disable(); |
| /* update PQR_ASSOC MSR to make resource group go into effect */ |
| intel_rdt_sched_in(); |
| preempt_enable(); |
| |
| kfree(callback); |
| } |
| |
| static int __rdtgroup_move_task(struct task_struct *tsk, |
| struct rdtgroup *rdtgrp) |
| { |
| struct task_move_callback *callback; |
| int ret; |
| |
| callback = kzalloc(sizeof(*callback), GFP_KERNEL); |
| if (!callback) |
| return -ENOMEM; |
| callback->work.func = move_myself; |
| callback->rdtgrp = rdtgrp; |
| |
| /* |
| * Take a refcount, so rdtgrp cannot be freed before the |
| * callback has been invoked. |
| */ |
| atomic_inc(&rdtgrp->waitcount); |
| ret = task_work_add(tsk, &callback->work, true); |
| if (ret) { |
| /* |
| * Task is exiting. Drop the refcount and free the callback. |
| * No need to check the refcount as the group cannot be |
| * deleted before the write function unlocks rdtgroup_mutex. |
| */ |
| atomic_dec(&rdtgrp->waitcount); |
| kfree(callback); |
| } else { |
| tsk->closid = rdtgrp->closid; |
| } |
| return ret; |
| } |
| |
| static int rdtgroup_task_write_permission(struct task_struct *task, |
| struct kernfs_open_file *of) |
| { |
| const struct cred *tcred = get_task_cred(task); |
| const struct cred *cred = current_cred(); |
| int ret = 0; |
| |
| /* |
| * Even if we're attaching all tasks in the thread group, we only |
| * need to check permissions on one of them. |
| */ |
| if (!uid_eq(cred->euid, GLOBAL_ROOT_UID) && |
| !uid_eq(cred->euid, tcred->uid) && |
| !uid_eq(cred->euid, tcred->suid)) |
| ret = -EPERM; |
| |
| put_cred(tcred); |
| return ret; |
| } |
| |
| static int rdtgroup_move_task(pid_t pid, struct rdtgroup *rdtgrp, |
| struct kernfs_open_file *of) |
| { |
| struct task_struct *tsk; |
| int ret; |
| |
| rcu_read_lock(); |
| if (pid) { |
| tsk = find_task_by_vpid(pid); |
| if (!tsk) { |
| rcu_read_unlock(); |
| return -ESRCH; |
| } |
| } else { |
| tsk = current; |
| } |
| |
| get_task_struct(tsk); |
| rcu_read_unlock(); |
| |
| ret = rdtgroup_task_write_permission(tsk, of); |
| if (!ret) |
| ret = __rdtgroup_move_task(tsk, rdtgrp); |
| |
| put_task_struct(tsk); |
| return ret; |
| } |
| |
| static ssize_t rdtgroup_tasks_write(struct kernfs_open_file *of, |
| char *buf, size_t nbytes, loff_t off) |
| { |
| struct rdtgroup *rdtgrp; |
| int ret = 0; |
| pid_t pid; |
| |
| if (kstrtoint(strstrip(buf), 0, &pid) || pid < 0) |
| return -EINVAL; |
| rdtgrp = rdtgroup_kn_lock_live(of->kn); |
| |
| if (rdtgrp) |
| ret = rdtgroup_move_task(pid, rdtgrp, of); |
| else |
| ret = -ENOENT; |
| |
| rdtgroup_kn_unlock(of->kn); |
| |
| return ret ?: nbytes; |
| } |
| |
| static void show_rdt_tasks(struct rdtgroup *r, struct seq_file *s) |
| { |
| struct task_struct *p, *t; |
| |
| rcu_read_lock(); |
| for_each_process_thread(p, t) { |
| if (t->closid == r->closid) |
| seq_printf(s, "%d\n", t->pid); |
| } |
| rcu_read_unlock(); |
| } |
| |
| static int rdtgroup_tasks_show(struct kernfs_open_file *of, |
| struct seq_file *s, void *v) |
| { |
| struct rdtgroup *rdtgrp; |
| int ret = 0; |
| |
| rdtgrp = rdtgroup_kn_lock_live(of->kn); |
| if (rdtgrp) |
| show_rdt_tasks(rdtgrp, s); |
| else |
| ret = -ENOENT; |
| rdtgroup_kn_unlock(of->kn); |
| |
| return ret; |
| } |
| |
| /* Files in each rdtgroup */ |
| static struct rftype rdtgroup_base_files[] = { |
| { |
| .name = "cpus", |
| .mode = 0644, |
| .kf_ops = &rdtgroup_kf_single_ops, |
| .write = rdtgroup_cpus_write, |
| .seq_show = rdtgroup_cpus_show, |
| }, |
| { |
| .name = "tasks", |
| .mode = 0644, |
| .kf_ops = &rdtgroup_kf_single_ops, |
| .write = rdtgroup_tasks_write, |
| .seq_show = rdtgroup_tasks_show, |
| }, |
| { |
| .name = "schemata", |
| .mode = 0644, |
| .kf_ops = &rdtgroup_kf_single_ops, |
| .write = rdtgroup_schemata_write, |
| .seq_show = rdtgroup_schemata_show, |
| }, |
| }; |
| |
| static int rdt_num_closids_show(struct kernfs_open_file *of, |
| struct seq_file *seq, void *v) |
| { |
| struct rdt_resource *r = of->kn->parent->priv; |
| |
| seq_printf(seq, "%d\n", r->num_closid); |
| |
| return 0; |
| } |
| |
| static int rdt_cbm_mask_show(struct kernfs_open_file *of, |
| struct seq_file *seq, void *v) |
| { |
| struct rdt_resource *r = of->kn->parent->priv; |
| |
| seq_printf(seq, "%x\n", r->max_cbm); |
| |
| return 0; |
| } |
| |
| static int rdt_min_cbm_bits_show(struct kernfs_open_file *of, |
| struct seq_file *seq, void *v) |
| { |
| struct rdt_resource *r = of->kn->parent->priv; |
| |
| seq_printf(seq, "%d\n", r->min_cbm_bits); |
| |
| return 0; |
| } |
| |
| /* rdtgroup information files for one cache resource. */ |
| static struct rftype res_info_files[] = { |
| { |
| .name = "num_closids", |
| .mode = 0444, |
| .kf_ops = &rdtgroup_kf_single_ops, |
| .seq_show = rdt_num_closids_show, |
| }, |
| { |
| .name = "cbm_mask", |
| .mode = 0444, |
| .kf_ops = &rdtgroup_kf_single_ops, |
| .seq_show = rdt_cbm_mask_show, |
| }, |
| { |
| .name = "min_cbm_bits", |
| .mode = 0444, |
| .kf_ops = &rdtgroup_kf_single_ops, |
| .seq_show = rdt_min_cbm_bits_show, |
| }, |
| }; |
| |
| static int rdtgroup_create_info_dir(struct kernfs_node *parent_kn) |
| { |
| struct kernfs_node *kn_subdir; |
| struct rdt_resource *r; |
| int ret; |
| |
| /* create the directory */ |
| kn_info = kernfs_create_dir(parent_kn, "info", parent_kn->mode, NULL); |
| if (IS_ERR(kn_info)) |
| return PTR_ERR(kn_info); |
| kernfs_get(kn_info); |
| |
| for_each_enabled_rdt_resource(r) { |
| kn_subdir = kernfs_create_dir(kn_info, r->name, |
| kn_info->mode, r); |
| if (IS_ERR(kn_subdir)) { |
| ret = PTR_ERR(kn_subdir); |
| goto out_destroy; |
| } |
| kernfs_get(kn_subdir); |
| ret = rdtgroup_kn_set_ugid(kn_subdir); |
| if (ret) |
| goto out_destroy; |
| ret = rdtgroup_add_files(kn_subdir, res_info_files, |
| ARRAY_SIZE(res_info_files)); |
| if (ret) |
| goto out_destroy; |
| kernfs_activate(kn_subdir); |
| } |
| |
| /* |
| * This extra ref will be put in kernfs_remove() and guarantees |
| * that @rdtgrp->kn is always accessible. |
| */ |
| kernfs_get(kn_info); |
| |
| ret = rdtgroup_kn_set_ugid(kn_info); |
| if (ret) |
| goto out_destroy; |
| |
| kernfs_activate(kn_info); |
| |
| return 0; |
| |
| out_destroy: |
| kernfs_remove(kn_info); |
| return ret; |
| } |
| |
| static void l3_qos_cfg_update(void *arg) |
| { |
| bool *enable = arg; |
| |
| wrmsrl(IA32_L3_QOS_CFG, *enable ? L3_QOS_CDP_ENABLE : 0ULL); |
| } |
| |
| static int set_l3_qos_cfg(struct rdt_resource *r, bool enable) |
| { |
| cpumask_var_t cpu_mask; |
| struct rdt_domain *d; |
| int cpu; |
| |
| if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL)) |
| return -ENOMEM; |
| |
| list_for_each_entry(d, &r->domains, list) { |
| /* Pick one CPU from each domain instance to update MSR */ |
| cpumask_set_cpu(cpumask_any(&d->cpu_mask), cpu_mask); |
| } |
| cpu = get_cpu(); |
| /* Update QOS_CFG MSR on this cpu if it's in cpu_mask. */ |
| if (cpumask_test_cpu(cpu, cpu_mask)) |
| l3_qos_cfg_update(&enable); |
| /* Update QOS_CFG MSR on all other cpus in cpu_mask. */ |
| smp_call_function_many(cpu_mask, l3_qos_cfg_update, &enable, 1); |
| put_cpu(); |
| |
| free_cpumask_var(cpu_mask); |
| |
| return 0; |
| } |
| |
| static int cdp_enable(void) |
| { |
| struct rdt_resource *r_l3data = &rdt_resources_all[RDT_RESOURCE_L3DATA]; |
| struct rdt_resource *r_l3code = &rdt_resources_all[RDT_RESOURCE_L3CODE]; |
| struct rdt_resource *r_l3 = &rdt_resources_all[RDT_RESOURCE_L3]; |
| int ret; |
| |
| if (!r_l3->capable || !r_l3data->capable || !r_l3code->capable) |
| return -EINVAL; |
| |
| ret = set_l3_qos_cfg(r_l3, true); |
| if (!ret) { |
| r_l3->enabled = false; |
| r_l3data->enabled = true; |
| r_l3code->enabled = true; |
| } |
| return ret; |
| } |
| |
| static void cdp_disable(void) |
| { |
| struct rdt_resource *r = &rdt_resources_all[RDT_RESOURCE_L3]; |
| |
| r->enabled = r->capable; |
| |
| if (rdt_resources_all[RDT_RESOURCE_L3DATA].enabled) { |
| rdt_resources_all[RDT_RESOURCE_L3DATA].enabled = false; |
| rdt_resources_all[RDT_RESOURCE_L3CODE].enabled = false; |
| set_l3_qos_cfg(r, false); |
| } |
| } |
| |
| static int parse_rdtgroupfs_options(char *data) |
| { |
| char *token, *o = data; |
| int ret = 0; |
| |
| while ((token = strsep(&o, ",")) != NULL) { |
| if (!*token) |
| return -EINVAL; |
| |
| if (!strcmp(token, "cdp")) |
| ret = cdp_enable(); |
| } |
| |
| return ret; |
| } |
| |
| /* |
| * We don't allow rdtgroup directories to be created anywhere |
| * except the root directory. Thus when looking for the rdtgroup |
| * structure for a kernfs node we are either looking at a directory, |
| * in which case the rdtgroup structure is pointed at by the "priv" |
| * field, otherwise we have a file, and need only look to the parent |
| * to find the rdtgroup. |
| */ |
| static struct rdtgroup *kernfs_to_rdtgroup(struct kernfs_node *kn) |
| { |
| if (kernfs_type(kn) == KERNFS_DIR) { |
| /* |
| * All the resource directories use "kn->priv" |
| * to point to the "struct rdtgroup" for the |
| * resource. "info" and its subdirectories don't |
| * have rdtgroup structures, so return NULL here. |
| */ |
| if (kn == kn_info || kn->parent == kn_info) |
| return NULL; |
| else |
| return kn->priv; |
| } else { |
| return kn->parent->priv; |
| } |
| } |
| |
| struct rdtgroup *rdtgroup_kn_lock_live(struct kernfs_node *kn) |
| { |
| struct rdtgroup *rdtgrp = kernfs_to_rdtgroup(kn); |
| |
| if (!rdtgrp) |
| return NULL; |
| |
| atomic_inc(&rdtgrp->waitcount); |
| kernfs_break_active_protection(kn); |
| |
| mutex_lock(&rdtgroup_mutex); |
| |
| /* Was this group deleted while we waited? */ |
| if (rdtgrp->flags & RDT_DELETED) |
| return NULL; |
| |
| return rdtgrp; |
| } |
| |
| void rdtgroup_kn_unlock(struct kernfs_node *kn) |
| { |
| struct rdtgroup *rdtgrp = kernfs_to_rdtgroup(kn); |
| |
| if (!rdtgrp) |
| return; |
| |
| mutex_unlock(&rdtgroup_mutex); |
| |
| if (atomic_dec_and_test(&rdtgrp->waitcount) && |
| (rdtgrp->flags & RDT_DELETED)) { |
| kernfs_unbreak_active_protection(kn); |
| kernfs_put(kn); |
| kfree(rdtgrp); |
| } else { |
| kernfs_unbreak_active_protection(kn); |
| } |
| } |
| |
| static struct dentry *rdt_mount(struct file_system_type *fs_type, |
| int flags, const char *unused_dev_name, |
| void *data) |
| { |
| struct dentry *dentry; |
| int ret; |
| |
| mutex_lock(&rdtgroup_mutex); |
| /* |
| * resctrl file system can only be mounted once. |
| */ |
| if (static_branch_unlikely(&rdt_enable_key)) { |
| dentry = ERR_PTR(-EBUSY); |
| goto out; |
| } |
| |
| ret = parse_rdtgroupfs_options(data); |
| if (ret) { |
| dentry = ERR_PTR(ret); |
| goto out_cdp; |
| } |
| |
| closid_init(); |
| |
| ret = rdtgroup_create_info_dir(rdtgroup_default.kn); |
| if (ret) { |
| dentry = ERR_PTR(ret); |
| goto out_cdp; |
| } |
| |
| dentry = kernfs_mount(fs_type, flags, rdt_root, |
| RDTGROUP_SUPER_MAGIC, NULL); |
| if (IS_ERR(dentry)) |
| goto out_cdp; |
| |
| static_branch_enable(&rdt_enable_key); |
| goto out; |
| |
| out_cdp: |
| cdp_disable(); |
| out: |
| mutex_unlock(&rdtgroup_mutex); |
| |
| return dentry; |
| } |
| |
| static int reset_all_cbms(struct rdt_resource *r) |
| { |
| struct msr_param msr_param; |
| cpumask_var_t cpu_mask; |
| struct rdt_domain *d; |
| int i, cpu; |
| |
| if (!zalloc_cpumask_var(&cpu_mask, GFP_KERNEL)) |
| return -ENOMEM; |
| |
| msr_param.res = r; |
| msr_param.low = 0; |
| msr_param.high = r->num_closid; |
| |
| /* |
| * Disable resource control for this resource by setting all |
| * CBMs in all domains to the maximum mask value. Pick one CPU |
| * from each domain to update the MSRs below. |
| */ |
| list_for_each_entry(d, &r->domains, list) { |
| cpumask_set_cpu(cpumask_any(&d->cpu_mask), cpu_mask); |
| |
| for (i = 0; i < r->num_closid; i++) |
| d->cbm[i] = r->max_cbm; |
| } |
| cpu = get_cpu(); |
| /* Update CBM on this cpu if it's in cpu_mask. */ |
| if (cpumask_test_cpu(cpu, cpu_mask)) |
| rdt_cbm_update(&msr_param); |
| /* Update CBM on all other cpus in cpu_mask. */ |
| smp_call_function_many(cpu_mask, rdt_cbm_update, &msr_param, 1); |
| put_cpu(); |
| |
| free_cpumask_var(cpu_mask); |
| |
| return 0; |
| } |
| |
| /* |
| * Move tasks from one to the other group. If @from is NULL, then all tasks |
| * in the systems are moved unconditionally (used for teardown). |
| * |
| * If @mask is not NULL the cpus on which moved tasks are running are set |
| * in that mask so the update smp function call is restricted to affected |
| * cpus. |
| */ |
| static void rdt_move_group_tasks(struct rdtgroup *from, struct rdtgroup *to, |
| struct cpumask *mask) |
| { |
| struct task_struct *p, *t; |
| |
| read_lock(&tasklist_lock); |
| for_each_process_thread(p, t) { |
| if (!from || t->closid == from->closid) { |
| t->closid = to->closid; |
| #ifdef CONFIG_SMP |
| /* |
| * This is safe on x86 w/o barriers as the ordering |
| * of writing to task_cpu() and t->on_cpu is |
| * reverse to the reading here. The detection is |
| * inaccurate as tasks might move or schedule |
| * before the smp function call takes place. In |
| * such a case the function call is pointless, but |
| * there is no other side effect. |
| */ |
| if (mask && t->on_cpu) |
| cpumask_set_cpu(task_cpu(t), mask); |
| #endif |
| } |
| } |
| read_unlock(&tasklist_lock); |
| } |
| |
| /* |
| * Forcibly remove all of subdirectories under root. |
| */ |
| static void rmdir_all_sub(void) |
| { |
| struct rdtgroup *rdtgrp, *tmp; |
| |
| /* Move all tasks to the default resource group */ |
| rdt_move_group_tasks(NULL, &rdtgroup_default, NULL); |
| |
| list_for_each_entry_safe(rdtgrp, tmp, &rdt_all_groups, rdtgroup_list) { |
| /* Remove each rdtgroup other than root */ |
| if (rdtgrp == &rdtgroup_default) |
| continue; |
| |
| /* |
| * Give any CPUs back to the default group. We cannot copy |
| * cpu_online_mask because a CPU might have executed the |
| * offline callback already, but is still marked online. |
| */ |
| cpumask_or(&rdtgroup_default.cpu_mask, |
| &rdtgroup_default.cpu_mask, &rdtgrp->cpu_mask); |
| |
| kernfs_remove(rdtgrp->kn); |
| list_del(&rdtgrp->rdtgroup_list); |
| kfree(rdtgrp); |
| } |
| /* Notify online CPUs to update per cpu storage and PQR_ASSOC MSR */ |
| get_online_cpus(); |
| rdt_update_closid(cpu_online_mask, &rdtgroup_default.closid); |
| put_online_cpus(); |
| |
| kernfs_remove(kn_info); |
| } |
| |
| static void rdt_kill_sb(struct super_block *sb) |
| { |
| struct rdt_resource *r; |
| |
| mutex_lock(&rdtgroup_mutex); |
| |
| /*Put everything back to default values. */ |
| for_each_enabled_rdt_resource(r) |
| reset_all_cbms(r); |
| cdp_disable(); |
| rmdir_all_sub(); |
| static_branch_disable(&rdt_enable_key); |
| kernfs_kill_sb(sb); |
| mutex_unlock(&rdtgroup_mutex); |
| } |
| |
| static struct file_system_type rdt_fs_type = { |
| .name = "resctrl", |
| .mount = rdt_mount, |
| .kill_sb = rdt_kill_sb, |
| }; |
| |
| static int rdtgroup_mkdir(struct kernfs_node *parent_kn, const char *name, |
| umode_t mode) |
| { |
| struct rdtgroup *parent, *rdtgrp; |
| struct kernfs_node *kn; |
| int ret, closid; |
| |
| /* Only allow mkdir in the root directory */ |
| if (parent_kn != rdtgroup_default.kn) |
| return -EPERM; |
| |
| /* Do not accept '\n' to avoid unparsable situation. */ |
| if (strchr(name, '\n')) |
| return -EINVAL; |
| |
| parent = rdtgroup_kn_lock_live(parent_kn); |
| if (!parent) { |
| ret = -ENODEV; |
| goto out_unlock; |
| } |
| |
| ret = closid_alloc(); |
| if (ret < 0) |
| goto out_unlock; |
| closid = ret; |
| |
| /* allocate the rdtgroup. */ |
| rdtgrp = kzalloc(sizeof(*rdtgrp), GFP_KERNEL); |
| if (!rdtgrp) { |
| ret = -ENOSPC; |
| goto out_closid_free; |
| } |
| rdtgrp->closid = closid; |
| list_add(&rdtgrp->rdtgroup_list, &rdt_all_groups); |
| |
| /* kernfs creates the directory for rdtgrp */ |
| kn = kernfs_create_dir(parent->kn, name, mode, rdtgrp); |
| if (IS_ERR(kn)) { |
| ret = PTR_ERR(kn); |
| goto out_cancel_ref; |
| } |
| rdtgrp->kn = kn; |
| |
| /* |
| * kernfs_remove() will drop the reference count on "kn" which |
| * will free it. But we still need it to stick around for the |
| * rdtgroup_kn_unlock(kn} call below. Take one extra reference |
| * here, which will be dropped inside rdtgroup_kn_unlock(). |
| */ |
| kernfs_get(kn); |
| |
| ret = rdtgroup_kn_set_ugid(kn); |
| if (ret) |
| goto out_destroy; |
| |
| ret = rdtgroup_add_files(kn, rdtgroup_base_files, |
| ARRAY_SIZE(rdtgroup_base_files)); |
| if (ret) |
| goto out_destroy; |
| |
| kernfs_activate(kn); |
| |
| ret = 0; |
| goto out_unlock; |
| |
| out_destroy: |
| kernfs_remove(rdtgrp->kn); |
| out_cancel_ref: |
| list_del(&rdtgrp->rdtgroup_list); |
| kfree(rdtgrp); |
| out_closid_free: |
| closid_free(closid); |
| out_unlock: |
| rdtgroup_kn_unlock(parent_kn); |
| return ret; |
| } |
| |
| static int rdtgroup_rmdir(struct kernfs_node *kn) |
| { |
| int ret, cpu, closid = rdtgroup_default.closid; |
| struct rdtgroup *rdtgrp; |
| cpumask_var_t tmpmask; |
| |
| if (!zalloc_cpumask_var(&tmpmask, GFP_KERNEL)) |
| return -ENOMEM; |
| |
| rdtgrp = rdtgroup_kn_lock_live(kn); |
| if (!rdtgrp) { |
| ret = -EPERM; |
| goto out; |
| } |
| |
| /* Give any tasks back to the default group */ |
| rdt_move_group_tasks(rdtgrp, &rdtgroup_default, tmpmask); |
| |
| /* Give any CPUs back to the default group */ |
| cpumask_or(&rdtgroup_default.cpu_mask, |
| &rdtgroup_default.cpu_mask, &rdtgrp->cpu_mask); |
| |
| /* Update per cpu closid of the moved CPUs first */ |
| for_each_cpu(cpu, &rdtgrp->cpu_mask) |
| per_cpu(cpu_closid, cpu) = closid; |
| /* |
| * Update the MSR on moved CPUs and CPUs which have moved |
| * task running on them. |
| */ |
| cpumask_or(tmpmask, tmpmask, &rdtgrp->cpu_mask); |
| rdt_update_closid(tmpmask, NULL); |
| |
| rdtgrp->flags = RDT_DELETED; |
| closid_free(rdtgrp->closid); |
| list_del(&rdtgrp->rdtgroup_list); |
| |
| /* |
| * one extra hold on this, will drop when we kfree(rdtgrp) |
| * in rdtgroup_kn_unlock() |
| */ |
| kernfs_get(kn); |
| kernfs_remove(rdtgrp->kn); |
| ret = 0; |
| out: |
| rdtgroup_kn_unlock(kn); |
| free_cpumask_var(tmpmask); |
| return ret; |
| } |
| |
| static int rdtgroup_show_options(struct seq_file *seq, struct kernfs_root *kf) |
| { |
| if (rdt_resources_all[RDT_RESOURCE_L3DATA].enabled) |
| seq_puts(seq, ",cdp"); |
| return 0; |
| } |
| |
| static struct kernfs_syscall_ops rdtgroup_kf_syscall_ops = { |
| .mkdir = rdtgroup_mkdir, |
| .rmdir = rdtgroup_rmdir, |
| .show_options = rdtgroup_show_options, |
| }; |
| |
| static int __init rdtgroup_setup_root(void) |
| { |
| int ret; |
| |
| rdt_root = kernfs_create_root(&rdtgroup_kf_syscall_ops, |
| KERNFS_ROOT_CREATE_DEACTIVATED, |
| &rdtgroup_default); |
| if (IS_ERR(rdt_root)) |
| return PTR_ERR(rdt_root); |
| |
| mutex_lock(&rdtgroup_mutex); |
| |
| rdtgroup_default.closid = 0; |
| list_add(&rdtgroup_default.rdtgroup_list, &rdt_all_groups); |
| |
| ret = rdtgroup_add_files(rdt_root->kn, rdtgroup_base_files, |
| ARRAY_SIZE(rdtgroup_base_files)); |
| if (ret) { |
| kernfs_destroy_root(rdt_root); |
| goto out; |
| } |
| |
| rdtgroup_default.kn = rdt_root->kn; |
| kernfs_activate(rdtgroup_default.kn); |
| |
| out: |
| mutex_unlock(&rdtgroup_mutex); |
| |
| return ret; |
| } |
| |
| /* |
| * rdtgroup_init - rdtgroup initialization |
| * |
| * Setup resctrl file system including set up root, create mount point, |
| * register rdtgroup filesystem, and initialize files under root directory. |
| * |
| * Return: 0 on success or -errno |
| */ |
| int __init rdtgroup_init(void) |
| { |
| int ret = 0; |
| |
| ret = rdtgroup_setup_root(); |
| if (ret) |
| return ret; |
| |
| ret = sysfs_create_mount_point(fs_kobj, "resctrl"); |
| if (ret) |
| goto cleanup_root; |
| |
| ret = register_filesystem(&rdt_fs_type); |
| if (ret) |
| goto cleanup_mountpoint; |
| |
| return 0; |
| |
| cleanup_mountpoint: |
| sysfs_remove_mount_point(fs_kobj, "resctrl"); |
| cleanup_root: |
| kernfs_destroy_root(rdt_root); |
| |
| return ret; |
| } |