On Fri, Feb 14, 2020 at 1:34 PM Daniel Vetter <daniel@xxxxxxxx> wrote: > > I think guidance from Tejun in previos discussions was pretty clear that > he expects cgroups to be both a) standardized and c) sufficient clear > meaning that end-users have a clear understanding of what happens when > they change the resource allocation. > > I'm not sure lgpu here, at least as specified, passes either. I disagree (at least on the characterization of the feedback provided.) I believe this series satisfied the sprite of Tejun's guidance so far (the weight knob for lgpu, for example, was specifically implemented base on his input.) But, I will let Tejun speak for himself after he considered the implementation in detail. Regards, Kenny > But I also > don't have much clue, so pulled Jason in - he understands how this all > gets reflected to userspace apis a lot better than me. > -Daniel > > > > > > > If it's carving up compute power, what's actually being carved up? Time? Execution units/waves/threads? Even if that's the case, what advantage does it give to have it in terms of a fixed set of lgpus where each cgroup gets to pick a fixed set. Does affinity matter that much? Why not just say how many waves the GPU supports and that they have to be allocated in chunks of 16 waves (pulling a number out of thin air) and let the cgroup specify how many waves it wants. > > > > > > Don't get me wrong here. I'm all for the notion of being able to use cgroups to carve up GPU compute resources. However, this sounds to me like the most AMD-specific solution possible. We (Intel) could probably do some sort of carving up as well but we'd likely want to do it with preemption and time-slicing rather than handing out specific EUs. > > > > This has been discussed in the RFC before > > (https://www.spinics.net/lists/cgroups/msg23469.html.) As mentioned > > before, the idea of a compute unit is hardly an AMD specific thing as > > it is in the OpenCL standard and part of the architecture of many > > different vendors. In addition, the interface presented here supports > > Intel's use case. What you described is what I considered as the > > "anonymous resources" view of the lgpu. What you/Intel can do, is to > > register your device to drmcg to have 100 lgpu and users can specify > > simply by count. So if they want to allocate 5% for a cgroup, they > > would set count=5. Per the documentation in this patch: "Some DRM > > devices may only support lgpu as anonymous resources. In such case, > > the significance of the position of the set bits in list will be > > ignored." What Intel does with the user expressed configuration of "5 > > out of 100" is entirely up to Intel (time slice if you like, change to > > specific EUs later if you like, or make it driver configurable to > > support both if you like.) > > > > Regards, > > Kenny > > > > > > > > On Fri, Feb 14, 2020 at 9:57 AM Kenny Ho <Kenny.Ho@xxxxxxx> wrote: > > >> > > >> drm.lgpu > > >> A read-write nested-keyed file which exists on all cgroups. > > >> Each entry is keyed by the DRM device's major:minor. > > >> > > >> lgpu stands for logical GPU, it is an abstraction used to > > >> subdivide a physical DRM device for the purpose of resource > > >> management. This file stores user configuration while the > > >> drm.lgpu.effective reflects the actual allocation after > > >> considering the relationship between the cgroups and their > > >> configurations. > > >> > > >> The lgpu is a discrete quantity that is device specific (i.e. > > >> some DRM devices may have 64 lgpus while others may have 100 > > >> lgpus.) The lgpu is a single quantity that can be allocated > > >> in three different ways denoted by the following nested keys. > > >> > > >> ===== ============================================== > > >> weight Allocate by proportion in relationship with > > >> active sibling cgroups > > >> count Allocate by amount statically, treat lgpu as > > >> anonymous resources > > >> list Allocate statically, treat lgpu as named > > >> resource > > >> ===== ============================================== > > >> > > >> For example: > > >> 226:0 weight=100 count=256 list=0-255 > > >> 226:1 weight=100 count=4 list=0,2,4,6 > > >> 226:2 weight=100 count=32 list=32-63 > > >> 226:3 weight=100 count=0 list= > > >> 226:4 weight=500 count=0 list= > > >> > > >> lgpu is represented by a bitmap and uses the bitmap_parselist > > >> kernel function so the list key input format is a > > >> comma-separated list of decimal numbers and ranges. > > >> > > >> Consecutively set bits are shown as two hyphen-separated decimal > > >> numbers, the smallest and largest bit numbers set in the range. > > >> Optionally each range can be postfixed to denote that only parts > > >> of it should be set. The range will divided to groups of > > >> specific size. > > >> Syntax: range:used_size/group_size > > >> Example: 0-1023:2/256 ==> 0,1,256,257,512,513,768,769 > > >> > > >> The count key is the hamming weight / hweight of the bitmap. > > >> > > >> Weight, count and list accept the max and default keywords. > > >> > > >> Some DRM devices may only support lgpu as anonymous resources. > > >> In such case, the significance of the position of the set bits > > >> in list will be ignored. > > >> > > >> The weight quantity is only in effect when static allocation > > >> is not used (by setting count=0) for this cgroup. The weight > > >> quantity distributes lgpus that are not statically allocated by > > >> the siblings. For example, given siblings cgroupA, cgroupB and > > >> cgroupC for a DRM device that has 64 lgpus, if cgroupA occupies > > >> 0-63, no lgpu is available to be distributed by weight. > > >> Similarly, if cgroupA has list=0-31 and cgroupB has list=16-63, > > >> cgroupC will be starved if it tries to allocate by weight. > > >> > > >> On the other hand, if cgroupA has weight=100 count=0, cgroupB > > >> has list=16-47, and cgroupC has weight=100 count=0, then 32 > > >> lgpus are available to be distributed evenly between cgroupA > > >> and cgroupC. In drm.lgpu.effective, cgroupA will have > > >> list=0-15 and cgroupC will have list=48-63. > > >> > > >> This lgpu resource supports the 'allocation' and 'weight' > > >> resource distribution model. > > >> > > >> drm.lgpu.effective > > >> A read-only nested-keyed file which exists on all cgroups. > > >> Each entry is keyed by the DRM device's major:minor. > > >> > > >> lgpu stands for logical GPU, it is an abstraction used to > > >> subdivide a physical DRM device for the purpose of resource > > >> management. This file reflects the actual allocation after > > >> considering the relationship between the cgroups and their > > >> configurations in drm.lgpu. > > >> > > >> Change-Id: Idde0ef9a331fd67bb9c7eb8ef9978439e6452488 > > >> Signed-off-by: Kenny Ho <Kenny.Ho@xxxxxxx> > > >> --- > > >> Documentation/admin-guide/cgroup-v2.rst | 80 ++++++ > > >> include/drm/drm_cgroup.h | 3 + > > >> include/linux/cgroup_drm.h | 22 ++ > > >> kernel/cgroup/drm.c | 324 +++++++++++++++++++++++- > > >> 4 files changed, 427 insertions(+), 2 deletions(-) > > >> > > >> diff --git a/Documentation/admin-guide/cgroup-v2.rst b/Documentation/admin-guide/cgroup-v2.rst > > >> index ce5dc027366a..d8a41956e5c7 100644 > > >> --- a/Documentation/admin-guide/cgroup-v2.rst > > >> +++ b/Documentation/admin-guide/cgroup-v2.rst > > >> @@ -2120,6 +2120,86 @@ DRM Interface Files > > >> Set largest allocation for /dev/dri/card1 to 4MB > > >> echo "226:1 4m" > drm.buffer.peak.max > > >> > > >> + drm.lgpu > > >> + A read-write nested-keyed file which exists on all cgroups. > > >> + Each entry is keyed by the DRM device's major:minor. > > >> + > > >> + lgpu stands for logical GPU, it is an abstraction used to > > >> + subdivide a physical DRM device for the purpose of resource > > >> + management. This file stores user configuration while the > > >> + drm.lgpu.effective reflects the actual allocation after > > >> + considering the relationship between the cgroups and their > > >> + configurations. > > >> + > > >> + The lgpu is a discrete quantity that is device specific (i.e. > > >> + some DRM devices may have 64 lgpus while others may have 100 > > >> + lgpus.) The lgpu is a single quantity that can be allocated > > >> + in three different ways denoted by the following nested keys. > > >> + > > >> + ===== ============================================== > > >> + weight Allocate by proportion in relationship with > > >> + active sibling cgroups > > >> + count Allocate by amount statically, treat lgpu as > > >> + anonymous resources > > >> + list Allocate statically, treat lgpu as named > > >> + resource > > >> + ===== ============================================== > > >> + > > >> + For example: > > >> + 226:0 weight=100 count=256 list=0-255 > > >> + 226:1 weight=100 count=4 list=0,2,4,6 > > >> + 226:2 weight=100 count=32 list=32-63 > > >> + 226:3 weight=100 count=0 list= > > >> + 226:4 weight=500 count=0 list= > > >> + > > >> + lgpu is represented by a bitmap and uses the bitmap_parselist > > >> + kernel function so the list key input format is a > > >> + comma-separated list of decimal numbers and ranges. > > >> + > > >> + Consecutively set bits are shown as two hyphen-separated decimal > > >> + numbers, the smallest and largest bit numbers set in the range. > > >> + Optionally each range can be postfixed to denote that only parts > > >> + of it should be set. The range will divided to groups of > > >> + specific size. > > >> + Syntax: range:used_size/group_size > > >> + Example: 0-1023:2/256 ==> 0,1,256,257,512,513,768,769 > > >> + > > >> + The count key is the hamming weight / hweight of the bitmap. > > >> + > > >> + Weight, count and list accept the max and default keywords. > > >> + > > >> + Some DRM devices may only support lgpu as anonymous resources. > > >> + In such case, the significance of the position of the set bits > > >> + in list will be ignored. > > >> + > > >> + The weight quantity is only in effect when static allocation > > >> + is not used (by setting count=0) for this cgroup. The weight > > >> + quantity distributes lgpus that are not statically allocated by > > >> + the siblings. For example, given siblings cgroupA, cgroupB and > > >> + cgroupC for a DRM device that has 64 lgpus, if cgroupA occupies > > >> + 0-63, no lgpu is available to be distributed by weight. > > >> + Similarly, if cgroupA has list=0-31 and cgroupB has list=16-63, > > >> + cgroupC will be starved if it tries to allocate by weight. > > >> + > > >> + On the other hand, if cgroupA has weight=100 count=0, cgroupB > > >> + has list=16-47, and cgroupC has weight=100 count=0, then 32 > > >> + lgpus are available to be distributed evenly between cgroupA > > >> + and cgroupC. In drm.lgpu.effective, cgroupA will have > > >> + list=0-15 and cgroupC will have list=48-63. > > >> + > > >> + This lgpu resource supports the 'allocation' and 'weight' > > >> + resource distribution model. > > >> + > > >> + drm.lgpu.effective > > >> + A read-only nested-keyed file which exists on all cgroups. > > >> + Each entry is keyed by the DRM device's major:minor. > > >> + > > >> + lgpu stands for logical GPU, it is an abstraction used to > > >> + subdivide a physical DRM device for the purpose of resource > > >> + management. This file reflects the actual allocation after > > >> + considering the relationship between the cgroups and their > > >> + configurations in drm.lgpu. > > >> + > > >> GEM Buffer Ownership > > >> ~~~~~~~~~~~~~~~~~~~~ > > >> > > >> diff --git a/include/drm/drm_cgroup.h b/include/drm/drm_cgroup.h > > >> index 2b41d4d22e33..619a110cc748 100644 > > >> --- a/include/drm/drm_cgroup.h > > >> +++ b/include/drm/drm_cgroup.h > > >> @@ -17,6 +17,9 @@ struct drmcg_props { > > >> > > >> s64 bo_limits_total_allocated_default; > > >> s64 bo_limits_peak_allocated_default; > > >> + > > >> + int lgpu_capacity; > > >> + DECLARE_BITMAP(lgpu_slots, MAX_DRMCG_LGPU_CAPACITY); > > >> }; > > >> > > >> void drmcg_bind(struct drm_minor (*(*acq_dm)(unsigned int minor_id)), > > >> diff --git a/include/linux/cgroup_drm.h b/include/linux/cgroup_drm.h > > >> index eae400f3d9b4..bb09704e7f71 100644 > > >> --- a/include/linux/cgroup_drm.h > > >> +++ b/include/linux/cgroup_drm.h > > >> @@ -11,10 +11,14 @@ > > >> /* limit defined per the way drm_minor_alloc operates */ > > >> #define MAX_DRM_DEV (64 * DRM_MINOR_RENDER) > > >> > > >> +#define MAX_DRMCG_LGPU_CAPACITY 256 > > >> + > > >> enum drmcg_res_type { > > >> DRMCG_TYPE_BO_TOTAL, > > >> DRMCG_TYPE_BO_PEAK, > > >> DRMCG_TYPE_BO_COUNT, > > >> + DRMCG_TYPE_LGPU, > > >> + DRMCG_TYPE_LGPU_EFF, > > >> __DRMCG_TYPE_LAST, > > >> }; > > >> > > >> @@ -32,6 +36,24 @@ struct drmcg_device_resource { > > >> s64 bo_limits_peak_allocated; > > >> > > >> s64 bo_stats_count_allocated; > > >> + > > >> + /** > > >> + * Logical GPU > > >> + * > > >> + * *_cfg are properties configured by users > > >> + * *_eff are the effective properties being applied to the hardware > > >> + * *_stg is used to calculate _eff before applying to _eff > > >> + * after considering the entire hierarchy > > >> + */ > > >> + DECLARE_BITMAP(lgpu_stg, MAX_DRMCG_LGPU_CAPACITY); > > >> + /* user configurations */ > > >> + s64 lgpu_weight_cfg; > > >> + DECLARE_BITMAP(lgpu_cfg, MAX_DRMCG_LGPU_CAPACITY); > > >> + /* effective lgpu for the cgroup after considering > > >> + * relationship with other cgroup > > >> + */ > > >> + s64 lgpu_count_eff; > > >> + DECLARE_BITMAP(lgpu_eff, MAX_DRMCG_LGPU_CAPACITY); > > >> }; > > >> > > >> /** > > >> diff --git a/kernel/cgroup/drm.c b/kernel/cgroup/drm.c > > >> index 5fcbbc13fa1c..a4e88a3704bb 100644 > > >> --- a/kernel/cgroup/drm.c > > >> +++ b/kernel/cgroup/drm.c > > >> @@ -9,6 +9,7 @@ > > >> #include <linux/seq_file.h> > > >> #include <linux/mutex.h> > > >> #include <linux/kernel.h> > > >> +#include <linux/bitmap.h> > > >> #include <linux/cgroup_drm.h> > > >> #include <drm/drm_file.h> > > >> #include <drm/drm_drv.h> > > >> @@ -41,6 +42,10 @@ enum drmcg_file_type { > > >> DRMCG_FTYPE_DEFAULT, > > >> }; > > >> > > >> +#define LGPU_LIMITS_NAME_LIST "list" > > >> +#define LGPU_LIMITS_NAME_COUNT "count" > > >> +#define LGPU_LIMITS_NAME_WEIGHT "weight" > > >> + > > >> /** > > >> * drmcg_bind - Bind DRM subsystem to cgroup subsystem > > >> * @acq_dm: function pointer to the drm_minor_acquire function > > >> @@ -98,6 +103,13 @@ static inline int init_drmcg_single(struct drmcg *drmcg, struct drm_device *dev) > > >> ddr->bo_limits_peak_allocated = > > >> dev->drmcg_props.bo_limits_peak_allocated_default; > > >> > > >> + bitmap_copy(ddr->lgpu_cfg, dev->drmcg_props.lgpu_slots, > > >> + MAX_DRMCG_LGPU_CAPACITY); > > >> + bitmap_copy(ddr->lgpu_stg, dev->drmcg_props.lgpu_slots, > > >> + MAX_DRMCG_LGPU_CAPACITY); > > >> + > > >> + ddr->lgpu_weight_cfg = CGROUP_WEIGHT_DFL; > > >> + > > >> return 0; > > >> } > > >> > > >> @@ -121,6 +133,120 @@ static inline void drmcg_update_cg_tree(struct drm_device *dev) > > >> mutex_unlock(&cgroup_mutex); > > >> } > > >> > > >> +static void drmcg_calculate_effective_lgpu(struct drm_device *dev, > > >> + const unsigned long *free_static, > > >> + const unsigned long *free_weighted, > > >> + struct drmcg *parent_drmcg) > > >> +{ > > >> + int capacity = dev->drmcg_props.lgpu_capacity; > > >> + DECLARE_BITMAP(lgpu_unused, MAX_DRMCG_LGPU_CAPACITY); > > >> + DECLARE_BITMAP(lgpu_by_weight, MAX_DRMCG_LGPU_CAPACITY); > > >> + struct drmcg_device_resource *parent_ddr; > > >> + struct drmcg_device_resource *ddr; > > >> + int minor = dev->primary->index; > > >> + struct cgroup_subsys_state *pos; > > >> + struct drmcg *child; > > >> + s64 weight_sum = 0; > > >> + s64 unused; > > >> + > > >> + parent_ddr = parent_drmcg->dev_resources[minor]; > > >> + > > >> + if (bitmap_empty(parent_ddr->lgpu_cfg, capacity)) > > >> + /* no static cfg, use weight for calculating the effective */ > > >> + bitmap_copy(parent_ddr->lgpu_stg, free_weighted, capacity); > > >> + else > > >> + /* lgpu statically configured, use the overlap as effective */ > > >> + bitmap_and(parent_ddr->lgpu_stg, free_static, > > >> + parent_ddr->lgpu_cfg, capacity); > > >> + > > >> + /* calculate lgpu available for distribution by weight for children */ > > >> + bitmap_copy(lgpu_unused, parent_ddr->lgpu_stg, capacity); > > >> + css_for_each_child(pos, &parent_drmcg->css) { > > >> + child = css_to_drmcg(pos); > > >> + ddr = child->dev_resources[minor]; > > >> + > > >> + if (bitmap_empty(ddr->lgpu_cfg, capacity)) > > >> + /* no static allocation, participate in weight dist */ > > >> + weight_sum += ddr->lgpu_weight_cfg; > > >> + else > > >> + /* take out statically allocated lgpu by siblings */ > > >> + bitmap_andnot(lgpu_unused, lgpu_unused, ddr->lgpu_cfg, > > >> + capacity); > > >> + } > > >> + > > >> + unused = bitmap_weight(lgpu_unused, capacity); > > >> + > > >> + css_for_each_child(pos, &parent_drmcg->css) { > > >> + child = css_to_drmcg(pos); > > >> + ddr = child->dev_resources[minor]; > > >> + > > >> + bitmap_zero(lgpu_by_weight, capacity); > > >> + /* no static allocation, participate in weight distribution */ > > >> + if (bitmap_empty(ddr->lgpu_cfg, capacity)) { > > >> + int c; > > >> + int p = 0; > > >> + > > >> + for (c = ddr->lgpu_weight_cfg * unused / weight_sum; > > >> + c > 0; c--) { > > >> + p = find_next_bit(lgpu_unused, capacity, p); > > >> + if (p < capacity) { > > >> + clear_bit(p, lgpu_unused); > > >> + set_bit(p, lgpu_by_weight); > > >> + } > > >> + } > > >> + > > >> + } > > >> + > > >> + drmcg_calculate_effective_lgpu(dev, parent_ddr->lgpu_stg, > > >> + lgpu_by_weight, child); > > >> + } > > >> +} > > >> + > > >> +static void drmcg_apply_effective_lgpu(struct drm_device *dev) > > >> +{ > > >> + int capacity = dev->drmcg_props.lgpu_capacity; > > >> + int minor = dev->primary->index; > > >> + struct drmcg_device_resource *ddr; > > >> + struct cgroup_subsys_state *pos; > > >> + struct drmcg *drmcg; > > >> + > > >> + if (root_drmcg == NULL) { > > >> + WARN_ON(root_drmcg == NULL); > > >> + return; > > >> + } > > >> + > > >> + rcu_read_lock(); > > >> + > > >> + /* process the entire cgroup tree from root to simplify the algorithm */ > > >> + drmcg_calculate_effective_lgpu(dev, dev->drmcg_props.lgpu_slots, > > >> + dev->drmcg_props.lgpu_slots, root_drmcg); > > >> + > > >> + /* apply changes to effective only if there is a change */ > > >> + css_for_each_descendant_pre(pos, &root_drmcg->css) { > > >> + drmcg = css_to_drmcg(pos); > > >> + ddr = drmcg->dev_resources[minor]; > > >> + > > >> + if (!bitmap_equal(ddr->lgpu_stg, ddr->lgpu_eff, capacity)) { > > >> + bitmap_copy(ddr->lgpu_eff, ddr->lgpu_stg, capacity); > > >> + ddr->lgpu_count_eff = > > >> + bitmap_weight(ddr->lgpu_eff, capacity); > > >> + } > > >> + } > > >> + rcu_read_unlock(); > > >> +} > > >> + > > >> +static void drmcg_apply_effective(enum drmcg_res_type type, > > >> + struct drm_device *dev, struct drmcg *changed_drmcg) > > >> +{ > > >> + switch (type) { > > >> + case DRMCG_TYPE_LGPU: > > >> + drmcg_apply_effective_lgpu(dev); > > >> + break; > > >> + default: > > >> + break; > > >> + } > > >> +} > > >> + > > >> /** > > >> * drmcg_register_dev - register a DRM device for usage in drm cgroup > > >> * @dev: DRM device > > >> @@ -143,7 +269,13 @@ void drmcg_register_dev(struct drm_device *dev) > > >> { > > >> dev->driver->drmcg_custom_init(dev, &dev->drmcg_props); > > >> > > >> + WARN_ON(dev->drmcg_props.lgpu_capacity != > > >> + bitmap_weight(dev->drmcg_props.lgpu_slots, > > >> + MAX_DRMCG_LGPU_CAPACITY)); > > >> + > > >> drmcg_update_cg_tree(dev); > > >> + > > >> + drmcg_apply_effective(DRMCG_TYPE_LGPU, dev, root_drmcg); > > >> } > > >> mutex_unlock(&drmcg_mutex); > > >> } > > >> @@ -297,7 +429,8 @@ static void drmcg_print_stats(struct drmcg_device_resource *ddr, > > >> } > > >> > > >> static void drmcg_print_limits(struct drmcg_device_resource *ddr, > > >> - struct seq_file *sf, enum drmcg_res_type type) > > >> + struct seq_file *sf, enum drmcg_res_type type, > > >> + struct drm_device *dev) > > >> { > > >> if (ddr == NULL) { > > >> seq_puts(sf, "\n"); > > >> @@ -311,6 +444,25 @@ static void drmcg_print_limits(struct drmcg_device_resource *ddr, > > >> case DRMCG_TYPE_BO_PEAK: > > >> seq_printf(sf, "%lld\n", ddr->bo_limits_peak_allocated); > > >> break; > > >> + case DRMCG_TYPE_LGPU: > > >> + seq_printf(sf, "%s=%lld %s=%d %s=%*pbl\n", > > >> + LGPU_LIMITS_NAME_WEIGHT, > > >> + ddr->lgpu_weight_cfg, > > >> + LGPU_LIMITS_NAME_COUNT, > > >> + bitmap_weight(ddr->lgpu_cfg, > > >> + dev->drmcg_props.lgpu_capacity), > > >> + LGPU_LIMITS_NAME_LIST, > > >> + dev->drmcg_props.lgpu_capacity, > > >> + ddr->lgpu_cfg); > > >> + break; > > >> + case DRMCG_TYPE_LGPU_EFF: > > >> + seq_printf(sf, "%s=%lld %s=%*pbl\n", > > >> + LGPU_LIMITS_NAME_COUNT, > > >> + ddr->lgpu_count_eff, > > >> + LGPU_LIMITS_NAME_LIST, > > >> + dev->drmcg_props.lgpu_capacity, > > >> + ddr->lgpu_eff); > > >> + break; > > >> default: > > >> seq_puts(sf, "\n"); > > >> break; > > >> @@ -329,6 +481,17 @@ static void drmcg_print_default(struct drmcg_props *props, > > >> seq_printf(sf, "%lld\n", > > >> props->bo_limits_peak_allocated_default); > > >> break; > > >> + case DRMCG_TYPE_LGPU: > > >> + seq_printf(sf, "%s=%d %s=%d %s=%*pbl\n", > > >> + LGPU_LIMITS_NAME_WEIGHT, > > >> + CGROUP_WEIGHT_DFL, > > >> + LGPU_LIMITS_NAME_COUNT, > > >> + bitmap_weight(props->lgpu_slots, > > >> + props->lgpu_capacity), > > >> + LGPU_LIMITS_NAME_LIST, > > >> + props->lgpu_capacity, > > >> + props->lgpu_slots); > > >> + break; > > >> default: > > >> seq_puts(sf, "\n"); > > >> break; > > >> @@ -358,7 +521,7 @@ static int drmcg_seq_show_fn(int id, void *ptr, void *data) > > >> drmcg_print_stats(ddr, sf, type); > > >> break; > > >> case DRMCG_FTYPE_LIMIT: > > >> - drmcg_print_limits(ddr, sf, type); > > >> + drmcg_print_limits(ddr, sf, type, minor->dev); > > >> break; > > >> case DRMCG_FTYPE_DEFAULT: > > >> drmcg_print_default(&minor->dev->drmcg_props, sf, type); > > >> @@ -415,6 +578,115 @@ static int drmcg_process_limit_s64_val(char *sval, bool is_mem, > > >> return rc; > > >> } > > >> > > >> +static void drmcg_nested_limit_parse(struct kernfs_open_file *of, > > >> + struct drm_device *dev, char *attrs) > > >> +{ > > >> + DECLARE_BITMAP(tmp_bitmap, MAX_DRMCG_LGPU_CAPACITY); > > >> + DECLARE_BITMAP(chk_bitmap, MAX_DRMCG_LGPU_CAPACITY); > > >> + enum drmcg_res_type type = > > >> + DRMCG_CTF_PRIV2RESTYPE(of_cft(of)->private); > > >> + struct drmcg *drmcg = css_to_drmcg(of_css(of)); > > >> + struct drmcg_props *props = &dev->drmcg_props; > > >> + char *cft_name = of_cft(of)->name; > > >> + int minor = dev->primary->index; > > >> + char *nested = strstrip(attrs); > > >> + struct drmcg_device_resource *ddr = > > >> + drmcg->dev_resources[minor]; > > >> + char *attr; > > >> + char sname[256]; > > >> + char sval[256]; > > >> + s64 val; > > >> + int rc; > > >> + > > >> + while (nested != NULL) { > > >> + attr = strsep(&nested, " "); > > >> + > > >> + if (sscanf(attr, "%255[^=]=%255[^=]", sname, sval) != 2) > > >> + continue; > > >> + > > >> + switch (type) { > > >> + case DRMCG_TYPE_LGPU: > > >> + if (strncmp(sname, LGPU_LIMITS_NAME_LIST, 256) && > > >> + strncmp(sname, LGPU_LIMITS_NAME_COUNT, 256) && > > >> + strncmp(sname, LGPU_LIMITS_NAME_WEIGHT, 256)) > > >> + continue; > > >> + > > >> + if (strncmp(sname, LGPU_LIMITS_NAME_WEIGHT, 256) && > > >> + (!strcmp("max", sval) || > > >> + !strcmp("default", sval))) { > > >> + bitmap_copy(ddr->lgpu_cfg, props->lgpu_slots, > > >> + props->lgpu_capacity); > > >> + > > >> + continue; > > >> + } > > >> + > > >> + if (strncmp(sname, LGPU_LIMITS_NAME_WEIGHT, 256) == 0) { > > >> + rc = drmcg_process_limit_s64_val(sval, > > >> + false, CGROUP_WEIGHT_DFL, > > >> + CGROUP_WEIGHT_MAX, &val); > > >> + > > >> + if (rc || val < CGROUP_WEIGHT_MIN || > > >> + val > CGROUP_WEIGHT_MAX) { > > >> + drmcg_pr_cft_err(drmcg, rc, cft_name, > > >> + minor); > > >> + continue; > > >> + } > > >> + > > >> + ddr->lgpu_weight_cfg = val; > > >> + continue; > > >> + } > > >> + > > >> + if (strncmp(sname, LGPU_LIMITS_NAME_COUNT, 256) == 0) { > > >> + rc = drmcg_process_limit_s64_val(sval, > > >> + false, props->lgpu_capacity, > > >> + props->lgpu_capacity, &val); > > >> + > > >> + if (rc || val < 0) { > > >> + drmcg_pr_cft_err(drmcg, rc, cft_name, > > >> + minor); > > >> + continue; > > >> + } > > >> + > > >> + bitmap_zero(tmp_bitmap, > > >> + MAX_DRMCG_LGPU_CAPACITY); > > >> + bitmap_set(tmp_bitmap, 0, val); > > >> + } > > >> + > > >> + if (strncmp(sname, LGPU_LIMITS_NAME_LIST, 256) == 0) { > > >> + rc = bitmap_parselist(sval, tmp_bitmap, > > >> + MAX_DRMCG_LGPU_CAPACITY); > > >> + > > >> + if (rc) { > > >> + drmcg_pr_cft_err(drmcg, rc, cft_name, > > >> + minor); > > >> + continue; > > >> + } > > >> + > > >> + bitmap_andnot(chk_bitmap, tmp_bitmap, > > >> + props->lgpu_slots, > > >> + MAX_DRMCG_LGPU_CAPACITY); > > >> + > > >> + /* user setting does not intersect with > > >> + * available lgpu */ > > >> + if (!bitmap_empty(chk_bitmap, > > >> + MAX_DRMCG_LGPU_CAPACITY)) { > > >> + drmcg_pr_cft_err(drmcg, 0, cft_name, > > >> + minor); > > >> + continue; > > >> + } > > >> + } > > >> + > > >> + bitmap_copy(ddr->lgpu_cfg, tmp_bitmap, > > >> + props->lgpu_capacity); > > >> + > > >> + break; /* DRMCG_TYPE_LGPU */ > > >> + default: > > >> + break; > > >> + } /* switch (type) */ > > >> + } > > >> +} > > >> + > > >> + > > >> /** > > >> * drmcg_limit_write - parse cgroup interface files to obtain user config > > >> * > > >> @@ -499,9 +771,15 @@ static ssize_t drmcg_limit_write(struct kernfs_open_file *of, char *buf, > > >> > > >> ddr->bo_limits_peak_allocated = val; > > >> break; > > >> + case DRMCG_TYPE_LGPU: > > >> + drmcg_nested_limit_parse(of, dm->dev, sattr); > > >> + break; > > >> default: > > >> break; > > >> } > > >> + > > >> + drmcg_apply_effective(type, dm->dev, drmcg); > > >> + > > >> mutex_unlock(&dm->dev->drmcg_mutex); > > >> > > >> mutex_lock(&drmcg_mutex); > > >> @@ -560,12 +838,51 @@ struct cftype files[] = { > > >> .private = DRMCG_CTF_PRIV(DRMCG_TYPE_BO_COUNT, > > >> DRMCG_FTYPE_STATS), > > >> }, > > >> + { > > >> + .name = "lgpu", > > >> + .seq_show = drmcg_seq_show, > > >> + .write = drmcg_limit_write, > > >> + .private = DRMCG_CTF_PRIV(DRMCG_TYPE_LGPU, > > >> + DRMCG_FTYPE_LIMIT), > > >> + }, > > >> + { > > >> + .name = "lgpu.default", > > >> + .seq_show = drmcg_seq_show, > > >> + .flags = CFTYPE_ONLY_ON_ROOT, > > >> + .private = DRMCG_CTF_PRIV(DRMCG_TYPE_LGPU, > > >> + DRMCG_FTYPE_DEFAULT), > > >> + }, > > >> + { > > >> + .name = "lgpu.effective", > > >> + .seq_show = drmcg_seq_show, > > >> + .private = DRMCG_CTF_PRIV(DRMCG_TYPE_LGPU_EFF, > > >> + DRMCG_FTYPE_LIMIT), > > >> + }, > > >> { } /* terminate */ > > >> }; > > >> > > >> +static int drmcg_online_fn(int id, void *ptr, void *data) > > >> +{ > > >> + struct drm_minor *minor = ptr; > > >> + struct drmcg *drmcg = data; > > >> + > > >> + if (minor->type != DRM_MINOR_PRIMARY) > > >> + return 0; > > >> + > > >> + drmcg_apply_effective(DRMCG_TYPE_LGPU, minor->dev, drmcg); > > >> + > > >> + return 0; > > >> +} > > >> + > > >> +static int drmcg_css_online(struct cgroup_subsys_state *css) > > >> +{ > > >> + return drm_minor_for_each(&drmcg_online_fn, css_to_drmcg(css)); > > >> +} > > >> + > > >> struct cgroup_subsys drm_cgrp_subsys = { > > >> .css_alloc = drmcg_css_alloc, > > >> .css_free = drmcg_css_free, > > >> + .css_online = drmcg_css_online, > > >> .early_init = false, > > >> .legacy_cftypes = files, > > >> .dfl_cftypes = files, > > >> @@ -585,6 +902,9 @@ void drmcg_device_early_init(struct drm_device *dev) > > >> dev->drmcg_props.bo_limits_total_allocated_default = S64_MAX; > > >> dev->drmcg_props.bo_limits_peak_allocated_default = S64_MAX; > > >> > > >> + dev->drmcg_props.lgpu_capacity = MAX_DRMCG_LGPU_CAPACITY; > > >> + bitmap_fill(dev->drmcg_props.lgpu_slots, MAX_DRMCG_LGPU_CAPACITY); > > >> + > > >> drmcg_update_cg_tree(dev); > > >> } > > >> EXPORT_SYMBOL(drmcg_device_early_init); > > >> -- > > >> 2.25.0 > > >> > > >> _______________________________________________ > > >> dri-devel mailing list > > >> dri-devel@xxxxxxxxxxxxxxxxxxxxx > > >> https://lists.freedesktop.org/mailman/listinfo/dri-devel > > -- > Daniel Vetter > Software Engineer, Intel Corporation > http://blog.ffwll.ch _______________________________________________ amd-gfx mailing list amd-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/amd-gfx