|  | // SPDX-License-Identifier: GPL-2.0-only | 
|  | /* | 
|  | * MIPS cacheinfo support | 
|  | */ | 
|  | #include <linux/cacheinfo.h> | 
|  |  | 
|  | /* Populates leaf and increments to next leaf */ | 
|  | #define populate_cache(cache, leaf, c_level, c_type)		\ | 
|  | do {								\ | 
|  | leaf->type = c_type;					\ | 
|  | leaf->level = c_level;					\ | 
|  | leaf->coherency_line_size = c->cache.linesz;		\ | 
|  | leaf->number_of_sets = c->cache.sets;			\ | 
|  | leaf->ways_of_associativity = c->cache.ways;		\ | 
|  | leaf->size = c->cache.linesz * c->cache.sets *		\ | 
|  | c->cache.ways;					\ | 
|  | leaf++;							\ | 
|  | } while (0) | 
|  |  | 
|  | int init_cache_level(unsigned int cpu) | 
|  | { | 
|  | struct cpuinfo_mips *c = ¤t_cpu_data; | 
|  | struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); | 
|  | int levels = 0, leaves = 0; | 
|  |  | 
|  | /* | 
|  | * If Dcache is not set, we assume the cache structures | 
|  | * are not properly initialized. | 
|  | */ | 
|  | if (c->dcache.waysize) | 
|  | levels += 1; | 
|  | else | 
|  | return -ENOENT; | 
|  |  | 
|  |  | 
|  | leaves += (c->icache.waysize) ? 2 : 1; | 
|  |  | 
|  | if (c->vcache.waysize) { | 
|  | levels++; | 
|  | leaves++; | 
|  | } | 
|  |  | 
|  | if (c->scache.waysize) { | 
|  | levels++; | 
|  | leaves++; | 
|  | } | 
|  |  | 
|  | if (c->tcache.waysize) { | 
|  | levels++; | 
|  | leaves++; | 
|  | } | 
|  |  | 
|  | this_cpu_ci->num_levels = levels; | 
|  | this_cpu_ci->num_leaves = leaves; | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static void fill_cpumask_siblings(int cpu, cpumask_t *cpu_map) | 
|  | { | 
|  | int cpu1; | 
|  |  | 
|  | for_each_possible_cpu(cpu1) | 
|  | if (cpus_are_siblings(cpu, cpu1)) | 
|  | cpumask_set_cpu(cpu1, cpu_map); | 
|  | } | 
|  |  | 
|  | static void fill_cpumask_cluster(int cpu, cpumask_t *cpu_map) | 
|  | { | 
|  | int cpu1; | 
|  | int cluster = cpu_cluster(&cpu_data[cpu]); | 
|  |  | 
|  | for_each_possible_cpu(cpu1) | 
|  | if (cpu_cluster(&cpu_data[cpu1]) == cluster) | 
|  | cpumask_set_cpu(cpu1, cpu_map); | 
|  | } | 
|  |  | 
|  | int populate_cache_leaves(unsigned int cpu) | 
|  | { | 
|  | struct cpuinfo_mips *c = ¤t_cpu_data; | 
|  | struct cpu_cacheinfo *this_cpu_ci = get_cpu_cacheinfo(cpu); | 
|  | struct cacheinfo *this_leaf = this_cpu_ci->info_list; | 
|  | int level = 1; | 
|  |  | 
|  | if (c->icache.waysize) { | 
|  | /* I/D caches are per core */ | 
|  | fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map); | 
|  | populate_cache(dcache, this_leaf, level, CACHE_TYPE_DATA); | 
|  | fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map); | 
|  | populate_cache(icache, this_leaf, level, CACHE_TYPE_INST); | 
|  | level++; | 
|  | } else { | 
|  | populate_cache(dcache, this_leaf, level, CACHE_TYPE_UNIFIED); | 
|  | level++; | 
|  | } | 
|  |  | 
|  | if (c->vcache.waysize) { | 
|  | /* Vcache is per core as well */ | 
|  | fill_cpumask_siblings(cpu, &this_leaf->shared_cpu_map); | 
|  | populate_cache(vcache, this_leaf, level, CACHE_TYPE_UNIFIED); | 
|  | level++; | 
|  | } | 
|  |  | 
|  | if (c->scache.waysize) { | 
|  | /* Scache is per cluster */ | 
|  | fill_cpumask_cluster(cpu, &this_leaf->shared_cpu_map); | 
|  | populate_cache(scache, this_leaf, level, CACHE_TYPE_UNIFIED); | 
|  | level++; | 
|  | } | 
|  |  | 
|  | if (c->tcache.waysize) | 
|  | populate_cache(tcache, this_leaf, level, CACHE_TYPE_UNIFIED); | 
|  |  | 
|  | this_cpu_ci->cpu_map_populated = true; | 
|  |  | 
|  | return 0; | 
|  | } |