qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [PATCH v4 09/16] target/i386: Cleanup and use the EPYC mode topology


From: Igor Mammedov
Subject: Re: [PATCH v4 09/16] target/i386: Cleanup and use the EPYC mode topology functions
Date: Tue, 3 Mar 2020 09:47:13 +0100

On Mon, 2 Mar 2020 11:09:14 -0600
Babu Moger <address@hidden> wrote:

> On 2/25/20 1:49 AM, Igor Mammedov wrote:
> > On Mon, 24 Feb 2020 11:29:37 -0600
> > Babu Moger <address@hidden> wrote:
> >   
> >> On 2/24/20 2:52 AM, Igor Mammedov wrote:  
> >>> On Thu, 13 Feb 2020 12:17:25 -0600
> >>> Babu Moger <address@hidden> wrote:
> >>>     
> >>>> Use the new functions from topology.h and delete the unused code. Given 
> >>>> the
> >>>> sockets, nodes, cores and threads, the new functions generate apic id 
> >>>> for EPYC
> >>>> mode. Removes all the hardcoded values.
> >>>>
> >>>> Signed-off-by: Babu Moger <address@hidden>    
> >>>
> >>> modulo MAX() macro, looks fine to me    
> >>
> >> Igor, Sorry. What do you mean here?  
> > 
> > I meant s/MAX(topo_info->nodes_per_pkg, 1)/topo_info->nodes_per_pkg/
> > 
> > after it's made sure that topo_info->nodes_per_pkg is always valid.
> >   
> Noticed that we cannot change it in all the places and assign to valid
> value 1. We need this information to know weather the system is numa
> configured in topology.h. This is similar to ms->numa_state->num_nodes.
> This value is  > 0 if system is numa configured else it is 0. I need this
> information while generating the apicid. I have added comments about it in
> topology.h. Hope this is not a problem.

It should be fine as far as that's the only and best way to handle it
and it's clear from patches why it's done this way.

> > 
> > (I believe I've commented on that somewhere. Series isn't split nicely,
> > so I've ended up applying it all and then reviewing so comments might
> > look out of the place sometimes, hopefully next revision will be easier
> > to review)
> >   
> >>>     
> >>>> ---
> >>>>  target/i386/cpu.c |  162 
> >>>> +++++++++++------------------------------------------
> >>>>  1 file changed, 35 insertions(+), 127 deletions(-)
> >>>>
> >>>> diff --git a/target/i386/cpu.c b/target/i386/cpu.c
> >>>> index 5d6edfd09b..19675eb696 100644
> >>>> --- a/target/i386/cpu.c
> >>>> +++ b/target/i386/cpu.c
> >>>> @@ -338,68 +338,15 @@ static void 
> >>>> encode_cache_cpuid80000006(CPUCacheInfo *l2,
> >>>>      }
> >>>>  }
> >>>>  
> >>>> -/*
> >>>> - * Definitions used for building CPUID Leaf 0x8000001D and 0x8000001E
> >>>> - * Please refer to the AMD64 Architecture Programmer’s Manual Volume 3.
> >>>> - * Define the constants to build the cpu topology. Right now, TOPOEXT
> >>>> - * feature is enabled only on EPYC. So, these constants are based on
> >>>> - * EPYC supported configurations. We may need to handle the cases if
> >>>> - * these values change in future.
> >>>> - */
> >>>> -/* Maximum core complexes in a node */
> >>>> -#define MAX_CCX 2
> >>>> -/* Maximum cores in a core complex */
> >>>> -#define MAX_CORES_IN_CCX 4
> >>>> -/* Maximum cores in a node */
> >>>> -#define MAX_CORES_IN_NODE 8
> >>>> -/* Maximum nodes in a socket */
> >>>> -#define MAX_NODES_PER_SOCKET 4
> >>>> -
> >>>> -/*
> >>>> - * Figure out the number of nodes required to build this config.
> >>>> - * Max cores in a node is 8
> >>>> - */
> >>>> -static int nodes_in_socket(int nr_cores)
> >>>> -{
> >>>> -    int nodes;
> >>>> -
> >>>> -    nodes = DIV_ROUND_UP(nr_cores, MAX_CORES_IN_NODE);
> >>>> -
> >>>> -   /* Hardware does not support config with 3 nodes, return 4 in that 
> >>>> case */
> >>>> -    return (nodes == 3) ? 4 : nodes;
> >>>> -}
> >>>> -
> >>>> -/*
> >>>> - * Decide the number of cores in a core complex with the given nr_cores 
> >>>> using
> >>>> - * following set constants MAX_CCX, MAX_CORES_IN_CCX, MAX_CORES_IN_NODE 
> >>>> and
> >>>> - * MAX_NODES_PER_SOCKET. Maintain symmetry as much as possible
> >>>> - * L3 cache is shared across all cores in a core complex. So, this will 
> >>>> also
> >>>> - * tell us how many cores are sharing the L3 cache.
> >>>> - */
> >>>> -static int cores_in_core_complex(int nr_cores)
> >>>> -{
> >>>> -    int nodes;
> >>>> -
> >>>> -    /* Check if we can fit all the cores in one core complex */
> >>>> -    if (nr_cores <= MAX_CORES_IN_CCX) {
> >>>> -        return nr_cores;
> >>>> -    }
> >>>> -    /* Get the number of nodes required to build this config */
> >>>> -    nodes = nodes_in_socket(nr_cores);
> >>>> -
> >>>> -    /*
> >>>> -     * Divide the cores accros all the core complexes
> >>>> -     * Return rounded up value
> >>>> -     */
> >>>> -    return DIV_ROUND_UP(nr_cores, nodes * MAX_CCX);
> >>>> -}
> >>>> -
> >>>>  /* Encode cache info for CPUID[8000001D] */
> >>>> -static void encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState 
> >>>> *cs,
> >>>> -                                uint32_t *eax, uint32_t *ebx,
> >>>> -                                uint32_t *ecx, uint32_t *edx)
> >>>> +static void encode_cache_cpuid8000001d(CPUCacheInfo *cache,
> >>>> +                                       X86CPUTopoInfo *topo_info,
> >>>> +                                       uint32_t *eax, uint32_t *ebx,
> >>>> +                                       uint32_t *ecx, uint32_t *edx)
> >>>>  {
> >>>>      uint32_t l3_cores;
> >>>> +    unsigned nodes = MAX(topo_info->nodes_per_pkg, 1);
> >>>> +
> >>>>      assert(cache->size == cache->line_size * cache->associativity *
> >>>>                            cache->partitions * cache->sets);
> >>>>  
> >>>> @@ -408,10 +355,13 @@ static void 
> >>>> encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs,
> >>>>  
> >>>>      /* L3 is shared among multiple cores */
> >>>>      if (cache->level == 3) {
> >>>> -        l3_cores = cores_in_core_complex(cs->nr_cores);
> >>>> -        *eax |= ((l3_cores * cs->nr_threads) - 1) << 14;
> >>>> +        l3_cores = DIV_ROUND_UP((topo_info->dies_per_pkg *
> >>>> +                                 topo_info->cores_per_die *
> >>>> +                                 topo_info->threads_per_core),
> >>>> +                                 nodes);
> >>>> +        *eax |= (l3_cores - 1) << 14;
> >>>>      } else {
> >>>> -        *eax |= ((cs->nr_threads - 1) << 14);
> >>>> +        *eax |= ((topo_info->threads_per_core - 1) << 14);
> >>>>      }
> >>>>  
> >>>>      assert(cache->line_size > 0);
> >>>> @@ -431,55 +381,17 @@ static void 
> >>>> encode_cache_cpuid8000001d(CPUCacheInfo *cache, CPUState *cs,
> >>>>             (cache->complex_indexing ? CACHE_COMPLEX_IDX : 0);
> >>>>  }
> >>>>  
> >>>> -/* Data structure to hold the configuration info for a given core index 
> >>>> */
> >>>> -struct core_topology {
> >>>> -    /* core complex id of the current core index */
> >>>> -    int ccx_id;
> >>>> -    /*
> >>>> -     * Adjusted core index for this core in the topology
> >>>> -     * This can be 0,1,2,3 with max 4 cores in a core complex
> >>>> -     */
> >>>> -    int core_id;
> >>>> -    /* Node id for this core index */
> >>>> -    int node_id;
> >>>> -    /* Number of nodes in this config */
> >>>> -    int num_nodes;
> >>>> -};
> >>>> -
> >>>> -/*
> >>>> - * Build the configuration closely match the EPYC hardware. Using the 
> >>>> EPYC
> >>>> - * hardware configuration values (MAX_CCX, MAX_CORES_IN_CCX, 
> >>>> MAX_CORES_IN_NODE)
> >>>> - * right now. This could change in future.
> >>>> - * nr_cores : Total number of cores in the config
> >>>> - * core_id  : Core index of the current CPU
> >>>> - * topo     : Data structure to hold all the config info for this core 
> >>>> index
> >>>> - */
> >>>> -static void build_core_topology(int nr_cores, int core_id,
> >>>> -                                struct core_topology *topo)
> >>>> -{
> >>>> -    int nodes, cores_in_ccx;
> >>>> -
> >>>> -    /* First get the number of nodes required */
> >>>> -    nodes = nodes_in_socket(nr_cores);
> >>>> -
> >>>> -    cores_in_ccx = cores_in_core_complex(nr_cores);
> >>>> -
> >>>> -    topo->node_id = core_id / (cores_in_ccx * MAX_CCX);
> >>>> -    topo->ccx_id = (core_id % (cores_in_ccx * MAX_CCX)) / cores_in_ccx;
> >>>> -    topo->core_id = core_id % cores_in_ccx;
> >>>> -    topo->num_nodes = nodes;
> >>>> -}
> >>>> -
> >>>>  /* Encode cache info for CPUID[8000001E] */
> >>>> -static void encode_topo_cpuid8000001e(CPUState *cs, X86CPU *cpu,
> >>>> +static void encode_topo_cpuid8000001e(X86CPUTopoInfo *topo_info, X86CPU 
> >>>> *cpu,
> >>>>                                         uint32_t *eax, uint32_t *ebx,
> >>>>                                         uint32_t *ecx, uint32_t *edx)
> >>>>  {
> >>>> -    struct core_topology topo = {0};
> >>>> -    unsigned long nodes;
> >>>> +    X86CPUTopoIDs topo_ids = {0};
> >>>> +    unsigned long nodes = MAX(topo_info->nodes_per_pkg, 1);
> >>>>      int shift;
> >>>>  
> >>>> -    build_core_topology(cs->nr_cores, cpu->core_id, &topo);
> >>>> +    x86_topo_ids_from_apicid_epyc(cpu->apic_id, topo_info, &topo_ids);
> >>>> +
> >>>>      *eax = cpu->apic_id;
> >>>>      /*
> >>>>       * CPUID_Fn8000001E_EBX
> >>>> @@ -496,12 +408,8 @@ static void encode_topo_cpuid8000001e(CPUState *cs, 
> >>>> X86CPU *cpu,
> >>>>       *             3 Core complex id
> >>>>       *           1:0 Core id
> >>>>       */
> >>>> -    if (cs->nr_threads - 1) {
> >>>> -        *ebx = ((cs->nr_threads - 1) << 8) | (topo.node_id << 3) |
> >>>> -                (topo.ccx_id << 2) | topo.core_id;
> >>>> -    } else {
> >>>> -        *ebx = (topo.node_id << 4) | (topo.ccx_id << 3) | topo.core_id;
> >>>> -    }
> >>>> +    *ebx = ((topo_info->threads_per_core - 1) << 8) | (topo_ids.node_id 
> >>>> << 3) |
> >>>> +            (topo_ids.core_id);
> >>>>      /*
> >>>>       * CPUID_Fn8000001E_ECX
> >>>>       * 31:11 Reserved
> >>>> @@ -510,9 +418,9 @@ static void encode_topo_cpuid8000001e(CPUState *cs, 
> >>>> X86CPU *cpu,
> >>>>       *         2  Socket id
> >>>>       *       1:0  Node id
> >>>>       */
> >>>> -    if (topo.num_nodes <= 4) {
> >>>> -        *ecx = ((topo.num_nodes - 1) << 8) | (cpu->socket_id << 2) |
> >>>> -                topo.node_id;
> >>>> +
> >>>> +    if (nodes <= 4) {
> >>>> +        *ecx = ((nodes - 1) << 8) | (topo_ids.pkg_id << 2) | 
> >>>> topo_ids.node_id;
> >>>>      } else {
> >>>>          /*
> >>>>           * Node id fix up. Actual hardware supports up to 4 nodes. But 
> >>>> with
> >>>> @@ -527,10 +435,10 @@ static void encode_topo_cpuid8000001e(CPUState 
> >>>> *cs, X86CPU *cpu,
> >>>>           * number of nodes. find_last_bit returns last set bit(0 
> >>>> based). Left
> >>>>           * shift(+1) the socket id to represent all the nodes.
> >>>>           */
> >>>> -        nodes = topo.num_nodes - 1;
> >>>> +        nodes -= 1;
> >>>>          shift = find_last_bit(&nodes, 8);
> >>>> -        *ecx = ((topo.num_nodes - 1) << 8) | (cpu->socket_id << (shift 
> >>>> + 1)) |
> >>>> -                topo.node_id;
> >>>> +        *ecx = (nodes << 8) | (topo_ids.pkg_id << (shift + 1)) |
> >>>> +               topo_ids.node_id;
> >>>>      }
> >>>>      *edx = 0;
> >>>>  }
> >>>> @@ -5318,6 +5226,7 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t 
> >>>> index, uint32_t count,
> >>>>      uint32_t signature[3];
> >>>>      X86CPUTopoInfo topo_info;
> >>>>  
> >>>> +    topo_info.nodes_per_pkg = env->nr_nodes;
> >>>>      topo_info.dies_per_pkg = env->nr_dies;
> >>>>      topo_info.cores_per_die = cs->nr_cores;
> >>>>      topo_info.threads_per_core = cs->nr_threads;
> >>>> @@ -5737,20 +5646,20 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t 
> >>>> index, uint32_t count,
> >>>>          }
> >>>>          switch (count) {
> >>>>          case 0: /* L1 dcache info */
> >>>> -            encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache, 
> >>>> cs,
> >>>> -                                       eax, ebx, ecx, edx);
> >>>> +            encode_cache_cpuid8000001d(env->cache_info_amd.l1d_cache,
> >>>> +                                       &topo_info, eax, ebx, ecx, edx);
> >>>>              break;
> >>>>          case 1: /* L1 icache info */
> >>>> -            encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache, 
> >>>> cs,
> >>>> -                                       eax, ebx, ecx, edx);
> >>>> +            encode_cache_cpuid8000001d(env->cache_info_amd.l1i_cache,
> >>>> +                                       &topo_info, eax, ebx, ecx, edx);
> >>>>              break;
> >>>>          case 2: /* L2 cache info */
> >>>> -            encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache, cs,
> >>>> -                                       eax, ebx, ecx, edx);
> >>>> +            encode_cache_cpuid8000001d(env->cache_info_amd.l2_cache,
> >>>> +                                       &topo_info, eax, ebx, ecx, edx);
> >>>>              break;
> >>>>          case 3: /* L3 cache info */
> >>>> -            encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache, cs,
> >>>> -                                       eax, ebx, ecx, edx);
> >>>> +            encode_cache_cpuid8000001d(env->cache_info_amd.l3_cache,
> >>>> +                                       &topo_info, eax, ebx, ecx, edx);
> >>>>              break;
> >>>>          default: /* end of info */
> >>>>              *eax = *ebx = *ecx = *edx = 0;
> >>>> @@ -5759,8 +5668,7 @@ void cpu_x86_cpuid(CPUX86State *env, uint32_t 
> >>>> index, uint32_t count,
> >>>>          break;
> >>>>      case 0x8000001E:
> >>>>          assert(cpu->core_id <= 255);
> >>>> -        encode_topo_cpuid8000001e(cs, cpu,
> >>>> -                                  eax, ebx, ecx, edx);
> >>>> +        encode_topo_cpuid8000001e(&topo_info, cpu, eax, ebx, ecx, edx);
> >>>>          break;
> >>>>      case 0xC0000000:
> >>>>          *eax = env->cpuid_xlevel2;
> >>>>    
> >>>     
> >>  
> >   
> 




reply via email to

[Prev in Thread] Current Thread [Next in Thread]