a697c19640
Previously privilege drop was happening "too early" and that caused some PMC IPC programming (performed in FSP) to fail because sideband was already locked out. This change set moves privilege drop to later stage, after last FSP notify call. BRANCH=reef BUG=chrome-os-partner:60657 TEST=iotools rdmsr X 0x121, make sure they can't be read. Also dmesg|grep -i IPC to make sure there are no errors related Change-Id: Ia3a774aee5fbf92805a5c69093bfbd3d7682c3a7 Signed-off-by: Andrey Petrov <andrey.petrov@intel.com> Reviewed-on: https://review.coreboot.org/17769 Tested-by: build bot (Jenkins) Reviewed-by: Aaron Durbin <adurbin@chromium.org>
211 lines
6 KiB
C
211 lines
6 KiB
C
/*
|
|
* This file is part of the coreboot project.
|
|
*
|
|
* Copyright (C) 2015-2016 Intel Corp.
|
|
* (Written by Andrey Petrov <andrey.petrov@intel.com> for Intel Corp.)
|
|
* (Written by Alexandru Gagniuc <alexandrux.gagniuc@intel.com> for Intel Corp.)
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*/
|
|
|
|
#include <console/console.h>
|
|
#include <cpu/cpu.h>
|
|
#include <cpu/x86/cache.h>
|
|
#include <cpu/x86/mp.h>
|
|
#include <cpu/intel/microcode.h>
|
|
#include <cpu/x86/msr.h>
|
|
#include <cpu/x86/mtrr.h>
|
|
#include <device/device.h>
|
|
#include <device/pci.h>
|
|
#include <reg_script.h>
|
|
#include <soc/cpu.h>
|
|
#include <soc/iomap.h>
|
|
#include <soc/pm.h>
|
|
#include <soc/smm.h>
|
|
#include <cpu/intel/turbo.h>
|
|
|
|
static const struct reg_script core_msr_script[] = {
|
|
/* Enable C-state and IO/MWAIT redirect */
|
|
REG_MSR_WRITE(MSR_PMG_CST_CONFIG_CONTROL,
|
|
(PKG_C_STATE_LIMIT_C2_MASK | CORE_C_STATE_LIMIT_C10_MASK
|
|
| IO_MWAIT_REDIRECT_MASK | CST_CFG_LOCK_MASK)),
|
|
/* Power Management I/O base address for I/O trapping to C-states */
|
|
REG_MSR_WRITE(MSR_PMG_IO_CAPTURE_BASE,
|
|
(ACPI_PMIO_CST_REG | (PMG_IO_BASE_CST_RNG_BLK_SIZE << 16))),
|
|
/* Disable C1E */
|
|
REG_MSR_RMW(MSR_POWER_CTL, ~0x2, 0),
|
|
/* Disable support for MONITOR and MWAIT instructions */
|
|
REG_MSR_RMW(MSR_IA32_MISC_ENABLES, ~MONITOR_MWAIT_DIS_MASK, 0),
|
|
/*
|
|
* Enable and Lock the Advanced Encryption Standard (AES-NI)
|
|
* feature register
|
|
*/
|
|
REG_MSR_RMW(MSR_FEATURE_CONFIG, ~FEATURE_CONFIG_RESERVED_MASK,
|
|
FEATURE_CONFIG_LOCK),
|
|
REG_SCRIPT_END
|
|
};
|
|
|
|
void enable_untrusted_mode(void)
|
|
{
|
|
msr_t msr = rdmsr(MSR_POWER_MISC);
|
|
msr.lo |= ENABLE_IA_UNTRUSTED;
|
|
wrmsr(MSR_POWER_MISC, msr);
|
|
}
|
|
|
|
static void soc_core_init(device_t cpu)
|
|
{
|
|
/* Set core MSRs */
|
|
reg_script_run(core_msr_script);
|
|
/*
|
|
* Enable ACPI PM timer emulation, which also lets microcode know
|
|
* location of ACPI_PMIO_BASE. This also enables other features
|
|
* implemented in microcode.
|
|
*/
|
|
enable_pm_timer_emulation();
|
|
}
|
|
|
|
static struct device_operations cpu_dev_ops = {
|
|
.init = soc_core_init,
|
|
};
|
|
|
|
static struct cpu_device_id cpu_table[] = {
|
|
{ X86_VENDOR_INTEL, CPUID_APOLLOLAKE_A0 },
|
|
{ X86_VENDOR_INTEL, CPUID_APOLLOLAKE_B0 },
|
|
{ 0, 0 },
|
|
};
|
|
|
|
static const struct cpu_driver driver __cpu_driver = {
|
|
.ops = &cpu_dev_ops,
|
|
.id_table = cpu_table,
|
|
};
|
|
|
|
/*
|
|
* MP and SMM loading initialization.
|
|
*/
|
|
struct smm_relocation_attrs {
|
|
uint32_t smbase;
|
|
uint32_t smrr_base;
|
|
uint32_t smrr_mask;
|
|
};
|
|
|
|
static struct smm_relocation_attrs relo_attrs;
|
|
|
|
static void read_cpu_topology(unsigned int *num_phys, unsigned int *num_virt)
|
|
{
|
|
msr_t msr;
|
|
msr = rdmsr(MSR_CORE_THREAD_COUNT);
|
|
*num_virt = (msr.lo >> 0) & 0xffff;
|
|
*num_phys = (msr.lo >> 16) & 0xffff;
|
|
}
|
|
|
|
/*
|
|
* Do essential initialization tasks before APs can be fired up
|
|
*
|
|
* 1. Prevent race condition in MTRR solution. Enable MTRRs on the BSP. This
|
|
* creates the MTRR solution that the APs will use. Otherwise APs will try to
|
|
* apply the incomplete solution as the BSP is calculating it.
|
|
*/
|
|
static void pre_mp_init(void)
|
|
{
|
|
x86_setup_mtrrs_with_detect();
|
|
x86_mtrr_check();
|
|
|
|
/* Make sure BSP is using the microcode from cbfs */
|
|
intel_update_microcode_from_cbfs();
|
|
}
|
|
|
|
/* Find CPU topology */
|
|
static int get_cpu_count(void)
|
|
{
|
|
unsigned int num_virt_cores, num_phys_cores;
|
|
|
|
read_cpu_topology(&num_phys_cores, &num_virt_cores);
|
|
|
|
printk(BIOS_DEBUG, "Detected %u core, %u thread CPU.\n",
|
|
num_phys_cores, num_virt_cores);
|
|
|
|
return num_virt_cores;
|
|
}
|
|
|
|
static void get_microcode_info(const void **microcode, int *parallel)
|
|
{
|
|
*microcode = intel_microcode_find();
|
|
*parallel = 1;
|
|
}
|
|
|
|
static void get_smm_info(uintptr_t *perm_smbase, size_t *perm_smsize,
|
|
size_t *smm_save_state_size)
|
|
{
|
|
void *smm_base;
|
|
size_t smm_size;
|
|
void *handler_base;
|
|
size_t handler_size;
|
|
|
|
/* All range registers are aligned to 4KiB */
|
|
const uint32_t rmask = ~((1 << 12) - 1);
|
|
|
|
/* Initialize global tracking state. */
|
|
smm_region(&smm_base, &smm_size);
|
|
smm_subregion(SMM_SUBREGION_HANDLER, &handler_base, &handler_size);
|
|
|
|
relo_attrs.smbase = (uint32_t)smm_base;
|
|
relo_attrs.smrr_base = relo_attrs.smbase | MTRR_TYPE_WRBACK;
|
|
relo_attrs.smrr_mask = ~(smm_size - 1) & rmask;
|
|
relo_attrs.smrr_mask |= MTRR_PHYS_MASK_VALID;
|
|
|
|
*perm_smbase = (uintptr_t)handler_base;
|
|
*perm_smsize = handler_size;
|
|
*smm_save_state_size = sizeof(em64t100_smm_state_save_area_t);
|
|
}
|
|
|
|
static void relocation_handler(int cpu, uintptr_t curr_smbase,
|
|
uintptr_t staggered_smbase)
|
|
{
|
|
msr_t smrr;
|
|
em64t100_smm_state_save_area_t *smm_state;
|
|
/* Set up SMRR. */
|
|
smrr.lo = relo_attrs.smrr_base;
|
|
smrr.hi = 0;
|
|
wrmsr(SMRR_PHYS_BASE, smrr);
|
|
smrr.lo = relo_attrs.smrr_mask;
|
|
smrr.hi = 0;
|
|
wrmsr(SMRR_PHYS_MASK, smrr);
|
|
smm_state = (void *)(SMM_EM64T100_SAVE_STATE_OFFSET + curr_smbase);
|
|
smm_state->smbase = staggered_smbase;
|
|
}
|
|
/*
|
|
* CPU initialization recipe
|
|
*
|
|
* Note that no microcode update is passed to the init function. CSE updates
|
|
* the microcode on all cores before releasing them from reset. That means that
|
|
* the BSP and all APs will come up with the same microcode revision.
|
|
*/
|
|
static const struct mp_ops mp_ops = {
|
|
.pre_mp_init = pre_mp_init,
|
|
.get_cpu_count = get_cpu_count,
|
|
.get_smm_info = get_smm_info,
|
|
.get_microcode_info = get_microcode_info,
|
|
.pre_mp_smm_init = southbridge_smm_clear_state,
|
|
.relocation_handler = relocation_handler,
|
|
.post_mp_init = southbridge_smm_enable_smi,
|
|
};
|
|
|
|
void apollolake_init_cpus(device_t dev)
|
|
{
|
|
/* Clear for take-off */
|
|
if (mp_init_with_smm(dev->link_list, &mp_ops) < 0)
|
|
printk(BIOS_ERR, "MP initialization failure.\n");
|
|
|
|
/* Temporarily cache the memory-mapped boot media. */
|
|
if (IS_ENABLED(CONFIG_BOOT_DEVICE_MEMORY_MAPPED))
|
|
mtrr_use_temp_range(-CONFIG_ROM_SIZE, CONFIG_ROM_SIZE,
|
|
MTRR_TYPE_WRPROT);
|
|
}
|