2020-04-02 23:48:19 +02:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0-only */
|
2017-09-02 15:41:43 +02:00
|
|
|
|
2019-08-16 19:34:25 +02:00
|
|
|
#include <arch/romstage.h>
|
2017-09-02 15:41:43 +02:00
|
|
|
#include <cbmem.h>
|
|
|
|
#include <console/console.h>
|
2019-06-21 07:20:12 +02:00
|
|
|
#include <commonlib/helpers.h>
|
2017-09-02 15:41:43 +02:00
|
|
|
#include <cpu/amd/mtrr.h>
|
|
|
|
#include <cpu/cpu.h>
|
|
|
|
#include <cpu/x86/msr.h>
|
|
|
|
#include <cpu/x86/mtrr.h>
|
|
|
|
#include <northbridge/amd/agesa/agesa_helper.h>
|
|
|
|
|
|
|
|
static void set_range_uc(u32 base, u32 size)
|
|
|
|
{
|
|
|
|
int i, max_var_mtrrs;
|
|
|
|
msr_t msr;
|
|
|
|
msr = rdmsr(MTRR_CAP_MSR);
|
|
|
|
max_var_mtrrs = msr.lo & MTRR_CAP_VCNT;
|
|
|
|
|
|
|
|
for (i = 0; i < max_var_mtrrs; i++) {
|
|
|
|
msr = rdmsr(MTRR_PHYS_MASK(i));
|
|
|
|
if (!(msr.lo & MTRR_PHYS_MASK_VALID))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (i == max_var_mtrrs)
|
|
|
|
die("Run out of unused MTRRs\n");
|
|
|
|
|
|
|
|
msr.hi = 0;
|
|
|
|
msr.lo = base | MTRR_TYPE_UNCACHEABLE;
|
|
|
|
wrmsr(MTRR_PHYS_BASE(i), msr);
|
|
|
|
|
|
|
|
msr.hi = (1 << (cpu_phys_address_size() - 32)) - 1;
|
|
|
|
msr.lo = ~(size - 1) | MTRR_PHYS_MASK_VALID;
|
|
|
|
wrmsr(MTRR_PHYS_MASK(i), msr);
|
|
|
|
}
|
|
|
|
|
|
|
|
void fixup_cbmem_to_UC(int s3resume)
|
|
|
|
{
|
|
|
|
if (s3resume)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* For normal path, INIT_POST has returned with all
|
|
|
|
* memory set WB cacheable. But we need CBMEM as UC
|
|
|
|
* to make CAR teardown with invalidation without
|
|
|
|
* writeback possible.
|
|
|
|
*/
|
|
|
|
|
|
|
|
uintptr_t top_of_ram = (uintptr_t) cbmem_top();
|
|
|
|
top_of_ram = ALIGN_UP(top_of_ram, 4 * MiB);
|
|
|
|
|
|
|
|
set_range_uc(top_of_ram - 4 * MiB, 4 * MiB);
|
|
|
|
set_range_uc(top_of_ram - 8 * MiB, 4 * MiB);
|
|
|
|
}
|
|
|
|
|
|
|
|
void recover_postcar_frame(struct postcar_frame *pcf, int s3resume)
|
|
|
|
{
|
|
|
|
msr_t base, mask;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
/* Replicate non-UC MTRRs as left behind by AGESA.
|
|
|
|
*/
|
2020-05-29 05:34:43 +02:00
|
|
|
for (i = 0; i < pcf->ctx.max_var_mtrrs; i++) {
|
2017-09-02 15:41:43 +02:00
|
|
|
mask = rdmsr(MTRR_PHYS_MASK(i));
|
|
|
|
base = rdmsr(MTRR_PHYS_BASE(i));
|
|
|
|
u32 size = ~(mask.lo & ~0xfff) + 1;
|
|
|
|
u8 type = base.lo & 0x7;
|
|
|
|
base.lo &= ~0xfff;
|
|
|
|
|
|
|
|
if (!(mask.lo & MTRR_PHYS_MASK_VALID) ||
|
|
|
|
(type == MTRR_TYPE_UNCACHEABLE))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
postcar_frame_add_mtrr(pcf, base.lo, size, type);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* For S3 resume path, INIT_RESUME does not return with
|
|
|
|
* memory covering CBMEM set as WB cacheable. For better
|
|
|
|
* speed make them WB after CAR teardown.
|
|
|
|
*/
|
|
|
|
if (s3resume) {
|
|
|
|
uintptr_t top_of_ram = (uintptr_t) cbmem_top();
|
|
|
|
top_of_ram = ALIGN_DOWN(top_of_ram, 4*MiB);
|
|
|
|
|
|
|
|
postcar_frame_add_mtrr(pcf, top_of_ram - 4*MiB, 4*MiB,
|
|
|
|
MTRR_TYPE_WRBACK);
|
|
|
|
postcar_frame_add_mtrr(pcf, top_of_ram - 8*MiB, 4*MiB,
|
|
|
|
MTRR_TYPE_WRBACK);
|
|
|
|
}
|
|
|
|
}
|