haswell/lynxpoint: Add native DMI init
Implement native DMI init for Haswell and Lynx Point. This is only needed on non-ULT platforms, and only when MRC.bin is not used. TEST=Verify DMI initialises correctly on Asrock B85M Pro4. Change-Id: I5fb1a2adc4ffbf0ebbf0d2d3a444055c53765faa Signed-off-by: Angel Pons <th3fanbus@gmail.com> Reviewed-on: https://review.coreboot.org/c/coreboot/+/64177 Tested-by: build bot (Jenkins) <no-reply@coreboot.org> Reviewed-by: Arthur Heymans <arthur@aheymans.xyz>
This commit is contained in:
parent
b739fd287d
commit
567ece44ea
|
@ -20,6 +20,7 @@ romstage-y += report_platform.c
|
|||
postcar-y += memmap.c
|
||||
|
||||
ifeq ($(CONFIG_USE_NATIVE_RAMINIT),y)
|
||||
romstage-y += early_dmi.c early_pcie.c vcu_mailbox.c
|
||||
subdirs-y += native_raminit
|
||||
|
||||
else
|
||||
|
|
|
@ -0,0 +1,96 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
|
||||
#include <console/console.h>
|
||||
#include <northbridge/intel/haswell/haswell.h>
|
||||
#include <southbridge/intel/lynxpoint/pch.h>
|
||||
#include <types.h>
|
||||
|
||||
static void dmi_print_link_status(int loglevel)
|
||||
{
|
||||
const uint16_t dmilsts = dmibar_read16(DMILSTS);
|
||||
printk(loglevel, "DMI: Running at Gen%u x%u\n", dmilsts & 0xf, dmilsts >> 4 & 0x1f);
|
||||
}
|
||||
|
||||
#define RETRAIN (1 << 5)
|
||||
|
||||
#define LTRN (1 << 11)
|
||||
|
||||
static void dmi_setup_physical_layer(void)
|
||||
{
|
||||
/* Program DMI AFE settings, which are needed for DMI to work */
|
||||
peg_dmi_recipe(false, 0);
|
||||
|
||||
/* Additional DMI programming steps */
|
||||
dmibar_setbits32(0x258, 1 << 29);
|
||||
dmibar_clrsetbits32(0x208, 0x7ff, 0x6b5);
|
||||
dmibar_clrsetbits32(0x22c, 0xffff, 0x2020);
|
||||
|
||||
/* Write SA reference code version */
|
||||
dmibar_write32(0x71c, 0x0000000f);
|
||||
dmibar_write32(0x720, 0x01060200);
|
||||
|
||||
/* We also have to bring up the PCH side of the DMI link */
|
||||
pch_dmi_setup_physical_layer();
|
||||
|
||||
/* Write-once settings */
|
||||
dmibar_clrsetbits32(DMILCAP, 0x3f00f, 2 << 0);
|
||||
|
||||
printk(BIOS_DEBUG, "Retraining DMI at Gen2 speeds...\n");
|
||||
dmi_print_link_status(BIOS_DEBUG);
|
||||
|
||||
/* Retrain link */
|
||||
dmibar_setbits16(DMILCTL, RETRAIN);
|
||||
do {} while (dmibar_read16(DMILSTS) & LTRN);
|
||||
dmi_print_link_status(BIOS_DEBUG);
|
||||
|
||||
/* Retrain link again for DMI Gen2 speeds */
|
||||
dmibar_setbits16(DMILCTL, RETRAIN);
|
||||
do {} while (dmibar_read16(DMILSTS) & LTRN);
|
||||
dmi_print_link_status(BIOS_INFO);
|
||||
}
|
||||
|
||||
#define VC_ACTIVE (1U << 31)
|
||||
|
||||
#define VCNEGPND (1 << 1)
|
||||
|
||||
#define DMI_VC_CFG(vcid, tcmap) (VC_ACTIVE | ((vcid) << 24) | (tcmap))
|
||||
|
||||
static void dmi_tc_vc_mapping(void)
|
||||
{
|
||||
printk(BIOS_DEBUG, "Programming SA DMI VC/TC mappings...\n");
|
||||
|
||||
if (CONFIG(INTEL_LYNXPOINT_LP))
|
||||
dmibar_setbits8(0xa78, 1 << 1);
|
||||
|
||||
/* Each TC is mapped to one and only one VC */
|
||||
const u32 vc0 = DMI_VC_CFG(0, (1 << 6) | (1 << 5) | (1 << 4) | (1 << 3) | (1 << 0));
|
||||
const u32 vc1 = DMI_VC_CFG(1, (1 << 1));
|
||||
const u32 vcp = DMI_VC_CFG(2, (1 << 2));
|
||||
const u32 vcm = DMI_VC_CFG(7, (1 << 7));
|
||||
dmibar_write32(DMIVC0RCTL, vc0);
|
||||
dmibar_write32(DMIVC1RCTL, vc1);
|
||||
dmibar_write32(DMIVCPRCTL, vcp);
|
||||
dmibar_write32(DMIVCMRCTL, vcm);
|
||||
|
||||
/* Set Extended VC Count (EVCC) to 1 if VC1 is active */
|
||||
dmibar_clrsetbits8(DMIPVCCAP1, 7, !!(vc1 & VC_ACTIVE));
|
||||
|
||||
/*
|
||||
* We also have to program the PCH side of the DMI link. Since both ends
|
||||
* must use the same Virtual Channel settings, we pass them as arguments.
|
||||
*/
|
||||
pch_dmi_tc_vc_mapping(vc0, vc1, vcp, vcm);
|
||||
|
||||
printk(BIOS_DEBUG, "Waiting for SA DMI VC negotiation... ");
|
||||
do {} while (dmibar_read16(DMIVC0RSTS) & VCNEGPND);
|
||||
do {} while (dmibar_read16(DMIVC1RSTS) & VCNEGPND);
|
||||
do {} while (dmibar_read16(DMIVCPRSTS) & VCNEGPND);
|
||||
do {} while (dmibar_read16(DMIVCMRSTS) & VCNEGPND);
|
||||
printk(BIOS_DEBUG, "done!\n");
|
||||
}
|
||||
|
||||
void dmi_early_init(void)
|
||||
{
|
||||
dmi_setup_physical_layer();
|
||||
dmi_tc_vc_mapping();
|
||||
}
|
|
@ -0,0 +1,121 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
|
||||
#include <console/console.h>
|
||||
#include <device/pci_def.h>
|
||||
#include <device/pci_mmio_cfg.h>
|
||||
#include <device/pci_ops.h>
|
||||
#include <northbridge/intel/haswell/haswell.h>
|
||||
#include <northbridge/intel/haswell/vcu_mailbox.h>
|
||||
#include <types.h>
|
||||
|
||||
#define PEG_DEV(func) PCI_DEV(0, 1, func)
|
||||
|
||||
#define MAX_PEG_FUNC 3
|
||||
|
||||
static void peg_dmi_unset_and_set_mask_pcicfg(
|
||||
volatile union pci_bank *const bank,
|
||||
const uint32_t offset,
|
||||
const uint32_t unset_mask,
|
||||
const uint32_t set_mask,
|
||||
const uint32_t shift,
|
||||
const bool valid)
|
||||
{
|
||||
if (!valid)
|
||||
return;
|
||||
|
||||
volatile uint32_t *const addr = &bank->reg32[offset / sizeof(uint32_t)];
|
||||
clrsetbits32(addr, unset_mask << shift, set_mask << shift);
|
||||
}
|
||||
|
||||
static void peg_dmi_unset_and_set_mask_common(
|
||||
const bool is_peg,
|
||||
const uint32_t offset,
|
||||
const uint32_t unset,
|
||||
const uint32_t set,
|
||||
const uint32_t shift,
|
||||
const bool valid)
|
||||
{
|
||||
const uint32_t unset_mask = unset << shift;
|
||||
const uint32_t set_mask = set << shift;
|
||||
if (is_peg) {
|
||||
for (uint8_t i = 0; i < MAX_PEG_FUNC; i++)
|
||||
pci_update_config32(PEG_DEV(i), offset, ~unset_mask, set_mask);
|
||||
} else {
|
||||
dmibar_clrsetbits32(offset, unset_mask, set_mask);
|
||||
}
|
||||
}
|
||||
|
||||
static void peg_dmi_unset_and_set_mask_vcu_mmio(
|
||||
const uint32_t addr,
|
||||
const uint32_t unset_mask,
|
||||
const uint32_t set_mask,
|
||||
const uint32_t shift,
|
||||
const bool valid)
|
||||
{
|
||||
if (!valid)
|
||||
return;
|
||||
|
||||
vcu_update_mmio(addr, ~(unset_mask << shift), set_mask << shift);
|
||||
}
|
||||
|
||||
#define BUNDLE_STEP 0x20
|
||||
|
||||
static void *const dmibar = (void *)(uintptr_t)CONFIG_FIXED_DMIBAR_MMIO_BASE;
|
||||
|
||||
void peg_dmi_recipe(const bool is_peg, const pci_devfn_t dev)
|
||||
{
|
||||
const bool always = true;
|
||||
const bool is_dmi = !is_peg;
|
||||
|
||||
/* Treat DMIBAR and PEG devices the same way */
|
||||
volatile union pci_bank *const bank = is_peg ? pci_map_bus(dev) : dmibar;
|
||||
|
||||
const size_t bundles = (is_peg ? 8 : 2) * BUNDLE_STEP;
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP) {
|
||||
/* These are actually per-lane */
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0xa00 + i, 0x1f, 0x0c, 0, always);
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0xa10 + i, 0x1f, 0x0c, 0, always);
|
||||
}
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x904 + i, 0x1f, 0x02, 0, is_peg);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x904 + i, 0x1f, 0x03, 5, is_peg);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x90c + i, 0x3f, 0x09, 5, always);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x90c + i, 0x0f, 0x05, 21, is_peg);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x910 + i, 0x0f, 0x08, 6, is_peg);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x910 + i, 0x0f, 0x00, 10, always);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x910 + i, 0x07, 0x00, 18, always);
|
||||
|
||||
peg_dmi_unset_and_set_mask_vcu_mmio(0x0c008001, 0x1f, 0x03, 25, is_peg);
|
||||
peg_dmi_unset_and_set_mask_vcu_mmio(0x0c0c8001, 0x3f, 0x00, 23, is_dmi);
|
||||
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0xc28, 0x1f, 0x13, 18, always);
|
||||
|
||||
peg_dmi_unset_and_set_mask_common(is_peg, 0xc38, 0x01, 0x00, 6, always);
|
||||
peg_dmi_unset_and_set_mask_common(is_peg, 0x260, 0x03, 0x02, 0, always);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x900 + i, 0x03, 0x00, 26, always);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x904 + i, 0x03, 0x03, 10, always);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x90c + i, 0x1f, 0x07, 25, is_peg);
|
||||
|
||||
for (size_t i = 0; i < bundles; i += BUNDLE_STEP)
|
||||
peg_dmi_unset_and_set_mask_pcicfg(bank, 0x91c + i, 0x07, 0x05, 27, is_peg);
|
||||
}
|
|
@ -34,6 +34,9 @@ void haswell_early_initialization(void);
|
|||
void haswell_late_initialization(void);
|
||||
void haswell_unhide_peg(void);
|
||||
|
||||
void dmi_early_init(void);
|
||||
void peg_dmi_recipe(const bool is_peg, const pci_devfn_t dev);
|
||||
|
||||
void report_platform_info(void);
|
||||
|
||||
struct acpi_rsdp;
|
||||
|
|
|
@ -1,7 +1,19 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
|
||||
#include <console/console.h>
|
||||
#include <northbridge/intel/haswell/haswell.h>
|
||||
#include <northbridge/intel/haswell/raminit.h>
|
||||
#include <types.h>
|
||||
|
||||
static bool early_init_native(int s3resume)
|
||||
{
|
||||
printk(BIOS_DEBUG, "Starting native platform initialisation\n");
|
||||
|
||||
if (!CONFIG(INTEL_LYNXPOINT_LP))
|
||||
dmi_early_init();
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
void perform_raminit(const int s3resume)
|
||||
{
|
||||
|
@ -9,6 +21,9 @@ void perform_raminit(const int s3resume)
|
|||
* See, this function's name is a lie. There are more things to
|
||||
* do that memory initialisation, but they are relatively easy.
|
||||
*/
|
||||
const bool cpu_replaced = early_init_native(s3resume);
|
||||
|
||||
(void)cpu_replaced;
|
||||
|
||||
/** TODO: Implement the required magic **/
|
||||
die("NATIVE RAMINIT: More Magic (tm) required.\n");
|
||||
|
|
|
@ -0,0 +1,155 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
|
||||
#include <assert.h>
|
||||
#include <console/console.h>
|
||||
#include <delay.h>
|
||||
#include <northbridge/intel/haswell/haswell.h>
|
||||
#include <northbridge/intel/haswell/vcu_mailbox.h>
|
||||
#include <types.h>
|
||||
|
||||
/*
|
||||
* This is a library for the VCU (Validation Control Unit) mailbox. This
|
||||
* mailbox is primarily used to adjust some magic PCIe tuning parameters.
|
||||
*
|
||||
* There are two revisions of the VCU mailbox. Rev1 is specific to Haswell
|
||||
* stepping A0, and all other steppings use Rev2. Haswell stepping A0 CPUs
|
||||
* are early Engineering Samples with undocumented errata, and most likely
|
||||
* need special microcode updates to boot. Thus, the code does not support
|
||||
* VCU mailbox Rev1, because no one should need it anymore.
|
||||
*/
|
||||
|
||||
#define VCU_MAILBOX_INTERFACE 0x6c00
|
||||
#define VCU_MAILBOX_DATA 0x6c04
|
||||
|
||||
#define VCU_RUN_BUSY (1U << 31)
|
||||
|
||||
enum vcu_opcode {
|
||||
VCU_INVALID_OPCODE = 0x00,
|
||||
VCU_OPCODE_READ_VCU_API_VER_ID = 0x01,
|
||||
VCU_OPCODE_OPEN_SEQ = 0x02,
|
||||
VCU_OPCODE_CLOSE_SEQ = 0x03,
|
||||
VCU_OPCODE_READ_DATA = 0x07,
|
||||
VCU_OPCODE_WRITE_DATA = 0x08,
|
||||
VCU_OPCODE_READ_CSR = 0x13,
|
||||
VCU_OPCODE_WRITE_CSR = 0x14,
|
||||
VCU_OPCODE_READ_MMIO = 0x15,
|
||||
VCU_OPCODE_WRITE_MMIO = 0x16,
|
||||
};
|
||||
|
||||
enum vcu_sequence {
|
||||
SEQ_ID_READ_CSR = 0x1,
|
||||
SEQ_ID_WRITE_CSR = 0x2,
|
||||
SEQ_ID_READ_MMIO = 0x3,
|
||||
SEQ_ID_WRITE_MMIO = 0x4,
|
||||
};
|
||||
|
||||
#define VCU_RESPONSE_MASK 0xffff
|
||||
#define VCU_RESPONSE_SUCCESS 0x40
|
||||
#define VCU_RESPONSE_BUSY 0x80
|
||||
#define VCU_RESPONSE_THREAD_UNAVAILABLE 0x82
|
||||
#define VCU_RESPONSE_ILLEGAL 0x90
|
||||
|
||||
/* FIXME: Use timer API */
|
||||
static void send_vcu_command(const enum vcu_opcode opcode, const uint32_t data)
|
||||
{
|
||||
if (opcode == VCU_INVALID_OPCODE)
|
||||
return;
|
||||
|
||||
for (unsigned int i = 0; i < 10; i++) {
|
||||
mchbar_write32(VCU_MAILBOX_DATA, data);
|
||||
mchbar_write32(VCU_MAILBOX_INTERFACE, opcode | VCU_RUN_BUSY);
|
||||
uint32_t vcu_interface;
|
||||
for (unsigned int j = 0; j < 100; j++) {
|
||||
vcu_interface = mchbar_read32(VCU_MAILBOX_INTERFACE);
|
||||
if (!(vcu_interface & VCU_RUN_BUSY))
|
||||
break;
|
||||
|
||||
udelay(10);
|
||||
}
|
||||
if (vcu_interface & VCU_RUN_BUSY)
|
||||
continue;
|
||||
|
||||
if ((vcu_interface & VCU_RESPONSE_MASK) == VCU_RESPONSE_SUCCESS)
|
||||
return;
|
||||
}
|
||||
printk(BIOS_ERR, "VCU: Failed to send command\n");
|
||||
}
|
||||
|
||||
static enum vcu_opcode get_register_opcode(enum vcu_sequence seq)
|
||||
{
|
||||
switch (seq) {
|
||||
case SEQ_ID_READ_CSR:
|
||||
return VCU_OPCODE_READ_CSR;
|
||||
case SEQ_ID_WRITE_CSR:
|
||||
return VCU_OPCODE_WRITE_CSR;
|
||||
case SEQ_ID_READ_MMIO:
|
||||
return VCU_OPCODE_READ_MMIO;
|
||||
case SEQ_ID_WRITE_MMIO:
|
||||
return VCU_OPCODE_WRITE_MMIO;
|
||||
default:
|
||||
BUG();
|
||||
return VCU_INVALID_OPCODE;
|
||||
}
|
||||
}
|
||||
|
||||
static enum vcu_opcode get_data_opcode(enum vcu_sequence seq)
|
||||
{
|
||||
switch (seq) {
|
||||
case SEQ_ID_READ_CSR:
|
||||
case SEQ_ID_READ_MMIO:
|
||||
return VCU_OPCODE_READ_DATA;
|
||||
case SEQ_ID_WRITE_CSR:
|
||||
case SEQ_ID_WRITE_MMIO:
|
||||
return VCU_OPCODE_WRITE_DATA;
|
||||
default:
|
||||
BUG();
|
||||
return VCU_INVALID_OPCODE;
|
||||
}
|
||||
}
|
||||
|
||||
static uint32_t send_vcu_sequence(uint32_t addr, enum vcu_sequence seq, uint32_t wr_data)
|
||||
{
|
||||
send_vcu_command(VCU_OPCODE_OPEN_SEQ, seq);
|
||||
|
||||
send_vcu_command(get_register_opcode(seq), addr);
|
||||
|
||||
send_vcu_command(get_data_opcode(seq), wr_data);
|
||||
|
||||
const uint32_t rd_data = mchbar_read32(VCU_MAILBOX_DATA);
|
||||
|
||||
send_vcu_command(VCU_OPCODE_CLOSE_SEQ, seq);
|
||||
|
||||
return rd_data;
|
||||
}
|
||||
|
||||
#define VCU_WRITE_IGNORED 0
|
||||
|
||||
uint32_t vcu_read_csr(uint32_t addr)
|
||||
{
|
||||
return send_vcu_sequence(addr, SEQ_ID_READ_CSR, VCU_WRITE_IGNORED);
|
||||
}
|
||||
|
||||
void vcu_write_csr(uint32_t addr, uint32_t data)
|
||||
{
|
||||
send_vcu_sequence(addr, SEQ_ID_WRITE_CSR, data);
|
||||
}
|
||||
|
||||
void vcu_update_csr(uint32_t addr, uint32_t andvalue, uint32_t orvalue)
|
||||
{
|
||||
vcu_write_csr(addr, (vcu_read_csr(addr) & andvalue) | orvalue);
|
||||
}
|
||||
|
||||
uint32_t vcu_read_mmio(uint32_t addr)
|
||||
{
|
||||
return send_vcu_sequence(addr, SEQ_ID_READ_MMIO, VCU_WRITE_IGNORED);
|
||||
}
|
||||
|
||||
void vcu_write_mmio(uint32_t addr, uint32_t data)
|
||||
{
|
||||
send_vcu_sequence(addr, SEQ_ID_WRITE_MMIO, data);
|
||||
}
|
||||
|
||||
void vcu_update_mmio(uint32_t addr, uint32_t andvalue, uint32_t orvalue)
|
||||
{
|
||||
vcu_write_mmio(addr, (vcu_read_mmio(addr) & andvalue) | orvalue);
|
||||
}
|
|
@ -0,0 +1,16 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
|
||||
#ifndef HASWELL_VCU_MAILBOX_H
|
||||
#define HASWELL_VCU_MAILBOX_H
|
||||
|
||||
#include <stdint.h>
|
||||
|
||||
uint32_t vcu_read_csr(uint32_t addr);
|
||||
void vcu_write_csr(uint32_t addr, uint32_t data);
|
||||
void vcu_update_csr(uint32_t addr, uint32_t andvalue, uint32_t orvalue);
|
||||
|
||||
uint32_t vcu_read_mmio(uint32_t addr);
|
||||
void vcu_write_mmio(uint32_t addr, uint32_t data);
|
||||
void vcu_update_mmio(uint32_t addr, uint32_t andvalue, uint32_t orvalue);
|
||||
|
||||
#endif /* HASWELL_VCU_MAILBOX_H */
|
|
@ -35,6 +35,8 @@ bootblock-y += early_pch.c
|
|||
romstage-y += early_usb.c early_me.c me_status.c early_pch.c
|
||||
romstage-y += pmutil.c
|
||||
|
||||
romstage-$(CONFIG_USE_NATIVE_RAMINIT) += early_pch_native.c
|
||||
|
||||
ifeq ($(CONFIG_INTEL_LYNXPOINT_LP),y)
|
||||
romstage-y += lp_gpio.c
|
||||
ramstage-y += lp_gpio.c
|
||||
|
|
|
@ -0,0 +1,52 @@
|
|||
/* SPDX-License-Identifier: GPL-2.0-or-later */
|
||||
|
||||
#include <console/console.h>
|
||||
#include <device/pci_ops.h>
|
||||
#include <southbridge/intel/lynxpoint/pch.h>
|
||||
#include <types.h>
|
||||
|
||||
void pch_dmi_setup_physical_layer(void)
|
||||
{
|
||||
/** FIXME: We need to make sure the SA supports Gen2 as well **/
|
||||
if ((RCBA32(0x21a4) & 0x0f) == 0x02) {
|
||||
/* Set Gen 2 Common Clock N_FTS */
|
||||
RCBA32_AND_OR(0x2340, ~0x00ff0000, 0x3a << 16);
|
||||
|
||||
/* Set Target Link Speed to DMI Gen2 */
|
||||
RCBA8_AND_OR(DLCTL2, ~0x07, 0x02);
|
||||
}
|
||||
}
|
||||
|
||||
#define VC_ACTIVE (1U << 31)
|
||||
|
||||
#define VCNEGPND (1 << 1)
|
||||
|
||||
void pch_dmi_tc_vc_mapping(const u32 vc0, const u32 vc1, const u32 vcp, const u32 vcm)
|
||||
{
|
||||
printk(BIOS_DEBUG, "Programming PCH DMI VC/TC mappings...\n");
|
||||
|
||||
RCBA32_AND_OR(CIR0050, ~(0xf << 20), 2 << 20);
|
||||
if (vcp & VC_ACTIVE)
|
||||
RCBA32_OR(CIR0050, 1 << 19 | 1 << 17);
|
||||
|
||||
RCBA32(CIR0050); /* Ensure posted write hits */
|
||||
|
||||
/* Use the same virtual channel mapping on both ends of the DMI link */
|
||||
RCBA32(V0CTL) = vc0;
|
||||
RCBA32(V1CTL) = vc1;
|
||||
RCBA32(V1CTL); /* Ensure posted write hits */
|
||||
RCBA32(VPCTL) = vcp;
|
||||
RCBA32(VPCTL); /* Ensure posted write hits */
|
||||
RCBA32(VMCTL) = vcm;
|
||||
|
||||
/* Lock the registers */
|
||||
RCBA32_OR(CIR0050, 1U << 31);
|
||||
RCBA32(CIR0050); /* Ensure posted write hits */
|
||||
|
||||
printk(BIOS_DEBUG, "Waiting for PCH DMI VC negotiation... ");
|
||||
do {} while (RCBA16(V0STS) & VCNEGPND);
|
||||
do {} while (RCBA16(V1STS) & VCNEGPND);
|
||||
do {} while (RCBA16(VPSTS) & VCNEGPND);
|
||||
do {} while (RCBA16(VMSTS) & VCNEGPND);
|
||||
printk(BIOS_DEBUG, "done!\n");
|
||||
}
|
|
@ -112,6 +112,9 @@ enum pch_platform_type {
|
|||
PCH_TYPE_ULT = 5,
|
||||
};
|
||||
|
||||
void pch_dmi_setup_physical_layer(void);
|
||||
void pch_dmi_tc_vc_mapping(u32 vc0, u32 vc1, u32 vcp, u32 vcm);
|
||||
|
||||
void usb_ehci_sleep_prepare(pci_devfn_t dev, u8 slp_typ);
|
||||
void usb_ehci_disable(pci_devfn_t dev);
|
||||
void usb_xhci_sleep_prepare(pci_devfn_t dev, u8 slp_typ);
|
||||
|
@ -405,9 +408,10 @@ void mainboard_config_rcba(void);
|
|||
|
||||
/* Southbridge IO BARs */
|
||||
|
||||
#define PMBASE 0x40
|
||||
#define GPIOBASE 0x48
|
||||
|
||||
#define PMBASE 0x40
|
||||
#define CIR0050 0x0050 /* 32bit */
|
||||
|
||||
#define RPC 0x0400 /* 32bit */
|
||||
#define RPFN 0x0404 /* 32bit */
|
||||
|
@ -430,6 +434,20 @@ void mainboard_config_rcba(void);
|
|||
#define IOTR2 0x1e90 /* 64bit */
|
||||
#define IOTR3 0x1e98 /* 64bit */
|
||||
|
||||
#define V0CTL 0x2014 /* 32bit */
|
||||
#define V0STS 0x201a /* 16bit */
|
||||
|
||||
#define V1CTL 0x2020 /* 32bit */
|
||||
#define V1STS 0x2026 /* 16bit */
|
||||
|
||||
#define VPCTL 0x2030 /* 32bit */
|
||||
#define VPSTS 0x2038 /* 16bit */
|
||||
|
||||
#define VMCTL 0x2040 /* 32bit */
|
||||
#define VMSTS 0x2048 /* 16bit */
|
||||
|
||||
#define DLCTL2 0x21b0
|
||||
|
||||
#define TCTL 0x3000 /* 8bit */
|
||||
|
||||
#define NOINT 0
|
||||
|
|
Loading…
Reference in New Issue