coreboot-kgpe-d16/util/docker/coreboot.org-status/board-status.html
Nico Huber 4829af17e3 cpu/intel: Rename socket_mFCPGA478 to socket_m
The name was wrong. mFCPGA478 is actually a pseudonym for mPGA478MN,
the successor of the socket that was meant.

The official name of this socket is mPGA478MT. But "Socket M" is much
easier to distinguish.

Change-Id: I4efeaca69acddfcdc5e957b0b521544314d46eeb
Signed-off-by: Nico Huber <nico.huber@secunet.com>
Reviewed-on: https://review.coreboot.org/c/31642
Tested-by: build bot (Jenkins) <no-reply@coreboot.org>
Reviewed-by: Arthur Heymans <arthur@aheymans.xyz>
Reviewed-by: Paul Menzel <paulepanter@users.sourceforge.net>
2019-02-28 13:47:54 +00:00
..
README
bucketize.sh
foreword.html board-status.html: improve instructions 2018-11-07 21:42:34 +00:00
status-to-html.sh
tohtml.sh cpu/intel: Rename socket_mFCPGA478 to socket_m 2019-02-28 13:47:54 +00:00

README

Scripts to publish board-status data to the wiki
================================================

These scripts parse the board-status repository (and the coreboot repository as companion)
to build a meaningful representation of the test coverage stored in board-status.

The server runs these nightly (CET/CEST), so no user interaction with the wiki page is needed.

How to use
----------
When modifying the scripts, or when publishing the results elsewhere, you might want to run them
yourself. You'll need the board-status and the coreboot repository checked out side by side, named
"board-status" and "coreboot" respectively (in particular without .git suffix).

To emit wiki-text, in the board-status repository's top-level directory, run

    $ ../util/board_status/to-wiki/status-to-wiki.sh

The output ends up on stdout, so you'll have to store it yourself, if you need it later.

`push-to-wiki.sh FILENAME TITLE` can be used to push a file into the wiki.
User credentials are looked up in ~/.wikiaccount, which should look like

    USERNAME=user
    USERPASS=password

How it works
------------
status-to-wiki collects the reports and sorts them in buckets by report date. These can have
weekly, monthly and quarterly granularity.
It then passes these into the towiki script, which reads the data in more details and prints
them in the output format.

Contributions
-------------
These scripts are rather bare, and you're welcome to extend them to extract more useful data
from both repositories, and to present the data in a nicer way.
A rewrite into another (reasonable) language is fine, too - shell quickly finds its limits
for this kind of text processing.