coreboot-kgpe-d16/util/board_status/to-wiki
Patrick Georgi 2a4f58ac12 board-status: Reorder the table categories
Show laptops and servers before desktop boards since that's where both
the market and coreboot are the most active these days.

Change-Id: I7de63975f3f2ff5e983b19e07558175a58870a1b
Signed-off-by: Patrick Georgi <pgeorgi@chromium.org>
Reviewed-on: http://review.coreboot.org/12292
Tested-by: build bot (Jenkins)
Reviewed-by: Paul Menzel <paulepanter@users.sourceforge.net>
Reviewed-by: Martin Roth <martinroth@google.com>
2015-11-03 21:29:16 +01:00
..
README
bucketize.sh
foreword.wiki board-status: Update the foreword 2015-11-03 21:27:03 +01:00
push-to-wiki.sh
status-to-wiki.sh
towiki.sh board-status: Reorder the table categories 2015-11-03 21:29:16 +01:00

README

Scripts to publish board-status data to the wiki
================================================

These scripts parse the board-status repository (and the coreboot repository as companion)
to build a meaningful representation of the test coverage stored in board-status.

The server runs these nightly (CET/CEST), so no user interaction with the wiki page is needed.

How to use
----------
When modifying the scripts, or when publishing the results elsewhere, you might want to run them
yourself. You'll need the board-status and the coreboot repository checked out side by side, named
"board-status" and "coreboot" respectively (in particular without .git suffix).

To emit wiki-text, in the board-status repository's top-level directory, run

    $ ../util/board_status/to-wiki/status-to-wiki.sh

The output ends up on stdout, so you'll have to store it yourself, if you need it later.

`push-to-wiki.sh FILENAME` can be used to push a file into the wiki. The page name is hard coded
in TITLE, while user credentials are looked up in ~/.wikiaccount, which should look like

    USERNAME=user
    USERPASS=password

How it works
------------
status-to-wiki collects the reports and sorts them in buckets by report date. These can have
weekly, monthly and quarterly granularity.
It then passes these into the towiki script, which reads the data in more details and prints
them in the output format.

Contributions
-------------
These scripts are rather bare, and you're welcome to extend them to extract more useful data
from both repositories, and to present the data in a nicer way.
A rewrite into another (reasonable) language is fine, too - shell quickly finds its limits
for this kind of text processing.