FrontPage 

Fuego 1.0 wiki

Login or create account

Fuego release 1.0.9-notes in 'raw' format

{{TableOfContents}}
Here is a list of problems encountered during the 1.0.9 release testing:
= Problems =
== general ==
 * /fuego-core/engine/scripts/ftc-unit-test.sh
   * can't download clitest
   * uses /userdata instead of /fuego-ro for accessing board stuff
 * had to create testplan_fuego (for fuego-specific tests)

== Functional.fuego_board_check ==
 * Jenkins jobs have the wrong TESTNAME
    * e.g. docker.testplan_fuego.Functional.fuego_board_check has 
"TESTNAME=fuego_board_check" instead of 'fuego_test')
      * fuego-create-jobs doesn't read test.yaml file
    * see if fuego-create-jobs uses the test name for the base script name
 * (worked around) can't find test with default plan in spec list
   * code is missing my patches for ignoring 'testplan_default'
   * you get an error from ovgen.py
 * docker board is missing reboot, according to Functional.fuego_board_check
   * scan_form_items.sh reports HAS_REBOOT=0
     * but docker has /sbin/reboot
   * test is run as user 'jenkins' which doesn't have /sbin in the path
   * Functional.fuego_board_check is correct!

== Benchmark.fuego_check_plots ==
 * (fixed) parser.py gets an exception
   * it's still using FUEGO_PARSER_PATH, instead of FUEGO_CORE environment var.
 * plot.png is empty
   * missing metrics.json file
   * multiplot mapping requires a specific metric name patterns (leading item
     of same name, to chart the data in a single chart?
     * plotting is VERY fragile!!
       * no one can debug this stuff
       * matplotlib producing empty graphs is hard to debug
 * FIXTHIS - should make plot API much easier, or at least document the rules better

= priorities for 1.1 release =
{{{#!Table
show_edit_links=0
sortby=priority
||item||priority||notes||
||testing by Daniel||high||not available until next week||
||testing by Jan-Simon||high||not available until next week||
||put post_test process kill in test_cleanup||high||need optional test_cleanup, that calls kill_procs, that calls ov_rootfs_kill||
||handle set -e differently||medium||
||use spec name instead of plan name in job name||medium||Use <board>.<spec>.<testname>||
||handle timeouts automatically||medium||should be in test and board, not in plan - but depends on spec, transport, build time etc||
||support separate build phase||defer||don't want to hold board reserved during build, for lava integration and handling timeouts better||
||support phase reordering||defer||||
||fix ftc device dictionary||defer||used for board check and pre-populating options||
||eliminate DISTRIB||medium||automatically detect, and use board vars||
||support ftc run-test||low||BUILD_TIMESTAMP won't work with separate post_test, but we're not doing that||
||ensure logs are linked even for failed tests||high||if log exists, there should be link, failure or not||
||have test do description-setting||medium||||
||ov_transport_connect/ov_transport_disconnect||medium||Sounds like Jan-Simon will do it||
||web file transport||defer||could be useful for some interesting target cases||
||rename all scripts to fuego_test.sh||low||||
||add nod and job scripts to ftc||high||||
||automatically create a view for each board||medium||use a command line option on fuego-create-node (--make_view)||
||move /fuego-ro/conf/boards to /fuego-ro/boards||medium||||
}}}

= post_test work =
List of functions that have post_test arguments:
 * Benchmark.netpipe: kills iperf (should be NPtcp)
    * netpipe test doesn't use benchmark.sh!! (there is no call to post_test)
 * Functional.LTP.Open_Posix: kills run-posix-option-group-tst.sh run-tests.sh run-test
    * this test is not in the 'next' branch
 * Functional.netperf: kills netperf
    * netperf doesn't use functional.sh!! (there is no call to post_test)
 * Functional.LTP.Filesystem: kills run-test
    * this test is not in the 'next' branch
 * (OK) Benchmark.fio: fills fio
 * Benchmark.OpenSSL: kills openssl
    * Openssl doesn't use benchmark.sh (there is no call to post_test)
 * (OK) Benchmark.lmbench2: kills lmbench lat_mem_rd par_mem
 * (OK) Benchmark.IOzone: kills iozone
 * (OK) Benchmark.iperf: kills iperf
 * Benchmark.Java: kills java
    * put in test_cleanup, but didn't call kill_procs
    * there could be other java programs running - kill string is too generic
 * (OK) Benchmark.Interbench: kills interbench
 * (OK) Benchmark.gtkperf: kills gtkperf

=== fix functions that need to call post_test ===
 * Benchmark.netpipe
 * Functional.netperf
 * Benchmark.OpenSSL









TBWiki engine 1.8.3 by Tim Bird