Categorize failing benchmarks

This commit is contained in:
justin 2016-04-06 09:53:54 -04:00
parent f1f9153578
commit c6e2ef2b9d

15
TODO
View file

@ -3,9 +3,22 @@ Initiatives:
Tier 1: Tier 1:
- Performance - most likely need CPS optimization and closure elimination - Performance - most likely need CPS optimization and closure elimination
- pass larceny benchmarks (will require some bugfixes): - pass larceny benchmarks (will require some bugfixes):
TODO: keep track of which ones will not run, and why
benchmarks that do not finish due to
missing features:
- gcbench.scm - Needed to manually import (srfi 9)
- mbrotZ.scm - No complex library
- read0.scm - needs open-input-string
- scheme.scm - Problem with (macro?) being a primitive and defined in application code
- slatex.scm - Need to handle more than one "args" argument to apply
crashes:
- compiler.scm - segfaults during compilation
- maze.scm - vector-ref invalid index -6
- nucleic.scm - Runs now, but segfaults
- simplex.scm - Compiles but quickly throws error Error: Divide by zero
OK if test errors out if outputs directory is missing, chibi does the same thing. OK if test errors out if outputs directory is missing, chibi does the same thing.
List of all benchmarks, along with comments if they are failing:
- ack.scm - ack.scm
- array1.scm - array1.scm