IsaSAT Releases
SAT Competition 2023
This year, once again, I submitted IsaSAT to the SAT Competition 2023 (the source code is as always available). According to the results, I did not finish last. However, compared to other solvers, I am trailing in the unsatisfiable instances. I believe that this is due to the fact that I still do not have bounded variable elimination. In the past, I assumed that vivification was also very important, but its importance is now less clear after Daniela and I attempted to examine it more precisely for POS'23 on the SAT solver CaDiCaL.
Something special happened this year: it turned out that my DRAT proofs are broken because the proof format is not strong enough to express what I want. I was aware of this problem before the competition (as it already existed last year).
My solver has two sets of clauses: the irredundant set N and the redundant set U, with the strict invariant N⊨U. To handle pure literals, I simply check for literals that appear only in one direction in N (all positive or all negative) and set the corresponding unit. For example, if N is A∨B, A∨¬B, B∨C, IsaSAT would first learn the unit clause A, delete all clauses containing A, and then learn the unit clauses B and C. This results in an empty problem.
This is legal (and is part of my proofs!). Technically, adding the units is a DRAT clause… but only if the negated literal does not appear in U. Given my strict invariants, I don’t care in IsaSAT. However, the DRAT checker has a different opinion on this because N and U cannot be distinguished (I have been told that VeriPB is able to do so, but did not look at it yet). Therefore, some of my proofs were rightfully rejected while IsaSAT did nothing incorrect and outputted the correct answer.
Other SAT solvers also face this issue, but they simply remove all the clauses containing a pure literal (both in N and U), making the transformation legal – but maybe a clause in U could have been useful for solving the problem.
Currently, there is no solution to this problem – execept that I only generate proofs for the SAT Competition, because it is mandatory in the main track, even though my SAT solver is entirely verified so the answer cannot be wrong.
Interestingly I found a heuristic bug after the submission, but it only helps on the SC2021 and SC2022 problems, not on the SC2023 problems. So this is a story for another time.
SAT Competition 2022 and v202201
The second release is now available. It is the version submitted to the SAT Competition 2022. This year I managed to make IsaSAT work on StarExec.
Compared to the version form the EDA Challenge:

It can now produces DRAT proofs. There are absolutely no guarantees on the correctness of those proofs… And in particular, I have managed to produce wrong proofs when I forgot to log unit clauses. This does not matter because the entire solver is verified, but it is mandatory to take part in the Main track.

I have implemented pure literal deletion. IsaSAT search for literals that do not appear or appear in only one polarity and sets them to true. It is the first transformation performed by IsaSAT that does not preserve models.
Shortly after the SAT Competition deadline, while drinking tea and looking at the HTML generated documentation, I found two heuristics bugs, which are solved in the release v202201. According to my benchmarks it solves 23 more problems on the benchmarks from the SAT Competition 2021:
solved  sat  uns  par2  

withproperscheduling  198  87  111  2273168 
sc2022  176  85  91  2457296 
(Better PDF version of the cactus plot). So, clearly a good fix. I will probably try to describe to the bug in more details, but in short: I was restarting way too little and reducing way too much. And I did not realise that because I swapped it when I printed it during the statistics.
EDAChallenge 2021
The first official release is available. It is the version submitted to the EDAChallenge 2021. To the surprise of everyone (including the author!), it actually won the challenge. The system description submitted to the EDA Challenge is available.
How good is the solver really? Another comparison is the wellknown SAT Competition that features hard instances with long running time. The comparison is unfair towards my SAT solver (because you need inprocessing nowadays to even have a chance for most instances), but the SAT Competition is the standard way to compare solvers.
Here is a CDF with kissat (who finished second) and the solver who finished last during the SAT Competition 2021 (in the plot, higher is better!):
(Better PDF version). The conclusion is that IsaSAT would clearly not have won the SAT Competition 2021, but also would probably not have been the slowest solver (if I had been able to make IsaSAT work on Starexec). It is not a great result, but it is an indication how far you can go without inprocessing.
Solved  

Solver  SAT+UNSAT  SAT  UNSAT  PAR2 
Kissat  283  137  146  1377523 
IsaSAT  137  69  68  2795128 
CleanMaple_PriPro  136  71  65  2868588 
The conclusion is threefold. First, it is possible to verify efficient code in a proof assistant. Second, there is no reason to believe that verified code is slower than handwritten one^{1} – even if it takes much more time to prove correctness. Third, the perfect SAT solver depends on the benchmarks you have.

I don’t believe that for verified code targetting Standard ML, however. Even if MLton is a very impressive compiler for Standard ML code. ↩︎