SMT-COMP

The International Satisfiability Modulo Theories (SMT) Competition.

GitHub

Home
Introduction
Benchmark Submission
Publications
SMT-LIB
Previous Editions

SMT-COMP 2023

Rules
Benchmarks
Specs
Model Validation Track
Proof Exhibition Track
Parallel & Cloud Tracks
Participants
Results
Statistics
Comparisons
Slides

QF_Bitvec (Model Validation Track)

Competition results for the QF_Bitvec division in the Model Validation Track.

Page generated on 2023-07-06 16:06:00 +0000

Benchmarks: 9069
Time Limit: 1200 seconds
Memory Limit: 60 GB

Logics:

Winners

Sequential PerformanceParallel Performance
STPSTP

Sequential Performance

Solver Error Score Correct Score CPU Time Score Wall Time ScoreAbstainedTimeout Memout
Bitwuzla Fixedn 0 9025 67902.025 67731.77345 0
STP 0 9024 57693.477 57556.81845 0
Bitwuzla 0 9019 71359.346 71216.29650 0
2020-Bitwuzlan 0 9006 91083.336 90763.56863 0
Yices2 0 8973 99192.946 99112.39396 0
cvc5 0 8650 134396.618 133952.196411 3

Parallel Performance

Solver Error Score Correct ScoreCPU Time ScoreWall Time ScoreAbstainedTimeout Memout
STP 0 902457693.47757556.81845 0
Bitwuzla Fixedn 0 902466703.50566531.74345 0
Bitwuzla 0 901971359.34671216.29650 0
2020-Bitwuzlan 0 900691083.33690763.56863 0
Yices2 0 897399192.94699112.39396 0
cvc5 0 8650134396.618133952.196411 3

n Non-competing.
Abstained: Total of benchmarks in logics in this division that solver chose to abstain from. For SAT/UNSAT scores, this column also includes benchmarks not known to be SAT/UNSAT.