qwencoder-eval/instruct/eval-dev-quality/docs/reports/v0.5.0/symbolic-execution/README.md
This report was generated by DevQualityEval benchmark in version 0.5.0.
PLEASE NOTE that we assume the results of the "symbolic execution" are deterministic. Hence they where only benchmarked over a single run and multiplied by 5 to be comparable to the remaining contestants.
Keep in mind that LLMs are nondeterministic. The following results just reflect a current snapshot.
The results of all models have been divided into the following categories:
The following sections list all models with their categories. The complete log of the evaluation with all outputs can be found here. Detailed scoring can be found here.
Models in this category could not be categorized.