Abstract. The DL implementation group (DIG) has specified a common interface for description logic reasoner. While the original intention was to make it easier for other tools to access these reasoning systems, the interface might also be used for evaluating the very reasoning systems. This paper presents a framework that is able to do that evaluation w.r.t. correctness of reasoning results and performance measurements. For usability, the evaluation framework is built on scriptable test cases. All the results of tests cases are stored in a relational database and can be queried by using standard SQL.