Abstract--In the Relational Reinforcement learning framework, we propose an algorithm that learns an action model allowing to predict the resulting state of each action in any given situation. The system incrementally learns a set of first order rules: each time an example contradicting the current model (a counter-example) is encountered, the model is revised to preserve coherence and completeness, by using data-driven generalization and specialization mechanisms. The system is proved to converge by storing counter-examples only, and experiments on RRL benchmarks demonstrate its good performance w.r.t state of the art RRL systems. Keywords-relational reinforcement learning; inductive logic programming; online and incremental learning