Contrary to indications made by prior researchers, digital logic circuits designed by artificial evolution to perform binary arithmetic tasks can generalise on inputs which were not seen during evolution. This phenomenon is demonstrated experimentally and speculatively explained in terms of the regular structure of binary arithmetic tasks and the nonoptimality of random circuits. This explanation rests on an assumption that evolution is relatively unbiased in its exploration of circuit space. Further experimental data is provided to support the proposed explanation.