This paper presents a method of hand posture estimation from silhouette images taken by two cameras. First, we extract the silhouette contour for a pair of images. We construct an eigenspace from images of hands with various postures. For effective matching, we define a shape complexity for each image to see how well the shape feature is represented. For a pair of input images, the total matching error is computed by combining the two matching errors according to the shape complexity. Thus the best-matched image is obtained for a pair of images. For rapid processing, we limit the matching candidate by using the constraint on the shape change. The possible shape transition is represented by a transition network. Because the network is hard to build, we apply offline learning, where nodes and links are automatically created by showing examples of hand shape sequences. We show experiments of building the transition networks and the performance of matching using the network.