The Web has established itself as the largest public data repository ever available. Even though the vast majority of information on the Web is formatted to be easily readable by the human eye, “meaningful information” is still largely inaccessible for the computer applications. In this paper, we present automated algorithms to gather meta-data and instance information by utilizing global regularities on the Web and incorporating with contextual information. Experimental evaluations successfully performed on the TAP knowledge base and the faculty-course home pages of computer science departments containing 16,861 Web pages. The system achieves this performance without any domain specific engineering requirement.