Abstract. Alternating least squares (ALS) is a powerful matrix factorization (MF) algorithm for both implicit and explicit feedback based recommender systems. We show that by using the Sherman-Morrison formula (SMF), we can reduce the computational complexity of several ALS based algorithms. It also reduces the complexity of greedy forward and backward feature selection algorithms by an order of magnitude. We propose linear kernel ridge regression (KRR) for users with few ratings. We show that both SMF and KRR can efficiently handle new ratings. Key words: matrix factorization, collaborative filtering, alternating least squares, Sherman-Morrison formula, kernel ridge regression, greedy feature selection