We consider the regression model in the situation when the number of available regressors pn is much bigger than the sample size n and the number of nonzero coefficients p0n is small (the sparse regression). To choose the regression model, we need to identify the nonzero coefficients. However, in this situation the classical model selection criteria for the choice of predictors like, e.g., the Bayesian Information Criterion (BIC) overestimate the number of regressors. To address this problem, several modifications of BIC have been recently proposed. In this paper we prove weak consistency of some of these modifications under the assumption that both n and pn as well as p0n go to infinity.