This paper introduces three (one linear and two nonlinear) automatic scaling techniques for NLPs with states and constraints spread over several orders of magnitude, without requiring complex off-the-shelf external tools. All of these methods have been compared to standard techniques and applied to three problems using SNOPT and IPOPT. The results confirm that the proposed techniques significantly improve the NLP conditioning, yielding more reliable and in some cases, faster NLP solutions.