Order of accuracy refers to the rate at which the error in a numerical approximation decreases as the step size of the approximation is reduced. It is a measure of the precision of a numerical method and is typically expressed as a power of the step size. A higher order of accuracy implies a more precise approximation, but it also typically requires more computational resources.