The idea is that with a suitable choice of measurement scheme, you can ask essentially any question "later", and estimate the information you're looking for efficiently.
In other words, you can fix the measurement strategy, if that alone allows to recover any observable of interest about input states.
Classical shadows are just estimators defined on individual measurement outcomes. If the measurement is informationally complete, as these schemes usually are, then any piece information is recoverable, with sufficient statistics.
You can prove that for suitable choices of measurement settings, you can estimate arbitrary observables with a number of resources that does not depend on the dimension of the measured state, which is what you'd consider "efficient" in this context. You can in fact estimate many observables at the same time in this way. You can estimate $M$ observables with a probability $\delta$ of the max error being less than $\epsilon$ with something like (and I'm going by memory here) $N\sim \log(M) \log(2/\delta)/\epsilon^2$. This is using measurement schemes that are very symmetric, such as the original "random projective measurements" scheme of Huang et al. That's efficient enough that it can make sense to not tailor the measurement apparatus to recover a specific observable, unless you really need to optimise for that.