Similar questions feature found about 10 QA's here on the subject with all about 1-5 cumulative upvotes. For example Why do power lines use high voltage? have 6 upvotes for accepted answer:
The line loss is given by $P_{loss}=I^2R$, or, substituting for $I$, $P_{loss} = \frac {P^2R}{V^2}$
However in wiki Joule_heating there is an assumption for using that formula:
Assuming the element behaves as a perfect resistor and that the power is completely converted into heat, the formula can be re-written by substituting Ohm's law
Why is Ohm law applied in many answers and articles explaining the reason for high voltage when it is applicable for complete conversion of energy (but losses are less than 100% in mains power)? or wiki incorrectly adds this condition?
P.S. the answer quoted above is actually much better than a place from where I started my "quest", those as far as I recall just stated voltage is stepped up because losses are square of current; but (if Ohm law is applied) losses are also square of voltage $P_{loss}=I^2R=V^2/R$ so I wanted better explanation and started web search and read wiki, but (at least wiki Joule_heating page) on the subject stated only:
However, in applications where heating is an unwanted by-product of current use (e.g., load losses in electrical transformers) the diversion of energy is often referred to as resistive loss. The use of high voltages in electric power transmission systems is specifically designed to reduce such losses in cabling by operating with commensurately lower currents.