Let us suppose that a transmission line is carrying power to a resistor $R$ The resistor $R$ requires a power of $P=IV$.
Applying the most simple case where a transmission line is supplying a current $I$. The Ohmic loss is $P_c=I^2R_c$ for the cable.
Putting $I=P/V$
$P_c=(P^2/V^2)R_c$ So to decrease Ohmic loss we need to increase $V$ but this is the potential difference across the resistor, right? I have read other questions here, but it still doesn't make sense. Please help.
What my textbook says
Consider a resistor R to which a power P is to be delivered via a transmission cable having res. $R_c$ to be dissipated finally. If $V$ is voltage across $R$ and $I$ the current then
$P=VI$
The connecting wires from the power station to the resistor has res. $R_c$. The Ohmic loss is $P_c=I^2R_c$
$P_c=(P^2/V^2)R_c$
Thus to drive a device with $P$ the power wasted in the connecting wires is inversely proportional to $V^2$.
A few paragraphs later...
The transmission cables from power stations are hundreds of miles long and their resistance $R_c$ is considerable. To reduce $P_c$, these wires carry current at enormous values of $V$