That's an amusing post. First you disagree with me, and then at the end do agree.
Fact is, no matter where you are, temperature rise will degrade mppt gain. It's just a matter of how much.
If you got 1000W/m^2 irradiance, then a one meter square panel will absorb about 800W continuous. It doesn't matter what the ambient temp is, the panel will get hot!
Sal
ktmrfs wrote:
Salvo wrote:
You're beginning to see the negative sides of mppt. From the get-go, mppt has about 30% advantage over pwm. But much of that advantage evaporates. Sal
not really. MPPT will do better, ...
And if I was strictly using a panel in the summer in the southern US in high temps, not sure that MPPT would be enough gain to justify the expense.