View Single Post
  #33  
Old 02-11-2004, 11:42 PM
ron101's Avatar
ron101 ron101 is offline
Member
 
Join Date: May 2002
Location: North Vancouver, BC
Posts: 262
ron101 is on a distinguished road
Default

Quote:
Would you not theoretically gain a slight efficiency advantage with higher voltage (ie. less heat produced by the ballasts)?

Quote:
If there is something like that, my guess is it's minimal or negligible. Of all the stats/specs I read about ballasts, they all more or less have the same "wattage" (which is roughly amps times voltage, give or take) regardless of what input voltage is used.
Overall power draw on the circuit is voltage x amperage; which should remain the same for both a 110 and a 220v setup. However power lost to heat is [correction : resistance x current squared] so a 220v setup should have less than a 110v and thus the difference should make it to the lamp given a constant line resitance.

Anyways, consider it a proposed experiment for someone with a 220v hookup and a PAR meter .[/i]
Reply With Quote