answersLogoWhite

0


Best Answer

The 200 watts on an amplifer would refer to output if it were turned up all the way. This does not tell you how much electricity the amplifier would use. There is some power lost from heat, but we don't know just how much, from the information above.

User Avatar

Wiki User

15y ago
This answer is:
User Avatar
More answers
User Avatar

AnswerBot

6mo ago

A 200-watt bulb will use approximately 1.67 amps at 120 volts. This is calculated by dividing the wattage of the bulb by the voltage of the circuit it is connected to.

This answer is:
User Avatar

User Avatar

Wiki User

11y ago

amps equals watts divided by volts.

This answer is:
User Avatar

Add your answer:

Earn +20 pts
Q: How many amps does a 200watt bulb use?
Write your answer...
Submit
Still have questions?
magnify glass
imp
Related questions

How do you work out how many amps a forty watt bulb uses?

To calculate the amperage of a 40-watt bulb, you need to use the formula: Amps = Watts / Volts. If the bulb operates at 120 volts (standard for US households), the amperage will be 0.33 amps (40 watts / 120 volts).


How many amperes does a standard 60 watt light bulb use?

Watts = Volts X Amps. Amps=Watt / Volts. So, with a 240V mains, a 60W bulb draws 0.25amps. On a 12 system (car/auto) a 60W bulb draws 5 amps. On a 110V mains, a 60W bulb draws .55 Amps.


How many amps does a 65 watt light bulb have at 120 volts?

To calculate the amperage, you can use the formula: Amps = Watts/Volts. For a 65-watt light bulb at 120 volts, the amperage would be 0.54 amps.


How many amps does a 50 watt bulb use?

You just have to divide the watts by the voltage to find the amps. For example 60 watts on a 120 v system would take ½ amp.


How many watts does a 12volt X 65milliamp bulb use?

Watts = Amps x Volts. Milliamps to amps move the decimal place three places to the left.


How many amps does a 65 watt light bulb use?

A 65-watt light bulb typically draws around 0.54 amps of current when operated at the standard voltage of 120 volts. This can vary slightly depending on the exact voltage and type of bulb.


How many ampere in 240 volt with 100 watts bulb?

To calculate the amperage, you can use the formula: Amperes = Watts / Volts. So for a 100-watt bulb connected to a 240-volt source, the amperage would be 0.42 Amps.


How many amp hours does a 23 watt bulb use?

The formula for finding amperage is I = W/E. Amps = Watts/Volts. Without the value of the voltage to the bulb this question can not be answered.


How many amps in a 300watt bulb?

There are zero amps in a 300 watt bulb. Watts are the product of amps times volts. W = A x V. To find amperage use the equation A = W/V. If you find the voltage of the 300 watt bulb uses, divide the wattage by the voltage and it will give you the amps of a 300 watt bulb.


What equation would you use to calculate how much current in AMPs is required to achieve the full 100 to output from a volt bulb?

The bulb is marked with the power (watts) and the voltage. Divide the watts by the volts and you have the amps.


How many amps does ATM use?

10 amps


How many amps does a 120v 50w halogen light bulb use?

At what voltage? Until you tell me the voltage I can't give you an answer. To find out Amps you need to divide the Watts by the Volts. At 120V you have 0.4 amps. At 12V you have 4 amps.