Table of Contents
Knowing how many watts a TV uses matters for several reasons: it affects your electricity bill, helps you size a generator or solar system for backup power, and impacts your carbon footprint. By understanding TV power usage, you can better manage household costs, plan off-grid systems, and make more eco-friendly choices.
TV Power Usage at a Glance
Most modern TVs use between 50 and 200 watts, depending on size, technology, and settings. Here’s a quick look:
- 32-inch LED TV: around 30–60 watts
- 55-inch LED TV: around 83–104 watts
- Older or larger TVs (like plasma): up to 200 watts or more
Volts and Amps
In North America, TVs typically run on 120 volts. You can calculate the current (amps) with:
- Watts = Volts × Amps
So a 100-watt TV on a 120-volt supply draws roughly 0.83 amps.
Monthly Electricity Costs
The cost to run a TV depends on its wattage, how long it’s used each day, and local electricity rates. For example:
- A 100-watt TV used 3 hours a day uses about 9 kWh per month, costing around $1–$2 monthly in most U.S. states. Larger or less efficient TVs will cost more.
Standby Power
Even when switched off, most TVs draw 0.5–5 watts in standby mode. Over a year, this can account for 2–5% of the TV’s total energy use.
How to Cut Your TV’s Energy Use?
- Turn off or unplug the TV completely to avoid standby draw.
- Lower brightness and contrast settings.
- Use built-in energy-saving modes.
- Set a sleep timer so it turns off automatically.
- Choose ENERGY STAR or similarly rated efficient models.
- Remember: bigger screens and higher resolutions consume more power.
Environmental Considerations
Running a TV contributes to carbon emissions, especially if your electricity comes from fossil fuels. Choosing an energy-efficient model and using it wisely can reduce your environmental impact.
In Short:
Most TVs use 50–200 watts, costing about $1–$2 a month for average use. By managing your TV’s power consumption, you can lower your bills, ensure your backup systems are properly sized, and help the environment.
How Many Watts Does a TV Use?
Quick Takeaway:
- Modern TVs generally use 30 to 200 watts, depending on size and type.
- LED TVs are the most energy-efficient, while plasma and CRT TVs consume the most.
Average TV wattage by size and type:
| TV Size | LED | LCD | OLED | Plasma |
| 32-inch | 30–55 W | 50–85 W | — | up to 160 W |
| 40–43-inch | 50–80 W | up to 100 W | — | up to 200 W |
| 50–55-inch | 60–90 W | 180 W | 98–110 W | up to 370 W |
| 65-inch | 115–150 W | 170–220 W | 190–300 W | up to 500 W |
Typical energy costs:
- A 100 W TV used 3 hours daily costs about $1–$2 per month, or roughly $15–$20 per year, depending on local electricity rates. Larger TVs or longer viewing times will cost more.
Watts vs Kilowatt-Hours vs Volts vs Amps
- Watt (W): Measures how much power the TV uses at a moment.
Example: A 100 W TV consumes 100 watts while it’s on. - Kilowatt-hour (kWh): Measures energy use over time.
Example: 100 W used for 10 hours = 1 kWh. Your electricity bill is based on kWh. - Volt (V): Electrical pressure. Most TVs in North America use 120 V outlets.
Amp (A): The flow of electric current.
Calculated as:
Amps = Watts ÷ Volts
- So a 100 W TV on 120 V draws about 0.83 amps.
| Term | What it means | Example for TV |
| Watt (W) | Instant power use | 100 W TV |
| kWh | Energy used over time | 100 W × 10 hrs = 1 kWh |
| Volt (V) | Electrical pressure | 120 V outlet |
| Amp (A) | Current flow | 100 W ÷ 120 V = 0.83 A |
Knowing these basics helps you understand energy labels and estimate costs.
TV Wattage by Size and Type
| TV Size | LED | OLED | LCD | CRT | Plasma |
| 24-inch | 35 W | — | 50 W | — | 120 W |
| 32-inch | 41 W | 55–60 W | 50–85 W | 150–200 W | 160 W |
| 50-inch | 72 W | 89 W | 150 W | — | 300 W |
| 55-inch | 80 W | 98 W | 180 W | — | 370 W |
| 65-inch | 88 W | 107 W | 200 W | — | 500 W |
| 75-inch+ | 100–150 W (est.) | 120–200 W (est.) | — | — | 500+ W (historical) |
Wattage by TV Type
- CRT: 60–200 W, with larger models on the higher end.
- Plasma: 100–500 W, very power-hungry especially at larger sizes.
- LCD: 50–200 W.
- LED: 30–100 W, the most efficient common type.
- OLED: 50–200 W, more efficient than plasma but usually draws more power than LED.
Smart TVs vs non-smart:
Smart TVs may use slightly more power due to added processors and Wi-Fi, but the difference is usually minor compared to screen type and size.
Full Quick Reference Table
| Size | LED (W) | OLED (W) | Plasma (W) | LCD (W) | CRT (W) |
| 24″ | 35 | — | 120 | 50 | 120 |
| 32″ | 41 | 55–60 | 160 | 50–85 | 150–200 |
| 50″ | 72 | 89 | 300 | 150 | — |
| 55″ | 80 | 98 | 370 | 180 | — |
| 65″ | 88 | 107 | 500 | 200 | — |
| 75″+ | 100–150 | 120–200 | 500+ | 220 | — |
These figures are averages; actual power use depends on brand, model, and settings. Always check the EnergyGuide label or manufacturer’s specs for precise numbers.
Bottom line:
LED TVs are generally the best choice for energy savings, plasma and CRTs are the worst. Knowing your TV’s wattage helps you estimate electricity costs, size a generator or solar system, and make smarter eco-friendly decisions.
How Much Electricity Does a TV Use?

Electricity Use Per Hour
To figure out how much electricity a TV uses each hour, convert its wattage to kilowatt-hours (kWh):
- kWh per hour = TV wattage ÷ 1000
For example, a 100 W TV uses:
- 100 ÷ 1000 = 0.1 kWh per hour
Electricity Use Per Day and Per Month
Average daily viewing:
In 2025, Americans watch roughly 2.8 hours of traditional TV daily.
Example for a 100 W TV:
- Daily:
0.1 kWh/hour × 2.8 hours = 0.28 kWh/day
- Monthly:
0.28 kWh/day × 30 = 8.4 kWh/month
Cost at $0.15 per kWh:
- Monthly:
8.4 × $0.15 = $1.26
Electricity Use Per Year
Annual example for a 100 W TV:
0.28 kWh/day × 365 = 102.2 kWh/year
Annual cost at $0.15 per kWh:
102.2 × $0.15 = $15.33
Summary Table for a 100 W TV at 2.8 hours/day
| Period | Energy Used | Cost (@$0.15/kWh) |
| 1 hour | 0.1 kWh | $0.015 |
| 1 day | 0.28 kWh | $0.042 |
| 1 month | 8.4 kWh | $1.26 |
| 1 year | 102.2 kWh | $15.33 |
Actual figures will vary with your TV’s wattage, how long you watch, and local electricity rates. Larger or older TVs will use more.
How Many Volts and Amps Does a TV Use?
Typical Voltage
- United States: Most TVs run on 110–120 volts.
- International: Many countries use 220–240 volts.
Modern TVs are often “dual voltage,” but always check the label before plugging into a different standard.
Typical Amps by TV Size
The current a TV draws depends on its wattage and voltage:
Amps = Watts ÷ Volts
| TV Size | Wattage | Amps (US, 120V) | Amps (Intl, 230V) |
| 24″ | 35 W | 0.29 A | 0.15 A |
| 32″ | 41 W | 0.34 A | 0.18 A |
| 50″ | 72 W | 0.60 A | 0.31 A |
| 55″ | 80 W | 0.67 A | 0.35 A |
| 65″ | 88 W | 0.73 A | 0.38 A |
| 75″+ | 120 W | 1.00 A | 0.52 A |
In short:
- Most TVs in the US draw 0.3 to 1 amp, depending on size.
- International homes see lower amps because of higher voltage, usually 0.15 to 0.5 amps.
What Factors Affect TV Power Consumption?
Screen Size
Larger TVs use more energy to light a bigger display and run more powerful electronics.
Display Technology
- CRT: Older, bulky TVs use 100–400 W.
- Plasma: Great contrast but inefficient, 100–500 W.
- LCD: About 50–200 W, depending on backlight.
- LED: Most efficient, often 30–100 W.
- OLED: Outstanding quality, typically 50–200 W, higher for large or bright models.
Resolution
Higher resolutions like 4K and 8K need more processing and brighter backlights, which increases power use.
Brightness & Settings
Using high brightness or vivid modes can raise energy use by 20–40% over standard or eco modes.
Input Source
Streaming or smart features can use slightly more power than HDMI or over-the-air broadcasts, but the difference is small compared to size and brightness.
Age of the TV
Newer models are much more efficient. ENERGY STAR TVs use up to 25% less energy than non-certified models. Older CRTs, plasmas, and early LCDs use significantly more.
Standby Power
Even when “off,” most TVs draw 0.5–5 W to stay ready for quick start or updates. This “vampire power” can add up if you leave TVs plugged in all year.
In summary:
A TV’s energy use mainly depends on screen size, technology, resolution, brightness, input source, and age. Adjusting settings, picking energy-efficient models, and managing standby features can all lower electricity use and save money.
How to Determine Your TV’s Power Usage
You can figure out how much power your TV uses with three simple approaches:
1. Check the Manufacturer Label
Look for a sticker on the back of your TV or near the power cord. It usually lists power consumption in watts (W) or amps (A). This gives you the maximum draw for your specific model — quick and straightforward.
2. Use a Kill-A-Watt Meter
A Kill-A-Watt or similar power meter plugs into your outlet, with your TV plugging into the meter. It shows exactly how many watts your TV is using in real time, whether it’s on or in standby. This is the most accurate way to see real-world usage based on your actual settings and habits.
3. Calculate Manually
If you know the TV’s wattage and how many hours you watch each day, you can estimate energy use.
Formulas:
- Watt-hours (Wh) = Watts × Hours Used
- Kilowatt-hours (kWh) = Wh ÷ 1000
Example:
A 100 W TV used for 3 hours daily:
- 100 W × 3 hours = 300 Wh = 0.3 kWh per day
- Over 30 days: 0.3 kWh × 30 = 9 kWh per month
In short:
- Check the label for a quick idea.
- Use a meter for precise measurements.
- Do the math to estimate monthly or yearly usage.
These methods help you track energy use, manage costs, and size generators or batteries if needed.
TV Power Consumption When Off
Standby Power (Vampire Draw)
When you turn off your TV with the remote, it doesn’t completely disconnect from power. Instead, it goes into standby mode, drawing about 0.5 to 3 watts (sometimes up to 5 watts) to stay ready for quick start, remote signals, and software updates. This small draw typically makes up 2–5% of a TV’s annual energy use, but it adds up over time, especially if you have multiple TVs.
Cost of Standby Power
Example with a TV using 1.3 W in standby:
- Daily: 1.3 W × 24 h = 31.2 Wh = 0.0312 kWh
- Monthly: 0.0312 kWh × 30 = 0.94 kWh
- Yearly: 0.0312 kWh × 365 = 11.4 kWh
At an average rate of $0.15 per kWh, that’s about $1.70 per year.
Example with a higher 3 W standby:
- Yearly: 3 W × 24 × 365 = 26.3 kWh
- Annual cost: 26.3 kWh × $0.15 = about $4.00 per year.
How to Avoid Standby Costs
To eliminate standby power completely, unplug your TV when not in use or plug it into a power strip that you can easily switch off. This cuts off the flow of electricity entirely and stops the hidden costs.
How Much Does It Cost to Power a TV?
Monthly and Yearly Costs by TV Size and Type
Here’s what it typically costs to run different TVs in the U.S., assuming an average electricity rate of $0.16 per kWh and about 5 hours of use per day.
| Screen Size | Typical Power (W) | Avg Monthly Cost | Avg Yearly Cost |
| 24″ LED | 40–50 | $1.54–$1.92 | $18.48–$23.04 |
| 32″ LED | 50–70 | $1.92–$2.69 | $23.04–$32.28 |
| 42″ LED | 80–120 | $3.07–$4.61 | $36.84–$55.32 |
| 50″ LED | 100–150 | $3.84–$5.76 | $46.08–$69.12 |
| 55″ LED | 110–160 | $4.22–$6.14 | $50.64–$73.68 |
| 65″ LED | 120–180 | $4.61–$6.91 | $55.32–$82.92 |
| 75″ LED | 150–220 | $5.76–$8.45 | $69.12–$101.40 |
| 32″ Plasma | 160 | $6.14 | $73.68 |
| 50″ Plasma | 300 | $11.52 | $138.24 |
| 55″ Plasma | 370 | $14.21 | $170.52 |
| 60″ Plasma | 500 | $19.20 | $230.40 |
- LED and LCD TVs are the most energy-efficient.
- Plasma and older CRT models are far more expensive to run.
- Smart TVs may use a bit more power due to extra processing and Wi-Fi, but the difference is minor compared to screen size and display type.
Note: Standby power isn’t included here but usually adds $1–$4 per year per TV.
State-by-State Cost Differences
Electricity rates vary widely across the U.S., which means your actual cost to run a TV can differ quite a bit.
- Lower-cost states: Idaho, Washington, Louisiana (often below $0.12/kWh).
- Higher-cost states: Hawaii, California, New England (can exceed $0.30/kWh).
Example for a 50″ LED TV (100 W used 5 hours daily):
- At $0.12/kWh: roughly $2.16 per month
- At $0.30/kWh: roughly $5.40 per month
In summary:
Most TVs cost between $1.50 and $8 per month to run, depending on size, technology, usage, and local electricity rates. Plasma and older models are much pricier to operate.
Can You Run a TV on a Generator or Solar?
Running a TV on a Generator
Most modern TVs use 50 to 200 watts, which means they draw less than 1 amp on a standard 120 V circuit. Nearly any portable generator—even small models rated at 500–1000 W—can easily power a TV, often alongside other low-wattage devices.
Typical TV wattage by size:
- 32″ LED: 41–70 W
- 50″ LED: 70–100 W
- 65″ LED: 95–200 W
- Plasma and CRT: 120–500+ W
So for general backup or camping needs, almost any household generator will handle a TV without issue.
Choosing the Right Solar Generator
To size a solar generator, multiply your TV’s wattage by expected hours of use.
Example:
A 100 W TV used for 3 hours needs 300 Wh per day.
Popular solar generator options:
| Model | Capacity (Wh) | Run Time for 100 W TV |
| Jackery 1000 | 1002 Wh | ~10 hours |
| Jackery 2000 | 2160 Wh | ~21 hours |
| Jackery 3000 | 3024 Wh | ~30 hours |
Actual run times vary based on inverter efficiency and whether you’re powering other devices too.
Who Should Get What?
- Jackery 1000: Great for short camping trips, a single TV during brief outages, or outdoor movie nights.
- Jackery 2000: Better for longer trips, running a TV plus a few small devices, or extended power cuts.
- Jackery 3000: Ideal for powering multiple devices all day or as a more complete home backup.
Bottom line:
You can easily run a modern TV on nearly any generator or solar system. Just size your solar setup to match your TV’s wattage and daily use. Even small portable power stations can keep you watching for hours, while larger systems support all-day or multi-device use.
Tips to Reduce TV Power Consumption
Lower Backlight or Brightness
One of the simplest ways to cut energy use is by reducing your TV’s backlight or brightness settings. TVs often come with brightness levels set much higher than necessary. Lowering this can cut power consumption by nearly half without noticeably affecting picture quality.
Use Eco or Energy-Saving Modes
Most modern TVs include an eco or energy-saving mode. These automatically adjust brightness, contrast, and sometimes backlight to use less power. Enabling this feature can reduce energy use by about a third.
Turn Off Completely vs Standby
Even when turned off by remote, TVs stay in standby mode, drawing “vampire power.” To stop this entirely, unplug the TV or use a power strip and switch it off when not in use.
Use a Power Strip
Plug your TV and related devices (like sound systems or game consoles) into a power strip. Turning it off cuts power to everything at once, ensuring no energy is wasted in standby.
Buy ENERGY STAR or High-Efficiency TVs
When replacing or upgrading, look for TVs with ENERGY STAR labels or high efficiency ratings. These models are designed to use significantly less electricity than standard models.
Other Helpful Tips
- Set a sleep timer so your TV turns off automatically.
- Use blank screen or audio-only modes if you just want sound.
- Let ambient light sensors adjust brightness to match your room.
- Lower contrast settings, which also affects energy draw.
- Automate your TV’s on/off schedule if your model supports it.
By combining these steps, you can meaningfully reduce your TV’s electricity use and cut down on your power bills—without sacrificing your viewing experience.
Conclusion
Most modern TVs use 30–100 watts while on and 0.5–3 watts in standby, costing around $1–$2 per month to run. Costs climb for larger or older models. To save energy and money, check your TV’s wattage, enable eco settings, lower brightness, and unplug or use a power strip to stop standby power draw. Always review your TV’s manual or energy label for the most accurate figures, and watch smart to save more.
TV Wattage FAQs
Do TVs use a lot of electricity?
Not usually. Modern LED and LCD TVs are energy-efficient, typically using 30–100 watts. Older plasma or CRT TVs can use several hundred watts and noticeably raise electricity bills. For most households today, TV power use is a small part of the total bill.
How much does it cost to run a TV each month?
For a typical modern TV, expect to pay $1–$2 per month, or about $12–$24 a year, based on average U.S. rates and normal daily viewing. Larger or older TVs can cost more, while efficient or smaller models cost less.
How much power does a TV use when off?
In standby mode, most TVs draw 0.5–3 watts, which can add up to $1–$4 per year. This small standby load can account for 2–5% of a TV’s total annual energy use.
Is there a best time of day to watch TV for lower costs?
If your utility has time-of-use rates, running your TV during off-peak hours (overnight or early morning) can save money. Otherwise, with flat-rate billing, time of day doesn’t impact cost.
Does screen saver mode save energy?
No. Screen savers on TVs generally use nearly as much power as watching video. To truly save, turn off the TV entirely.

Aditya Kumar B. is your go to lifestyle curator sharing daily entertainment rich tips and tricks for streamlined, joyful living. Avidly exploring the worlds of streaming and leisure, Adi blends chill vibes with thoughtful recommendations, helping audiences stay on top of trending shows, streaming hacks, and laid back self care routines. Whether you’re searching for your next binge worthy series, the perfect TV setup, or clever ways to unwind, Aditya has your back making every moment feel relaxed and effortlessly enjoyable.
