this post was submitted on 26 Aug 2024
167 points (93.7% liked)

Technology

59554 readers
3393 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

As we all know, AC won the "War of the Currents". The reasoning behind this is that AC voltage is easy to convert up/down with just a ring of iron and two coils. And high voltage allows us to transport current over longer distances, with less loss.

Now, the War of the Currents happened in 1900 (approximately), and our technology has improved a lot since then. We have useful diodes and transistors now, we have microcontrollers and Buck/Boost converters. We can transform DC voltage well today.

Additionally, photovoltaics produces DC naturally. Whereas the traditional generator has an easier time producing AC, photovoltaic plants would have to transform the power into AC, which, if I understand correctly, has a massive loss.

And then there's the issue of stabilizing the frequency. When you have one big producer (one big hydro-electric dam or coal power plant), then stabilizing the frequency is trivial, because you only have to talk to yourself. When you have 100000 small producers (assume everyone in a bigger area has photovoltaics on their roof), then suddenly stabilizing the frequency becomes more challenging, because everybody has to work in exactly the same rhythm.

I wonder, would it make sense to change our power grid from AC to DC today? I know it would obviously be a lot of work, since every consuming device would have to change what power it accepts from the grid. But in the long run, could it be worth it? Also, what about insular networks. Would it make sense there? Thanks for taking the time for reading this, and also, I'm willing to go into the maths, if that's relevant to the discussion.

you are viewing a single comment's thread
view the rest of the comments
[โ€“] BearOfaTime@lemm.ee 2 points 2 months ago (1 children)

Does it? (Genuine question).

That seems rather simplistic - I'm sure there are plenty of use-cases where AC is a better choice than DC and vice-versa (though I'm no EE).

Very rarely is any tech that binary.

[โ€“] bastion@feddit.nl 2 points 2 months ago* (last edited 2 months ago)

Yeah. Basically, the biggest reasons for AC have to do with voltage stepping up and down, and for instant grid load knowledge. Well, and of course, existing infrastructure.

Both have solutions, but aren't as cheap as they are for AC. But, aside from that, DC has a lot of benefits, particularly in end usage efficiency and transmission over distance.

Back in the day, the capability to easily bump up or down the voltage of electricity just wasn't there for DC, so AC was the distance winner (high voltage is needed for distance, low voltage typically needed for usage).