this post was submitted on 26 Aug 2024
168 points (93.8% liked)

Technology

59598 readers
3513 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

As we all know, AC won the "War of the Currents". The reasoning behind this is that AC voltage is easy to convert up/down with just a ring of iron and two coils. And high voltage allows us to transport current over longer distances, with less loss.

Now, the War of the Currents happened in 1900 (approximately), and our technology has improved a lot since then. We have useful diodes and transistors now, we have microcontrollers and Buck/Boost converters. We can transform DC voltage well today.

Additionally, photovoltaics produces DC naturally. Whereas the traditional generator has an easier time producing AC, photovoltaic plants would have to transform the power into AC, which, if I understand correctly, has a massive loss.

And then there's the issue of stabilizing the frequency. When you have one big producer (one big hydro-electric dam or coal power plant), then stabilizing the frequency is trivial, because you only have to talk to yourself. When you have 100000 small producers (assume everyone in a bigger area has photovoltaics on their roof), then suddenly stabilizing the frequency becomes more challenging, because everybody has to work in exactly the same rhythm.

I wonder, would it make sense to change our power grid from AC to DC today? I know it would obviously be a lot of work, since every consuming device would have to change what power it accepts from the grid. But in the long run, could it be worth it? Also, what about insular networks. Would it make sense there? Thanks for taking the time for reading this, and also, I'm willing to go into the maths, if that's relevant to the discussion.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 11 points 3 months ago (2 children)

How does the efficiency and cost of buck converters compare to AC transformers? Seems like the cost and efficiency of the voltage converter should be the prime determinant, rather than specific applications of generation/consumption.

What would a 400A 10kV utility scale DC converter look like?

[–] [email protected] 4 points 3 months ago* (last edited 3 months ago)

A lot of power blocks rectify the AC and then shift the DC voltage anyway, because transformers are bulky and heavy. This is why power bricks are so much smaller now than they used to be. See the modes section of the Wikipedia page about AC adapters for a longer explanation.

[–] [email protected] 0 points 3 months ago* (last edited 3 months ago)

well, a large part why I asked the question is because I hope that somebody knows more about what buck/boost-converters can do today. I know they work well enough on small scales, but I have no experimental data for them on larger scales.

I assume they would work well, but I'd like that somebody links me to the right datasheet or something.

Edit: you have a very important point there. " Seems like the [...] voltage converter should be the prime determinant, rather than specific applications of generation/consumption." YES. So, let me rephrase my question: does anybody have experience with high-power DC voltage converters?