this post was submitted on 07 Sep 2024
58 points (100.0% liked)

TechTakes

1489 readers
77 users here now

Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.

This is not debate club. Unless it’s amusing debate.

For actually-good tech, you want our NotAwfulTech community

founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 34 points 3 months ago (6 children)

The learning facilitators they mention are the key to understanding all of this. They need them to actually maintain discipline and ensure the kids engage with the AI, so they need humans in the room still. But now roles that were once teachers have been redefined as "Learning facilitators". Apparently former teachers have rejoined the school in these new roles.

Like a lot of automation, the main selling point is deskilling roles, reducing pay, making people more easily replaceable (don't need a teaching qualification to be a "learning facilitator to the AI) and producing a worse service which is just good enough if it is wrapped in difficult to verify claims and assumptions about what education actually is. Of course it also means that you get a new middleman parasite siphoning off funds that used to flow to staff.

[–] [email protected] 12 points 3 months ago

Aren't there laws about who gets to teach kids? I know there are strictures on teacher to student ratios, but how can that exist without a written definition of what a teacher is?

[–] [email protected] 12 points 3 months ago

They could just have the kids read actual books designed by actual pedagogic experts which actually help to learn through studying it.

Now nobody knows if the "AI" is even teaching real things or if it is only using properly vetted material, if the structure it proposes makes sense.

Yes teachers are fallible, but they are also human and can emotionally understand what is going on during learning that a trained algorithm just cannot get. In so far also it means there needs to be a clearly defined "goal" of knowledge and competencies and the algorithm can only fill the holes, rather than encourage students to maybe seek knowledge beyond the established set.

Also i am skeptical how much of it is even "AI" in the sense of needing a machine learning approach, or it is just regular computer tests of which "level" is reached in each category and where to improve still. Chance is, this could be done with an excel sheet.

[–] [email protected] 8 points 3 months ago

so the thing is this is a private school at the sort of fees that attract really good teachers and use them as a selling point, so I don't actually think being cheap is the goal here. I think some idiot thinks this is actually a good idea.

[–] [email protected] 7 points 3 months ago

In some areas of the USA, teaching degrees aren't required to actually teach. I hope I don't see this world-wide.

[–] [email protected] 7 points 3 months ago

Unfortunately this trend is happening in the States even without the AI buzzwords (though it is there). You give every kid a tablet with educational apps that feed into a curriculum algo. Teachers are told by the algo which student needs help on what, basically they become facilitators to the app. Then you also have "student summarizers" which will "analyze" a student written or audio submission and flatten it down to some unform stats.