this post was submitted on 10 Sep 2024
503 points (96.8% liked)

Programmer Humor

19175 readers
952 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS
 
top 47 comments
sorted by: hot top controversial new old
[–] [email protected] 69 points 1 week ago (3 children)

So much of the drift has just been marketing. Rebranding a Markov Chain stapled onto a particularly large graph as Master Computer from Tron.

[–] [email protected] 3 points 6 days ago (1 children)

I'm not sure what people think AI was ever going to be... every time something new comes out it's always dismissed because "it's basically just a X that does Y". I think that will continue to be the case until there is some literal connection to actual brains, in which case the concept of what a brain is will probably be questioned as well.

[–] [email protected] 1 points 6 days ago

I’m not sure what people think AI was ever going to be…

The heavy investment in AI is coming under the assumption that these advanced processes will replace huge portions of the human workforce.

So we don't need lawyers, because we just put prompts into a Law AI and it gives us a verdict. We don't need doctors, because we just put symptoms into a Medical AI and it gives us a diagnosis and treatment plan. We don't need salespeople, because we just put the product into a Marketing AI and it spits out a bunch of comvincing ad copy.

the concept of what a brain is will probably be questioned as well.

We already connect our brains to our computers. We just use screens and keyboards as our interface.

I suppose you could argue that a guy with a calculator or a camera or a chat app is mentally different than one without it. But I think the goal with AI is supplementing human minds, not complementing then.

[–] [email protected] 8 points 1 week ago

Fr I was about to say the same thing. Aside from better hardware and more layers the technology really hasn't changed from this level to begin with.

We've learned a little about what emergent behavior and trends look like in machine learning algorithms when graphed, though: it becomes more and more convergent, as if it forms its own little confirmation bias it will produce more and more samey results.

[–] [email protected] 5 points 1 week ago* (last edited 6 days ago)

Rebranding a Markov Chain stapled onto a particularly large graph

Could you elaborate how this applies to various areas of AI in your opinion?

Several models are non-markovian. Then there are also a lot of models and algorithms, where the description as or even comparison to Markov-chains would be incorrect and not suitable.

[–] [email protected] 35 points 1 week ago (3 children)

Sorry for the potato photograph, my phone was a potato, I was a student after all

[–] [email protected] 18 points 1 week ago (1 children)

Our phones back then were actual potatoes and we wore them next to the turnip on our belt, as was the fashion back then.

[–] [email protected] 4 points 1 week ago

That was before the dicketies.

[–] [email protected] 3 points 1 week ago

A propos potato cameras (i recommend checking out the channel):

https://youtu.be/BlWyKKJF0r4

[–] [email protected] 1 points 1 week ago

Looks more like an egg to me

[–] [email protected] 33 points 1 week ago (2 children)

As a kid I used to check out books from the library that had little BASIC games you could transcribe into your PC. Times have certainly changed.

[–] [email protected] 14 points 1 week ago (1 children)

I'm literally on an internship training course where the Exercises Left For The Readers are implementing Number Guessing Games on the various technologies talked about on the course. I'm like "thanks, but I read about this particular exercise extensively the BASIC age. I'm not going to redo these things unless your training material will have little cartoon robots. Like, you know, in the Usborne books or something."

[–] [email protected] 2 points 1 week ago* (last edited 1 week ago) (1 children)

Why would an internship train you on an antiquated language that hasn't been used in decades?

[–] [email protected] 3 points 1 week ago (1 children)

Well they train me in JavaScript frameworks and such. I allege this knowledge will be useless in a few decades. Or even less so, based on my meagre knowledge so far.

[–] [email protected] 5 points 1 week ago (1 children)

JavaScript is used on virtually every website on the planet today though. BASIC hasn't been used for anything in like 40 years.

[–] [email protected] 4 points 1 week ago (1 children)

20-ish, rather. You can still find some legacy systems running some version of VisualBasic on Windows.

Also disagree with op, javascript is the current "lingua franca" of programming. Unless every browser decides to allow scripting in a less shotgun-your-foot language, javascript will remain widely used.

[–] [email protected] 1 points 1 week ago (1 children)

Unless every browser decides to allow scripting in a less shotgun-your-foot language, javascript will remain widely used.

It's called web assembly, and all the major browsers (Safari, Chrome, Firefox, Edge) now support it. That's not to say that Javascript is going to disappear, but other languages might take over much of its marketshare.

[–] [email protected] 1 points 6 days ago

Even wasm relies on javascript and frankly, I see its existence as a failure of many layers. Machine code -> Operating system -> Browser -> WASM (emulated machine code)

[–] [email protected] 7 points 1 week ago (1 children)

Similar with the computer magazines, before they started coming with floppy disks.

[–] [email protected] 4 points 1 week ago

me proudly showing off to my dad that I had spent hours teaching the Timex Sinclair to... balance a checkbook!

dad: my checkbook is already balanced.

me: ahhh, yes well, just imagine though if things had lined up huh??

[–] [email protected] 32 points 1 week ago (1 children)

You can tell just from the font that this book is from the 80s

[–] [email protected] 32 points 1 week ago (2 children)

The font is Revue! People often say that their first love-hate font was Comic Sans - well, this was the first font I thought was pretty damn cool and I saw it getting run to the ground with overuse in early 1990s. It was pretty much in half of the ads in early 1990s. (My theory: It was bundled with a popular graphic design passion package / clipart bundle, Arts & Letters, and everyone made their ads with it. I can't wait for the day when I finally get arsed to install Windows 3.0 environment and my copy of Arts & Letters and prove the doubters wrong)

I half expected the first comment about the font to be about The Room to be honest.

[–] [email protected] 1 points 6 days ago* (last edited 6 days ago)

Randomly stumbling upon a font nerd was not on my bingo card today. I didn't know about Revue, thanks.

[–] [email protected] 15 points 1 week ago

Thanks! I was racking my brain trying to think of where I knew it from, and after seeing the page that you linked I'm almost certain that it's After Burner that is causing my brain to trigger the 80s association.

[–] [email protected] 23 points 1 week ago

Its object oriented programming. The object is to get investors $$$

[–] [email protected] 19 points 1 week ago* (last edited 1 week ago) (3 children)

Wait does it teach ai programming in BASIC?

10 dim prompt
20 input(prompt)
30 pring "Nice to meet you"+prompt
40 play g3 c4 e4
50 goto 20
[–] [email protected] 5 points 1 week ago
Syntax error on line 30
Please insert Pringles verification can to use pring function 
[–] [email protected] 4 points 1 week ago* (last edited 1 week ago) (1 children)

I'm in middle of a Rust module of a course, so I'll do some Programmer Friendly Error Messages:

Line 10: You do not need to dimension a dimensionless variable such as a standalone string variable. (This ain't Visual Basic.)
Line 20: input doesn't do parentheses, sorry
Line 20: Input accepts a string: Perhaps you meant prompt$?
Line 30: Concatenation is too modern, perhaps instead of + you meant ; just saying?
Line 40: Invalid syntax with play, maybe you meant play "g3c4e4"?

[–] [email protected] 1 points 1 week ago

Oh my! Haha all these modern languages poisoned my memory

[–] [email protected] 1 points 1 week ago

What does pring do?

[–] [email protected] 13 points 1 week ago (1 children)

Fun fact, the Perceptron is basically the first machine learning AI, and it was invented in 1943. It took a long time and many advancements in hardware before it became recognizable as the AI of today, but it's hardly a new idea.

[–] [email protected] 3 points 1 week ago (1 children)

People only discovered that multi-layer non-linear neural networks work at the 90s. It's not really reasonable to equate perceptrons with the stuff people use today.

[–] [email protected] 3 points 1 week ago

Obviously there have been major improvements over the past 80 years, but that's still considered the first neural network. The need for multi-layer neural networks was recognized by 1969, but the knowledge of how to do that took awhile to be worked out.

[–] [email protected] 12 points 1 week ago

you could say it was very basic AI

[–] [email protected] 12 points 1 week ago* (last edited 1 week ago) (1 children)
[–] [email protected] 1 points 1 week ago

Not longer in structured basic fortunately.

[–] [email protected] 10 points 1 week ago (1 children)

I ran ELIZA on a Timex/Sinclair 2068.

[–] [email protected] 14 points 1 week ago (1 children)

Why do you think you wanted to run ELIZA on a Timex/Sinclair 2068?

[–] [email protected] 2 points 1 week ago

Tell me about "Why do you think you wanted to run ELIXA on a Times/Sinclair 2968?".

[–] [email protected] 9 points 1 week ago (1 children)

oh god, you're making me think of the difference between a procedurally generated level and a randomly generated one.

The fucking generative AIs don't even have anything to stop it spewing crap

[–] [email protected] 2 points 1 week ago (1 children)

wave function collapse is a vastly superior procedural generation technique to generative pretrained transformer

[–] [email protected] 2 points 1 week ago

Same for enemy AI, for which I'm planning to implement some fuzzy logic support into my game engine, if there's no solution for it already in some D library.

[–] [email protected] 9 points 1 week ago (1 children)

I remember that book! Wasn't it basically, how to make your own Eliza with a bunch of If...Then's?

[–] [email protected] 5 points 1 week ago (1 children)
[–] [email protected] 12 points 1 week ago (1 children)

My lukewarm take is that all this hate on AI is really just hating on big companies trying to squeeze money out of some new development. AI (or rather LLMs) are not the problem.

[–] [email protected] 4 points 1 week ago

I personally like AI in its current state, still goofy and not a valid replacement for people. I definitely don't like that all of these corporations are justifying their billions in spending by forcing it into any and every app they can think of!

Why do I need Gemini to suggest a list of items I'm gonna need in my notes app that I likely opened to create a list of things I knew I already needed?