this post was submitted on 21 Oct 2024
392 points (96.4% liked)

Programmer Humor

32533 readers
587 users here now

Post funny things about programming here! (Or just rant about your favourite programming language.)

Rules:

founded 5 years ago
MODERATORS
 
all 34 comments
sorted by: hot top controversial new old
[–] [email protected] 62 points 1 month ago (2 children)

Don't copilot anything longer than a function of about 15 lines. That way you can quickly see if it made mistakes. Ensure it works, move on to the next.

And only do that for boring, repetitive work. The tough challenges and critical parts you're (for now) better off solving yourself.

[–] [email protected] 23 points 1 month ago (2 children)

Absolutely, I think the people who say it's completely useless for code are in denial

Definitely not replacing anyone but my god it has sped up development by generating code I already know how to write 90% of

No more having to look up "what was the for loop syntax in this language again?"

[–] [email protected] 18 points 1 month ago* (last edited 1 month ago) (1 children)

"Copilot is really good at things which I already know" and that is perfectly fine

[–] [email protected] 9 points 1 month ago (1 children)

Exactly.

It's to speed up boilerplate and save you having to look up function names or language specific syntax for that one feature you want to use, not to entirely do your job for you

[–] [email protected] 2 points 1 month ago (1 children)

If I've been working in the same language for at least a year or two, I don't have to look up any of that. Copilot might be actually helpful if I'm working in a language I'm not used to, but it's been a long time since I've had to look up syntax or functions (excluding 3rd party packages) for the language I work in.

[–] [email protected] 1 points 1 month ago (1 children)

Of course but presumably on occasion you do work in other languages? I work in all kinds of languages and so jumping between them it's pretty handy to bridge the gap

I think you could definitely still get value out of generating simple stuff though, at least for me it really helps get projects done quickly without burning myself out

For small one off scripts it makes them actually save more time than they take to write (for example colleague had to write the permissions of a bunch of files recursively into an excel doc, chatgpt did 90% of that I did 9 and he did 1 lol)

[–] [email protected] 1 points 1 month ago

Of course but presumably on occasion you do work in other languages? I work in all kinds of languages and so jumping between them it’s pretty handy to bridge the gap.

If I were jumping languages a lot, I definitely think it would be helpful. But pretty much 100% of what I've done for the last 3-4 years is Go (mostly) or JavaScript (occasionally). I have used chatgpt the few times I needed to work in some other language, but that has been pretty rare.

I think you could definitely still get value out of generating simple stuff though, at least for me it really helps get projects done quickly without burning myself out

If simple stuff == for loops and basic boilerplate, the kind of stuff that copilot can autocomplete, I write that on autopilot and it doesn't really register. So it doesn't contribute to my burnout. If simple stuff == boring, boilerplate tests, I'll admit that I don't do nearly enough of that. But doing the 'prompt engineering' to get copilot to write that wasn't any less painful that writing it myself.

For small one off scripts it makes them actually save more time than they take to write

The other day I wrote a duplicate image detector for my sister (files recovered from a dying drive). In hindsight I could have asked chatgpt to do it. But it was something I've never done before and an interesting problem so it was more fun to do it myself. And most of the one off stuff I'm asked to do by coworkers is tied to our code and our system and not the kind of thing chatgpt would know how to do.

[–] [email protected] 3 points 1 month ago (1 children)

I won't say copilot is completely useless for code. I will say that it's near useless for me. The kind of code that it's good at writing is the kind of code that I can write in my sleep. When I write a for-loop to iterate over an array and print it out (for example), it takes near zero brain power. I'm on autopilot, like driving to work. On the other hand, when I was trialing copilot I'd have to check each suggestion it made to verify that it wasn't giving me garbage. Verifying copilot's suggestions takes a lot more brain power than just writing it myself. And the difference in time is minimal. It doesn't take me much longer to write it myself than it does to validate copilot's work.

[–] [email protected] 1 points 1 month ago* (last edited 1 month ago) (1 children)

You can think bigger than that, as an example from the other day, I got it to a Display implementation for all of my error types in rust, it generated nice user friendly error messages based on context and wrote all the boilerplate around displaying them

Also got it to generate a function that generated a unique RGB colour from a user ID, did it first try and I could use it straight away

Both those things would've taken me maybe 15 minutes by hand but I can generate and proofread them in seconds

That said, I don't use copilot I use chatgpt, it's intentional when I use it not just being shoved in my face all the time which might help my opinion of it

[–] [email protected] 1 points 1 month ago
func randomRGB(uid int) color.RGBA {
	b := binary.BigEndian.AppendUint64(nil, uint64(uid))
	h := sha256.Sum256(b)
	return color.RGBA{h[0], h[1], h[2], 255}
}

That took me under three minutes and half of that was remembering that RGBA is in the color package, not the image package, and uint-to-bits is in the binary package, not the math package. I have found chatgpt useful when I was working in a different language. But trying to get chatgpt or copilot to write tests or documentation for me (the kind of work that bores me to death), doing the prompt engineering to get it to spit out something useful was more work than just writing the tests/documentation myself. Except for the time when I needed to write about 100 tests that were all nearly the same. In that case, using chatgpt was worth it.

[–] BudgetBandit 9 points 1 month ago (2 children)

Tried to learn coding using chatGPT. Wanted to make my own game engine for a phone game. Ended up looking up tutorials.

[–] Bread 1 points 1 month ago

ChatGPT as a programming tool like any other tool works a whole lot better when you are well versed in how the process should go. It speeds up the workflow of a professional, it doesn't make a new worker better.

[–] coffee_with_cream 1 points 1 month ago

If you are using "game engine" in the industry standard way, you would want to learn object oriented programming first, then learn how to use an existing game engine, and then MAYBE, in a long time, with a big team, build your own game engine.

[–] [email protected] 22 points 1 month ago (1 children)

Ai is great for finding small flaws or reciting documentation in a more succinct way. But writing new code and functions? That's a fools errand hoping it works out

[–] [email protected] 10 points 1 month ago* (last edited 1 month ago) (1 children)

I use it for writing functions and snippets all the time, at least in python and rust as long as you describe what you want it to do properly it works great

Example I used recently: "Please generate me a rust function that will take a u32 user id and return a unique RGB colour"

Generated the function, I plugged it in and it worked perfectly first time

[–] [email protected] 6 points 1 month ago

To be honest yes. That is the sort of thing that sounds great. I have a little project I'm about to start so I'll take a look

[–] [email protected] 17 points 1 month ago (3 children)

I haven't been in development for nearly 20 years now, but I assumed it worked like that:

You generate unit tests for a very specific function of rather limited magnitude, then you let AI generate the function. How could this work otherwise?

Bonus points if you let the AI divide your overall problem into smaller problems of manageable magnitudes. That wouldn't involve code generation as such...

Am I wrong with this approach?

[–] [email protected] 8 points 1 month ago

At that point you should be able to just write the code yourself.

The A"I" will either make mistakes even under defined bounds, or it will never make any mistakes ever in which case it's not an autocomplete, it's a compiler and we've just gone full circle.

[–] [email protected] 7 points 1 month ago (1 children)

The complexity here lies in having to craft a comprehensive enough spec. Correctness is one aspect, but another is performance. If the AI craps out code that passes your tests, but does it in really inefficient way then it's still a problem.

Also worth noting that you don't actually need AI to do such things. For example, Barliman is a tool that can do program synthesis. Given a set of tests to pass, it attempts to complete the program for you. Synthesis is performed using logic programming. Not only is it capable of generating code, but it can also reuse code it's already come up with as basis for solving bigger problems.

https://github.com/webyrd/Barliman

here's a talk about how it works https://www.youtube.com/watch?v=er_lLvkklsk

[–] [email protected] 2 points 1 month ago (1 children)

So it's like AI, but tailored for one purpose and without the marketing

[–] [email protected] 4 points 1 month ago

also doesn't require burning down a rain forest each time you run a query

[–] [email protected] 6 points 1 month ago (1 children)

I tend to write a comment of what I want to do, and have Copilot suggest the next 1-8 lines for me. I then check the code if it's correct and fix it if necessary.

For small tasks it's usually good enough, and I've already written a comment explaining what the code does. It can also be convenient to use it to explore an unknown library or functionality quickly.

[–] [email protected] 3 points 1 month ago (1 children)

"Unknown library" often means a rather small and sparely documented and used library tho, for me. Which means AI makes everything even worse by hallucinating.

[–] [email protected] 4 points 1 month ago

I meant a library unknown to me specifically. I do encounter hallucinations every now and then but usually they're quickly fixable.

It's made me a little bit faster, sometimes. It's certainly not like a 50-100% increase or anything, maybe like a 5-10% at best?

[–] coffee_with_cream 6 points 1 month ago (1 children)

I told it to generate a pretty complex react component and it worked on the first try yesterday. It even made a style sheet. And it actually looks good.

[–] [email protected] 2 points 1 month ago* (last edited 1 month ago)

It's so good when it works on the first try. But when it doesn't work it can really fool people with totally nonfunctional code. AI is the genie in the bottle where you really need to ask the right question.

[–] [email protected] 4 points 1 month ago

Sloppy joes is the new spaghetti code