this post was submitted on 11 Nov 2023
1 points (100.0% liked)
Hardware
47 readers
1 users here now
A place for quality hardware news, reviews, and intelligent discussion.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
How much does having the best performing hardware actually matter in AI?
not much, garbage (quickly) in, garbage (quickly) out, the training data set and novel techniques matter more than how quickly it's processed after a certain point.
I work in the field and I don't know what the others here are saying. In practice the difference in performance doesn't matter, we literally just launch a bunch of jobs on the cluster, and if they finish faster it doesn't matter because I either haven't checked my jobs, or I'm not ready to collect the results yet, or I'm just dicking around with some experiments or different parameters so the actual speed of completion doesn't matter. A bunch of weaker GPUs can do the same task as a stronger one, only memory really matters. Doubly true if the company is big enough that power consumption is a drop in the bucket in terms of operational costs.
What actually matters is the overall workflow, stuff like the cluster having downtime is way more impactful to my work than the performance of a GPU, or the ease of designing and scheduling jobs/experiments on the cluster.
Also in the end all of this is moot, this type of training is probably the wrong approach to AI. Note that a child does not need a million images of a ball to recognize a ball, and it would instantly be able to recognize soccer balls, basket balls, etc as all balls after learning one. The way we train our AIs cannot do this, our current approach to AI is just brute force.
When you are able to choose between 1 rack of latest Nvidia products vs 3 racks of last gen to get results in the same time frame, you find that the performance matters a lot