Read more.The Tesla FSD is "the best chip in the world… by a huge margin," claims Elon Musk.
Read more.The Tesla FSD is "the best chip in the world… by a huge margin," claims Elon Musk.
With time Musk claims start to appear less and less truthful. But I might be wrong.
Pleiades (23-04-2019)
Nvidia are pretty hurt by this. No wonder they want to compare their mega-bucks 500W self-driving Pegasus board with this <80W board.
Note that 144 TOPS includes the redundancy, it's 72 TOPS. However it may be that the board can use all 144 TOPS, and drop back to 72 TOPS in the case of a single chip failure.
And the 21x performance increase clearly targets a bottleneck in the Nvidia Drive Xavier solution that is outside of the NNP, as it's only about 3x more powerful there.
Tabbykatze (23-04-2019)
I'm surprised it's using ARM chips but I guess because of the reduced instructions set, it performs better? Probably wouldn't make sense to start a whole new division for just chips anyway as that would cost a lot and require a lot of expertise.
ARM is aimed at this sort of usage. Tesla can choose whatever level of cost/performance tradeoff they want and just buy a hard cell ready to place in their SoC layout. I'm not sure what else you think they could use, PowerPC is basically dead at this stage, MIPS don't have the same range of choice, RISC-V whilst one to watch would be rather new and risky when they are already going for it with the neural part.
Intel/AMD just aren't in the running for stuff like this, they are too big and power hungry.
Musk is most likely a narcissist but he has the slight advantage over Trump of having a reason to think that he's superior ... Trump is a slobbering moron in comparison.
There's also very little sense that Musk is a sociopath, which checklist Trump has plenty of ticks on.
Musk may have a touch of manic depression. Certain, er, "less than advisable" actions and outbursts on his part suggest at least hypomania.
All round, Musk is a good guy, which Trump is not, and, as with most divas, you have to cut him a lot of slack even if it grates to do so. :-)
That said, don't ask me to pledge any support for the "Musk for President" campaign, should it ever happen. )
From a position of total ignorance I was wondering - in ultrasound we've moved from dedicated hardware towards GPGPU stuff to do the same job. Requires an awful amount of oomph but the system can be upgraded far more by software updates than it could be when running on hardware beam formers, etc.
Would this kind of approach work for self driving cars? Just have a load of raw compute power in there, probably with a PSU to match and do it all using software, allowing for OTA / easy / cheap upgrades when things inevitably go wrong down the line and recalls happen?
AIUI that's what they're doing, just with a different flavour of silicon than GPUs. I don't believe there's much difference in the silicon between the neural networks doing DLSS in 2XXX series GPUs and the neural networks in the nvidia drive chips - the inferences are different enough from GPU work for there to be a big benefit to going custom silicon, but also general enough they can be applied wherever people want the "deep learning" buzzword (to share the R&D cost of the design of the neural network bits).
I'm interested to hear you have GPU powered ultrasound, for some years I have wondered why ultrasound scanners aren't a cheap(ish) USB peripheral knocked out by the million in China that can use a laptop graphics card given how compute. I'm sure driving an ultrasound scanner has it's complications and risks of false positives etc, it feels to me as an engineer that every GP should have one by now. Hopefully from what you are saying we are heading slowly down that path.
Anyway, back to your question...
Neural networks aren't really about computational power as the sums aren't that hard. The precision isn't that high, GPUs go for fp16 as the lowest they can do though fp16 was generally abandoned for desktop use in the early gpu generations for poor image quality so it has had to be re-introduced. Some systems I believe used fixed point getting down to 8 bit precision, again something that in a 64 bit world CPUs aren't usually optimised for. On top of that, the output from the multiply-accumulate maths that tallys up the neural input is then put through a non linear function like an S curve to get the final fire/not fire result. But the killer is memory access, just like in a living brain the results are not so much down to the simple processing that a neuron does but the massive connectivity of the dendrites linking to other brain cells. That is replaced with what comes down to a lot of matrix lookup.
So really whilst you can get perfectly decent results with a GPU, such low precision non linear arithmetic with massive memory bandwidth isn't something that fits any other compute job that well. Custom hardware seems a really good idea.
Note that this is for the mass market of deploying a trained network. The actual network training requires higher precision, a big GPU cluster with HBM is ideal for that.
I'm not gonna lie but I don't actually know much about chips etc. I know Intel and AMD can't perform well because they're not aimed at neural stuff as they are more general chips aimed at a wide range of applications.
I do however think it will be interesting to see a RISC-V neural processor if it's possible but I guess it does indeed make sense for Tesla to use ARM.
It isn't really that they are general purpose, Intel/AMD are based on the 8086 which was a rubbish processor when new. In some ways it was improved with the 386 and again with AMD64, but all that history drags a lot of transistors with it to remain compatible. At the very high end like we get in our desktop machines processors are so complicated anyway that the overhead is only something like 5% (according to Intel some years ago, so apply a big pinch of salt to that number). That is why the Atom chips were always doomed to failure in the tablet market, the likes of ARM will always be faster and lower power for a smaller cpu.
Edit: Note the cpu isn't doing any neural processing, just shuffling data to and from the neural unit and acting on results.
What I have seen of the video so far is actually rather good, though it is 4 hours long and I did skip to the first hour to get to the silicon description. Includes a primer for how neural networks can drive a car if people are interested in that.
"In a perfect world... spammers would get caught, go to jail, and share a cell with many men who have enlarged their penises, taken Viagra and are looking for a new relationship."
There are currently 1 users browsing this thread. (0 members and 1 guests)