After KrispyKreme's technical assault, I decided I Need to understand the electronics that handles the ai. So I started the search 2 days ago and finally found the thread with AP circuit board teardown and subsequent discussion today. Then took a deep dive into Nvidia's Parker (TA795SA-A2) general purpose gpu architecture and also the discreet Pascal GPU (GP106-510-KC) used. It was a deep dive, but I came out of it understanding that it is possible for Tesla to drop in a chip and enable Full Self driving. I will try to be as non technical and easy to understand as possible.
There were many experts chiming in both on this forum (for the sake of the person doing the teardown, will not link to it) and on Reddit's thread. However, not many have all the specialties required to see the full picture. I happen to have been an ASIC designer, firmware engineer, Machine Vision Application Engineer and in a jam, my managers have forced me to solder and troubleshoot PCBs so I know a bit of that. The only discipline I do not have is chip layout/routing and PCB layout/routing. (yes I jump ship every 3 years, jack of all trade and not an expert in any)
Here's my guess of how they use the hardware from looking at just the architecture.
Nvidia's Parker general purpose GPU is probably what Tesla intends to replace. Contained within is a multipurpose ARM Cpu and a small piece of its Discreet GPU. These GPUs have a bunch of units that process pixels in parallel and shove them into what they call Tensor units. Tensors units are programmable units that can be changed for other operations. I believe that these are used as the "Neurons" in neural network for decision making.
Some GPU parts not needed for AI
Nvidia, probably in a rush to bring a product to market, did not really design a GPU specifically for Neural Network, instead they brought their gaming discreet gpu and stuffed it with some control logics and called it a day. The Processing part of their GPU has a lot of waste. Things specifically made for gaming and displaying images can probably be taken out completely. Also all the calculations can probably be narrowed down to int8 once the neural network is sufficiently trained and the results can be locked down. For simple pre-processing, drawing contours and recognizing an object in an image. a 8 bit black and white image will suffice and TSLA uses the red spectrum to do that. Currently, many calculations passes through 16 bit floating points and 32 bit floating points (most likely necessary for training the neural network). These are very expensive operations and take up a lot of spaces.
AP 2.0 Model S, AP 2.0 Model 3, AP 2.5 Model 3
Model S tear down shows that it has 1 Parker general purpose and 1 discreet gpu. Model 3 tear down from Munroe shows 1 Parker general purpose and 2 discreet gpu. So there's some upgrade there. Then there's a potential AP 2.5 hardware showing up where extra connectors to a potential new board are present. In the future, there may be two board linked together to perform Full Self driving. My own guess is that it will evolve to two of the Nvidia board in parrallel for a while before Tesla have finished designing and testing their own chip. Both of the chips seems drop in replaceable as they both seems to sit on MXM. My guess is that there's a strong chance TSLA replaces both the General purpose and Discreet GPU.
I personally haven't done any calculation on how many chips is necessary to achieve full self driving. Potentially AP 2.0 Model S might have some trouble, but Tesla can just tapeout a chip with 7nm process for the older Model S which effectively doubles the amount of operations it can do. But this probably won't be possible if we stick to Nvidia solutions for the interim.
COST
However, the question is, is it worth it. Nvidia is forging ahead with their new plex 2.0 platform. More power consumption more heat generated, but it'd save TSLA the R&D cost as well as the design and tape out cost + equipment for analysis. I wouldn't be surprised that it will cost 1 mil or 2 to tapeout the first batch of chip (plus ~3$ a chip) and the chip design team of around 20 person x 100k salary (probalby 200k + if TSLA is using the best) x 2 years of time + equipment. On top of that, Taping out is a very rigid process and cannot be modified too much on the fly. There's only so many Engineering Change order you can do before the mass production.
What is so bad with sticking with NVidia and just increase cooling and power consumption, which was the reason why KrispyKreme was attacking Tsla's automated driving electronics. That reasoning is lost to me.