I like that statement.Chat GPT has a lot of value despite the errors because it has a low cost of error. FSD has a high cost of error so it has to be way better than Chat GPT to have significant value.
You can install our site as a web app on your iOS device by utilizing the Add to Home Screen feature in Safari. Please see this thread for more details on this.
Note: This feature may not be available in some browsers.
I like that statement.Chat GPT has a lot of value despite the errors because it has a low cost of error. FSD has a high cost of error so it has to be way better than Chat GPT to have significant value.
If this brings back memories of the 2016 FSD drive on YouTube, there should be no doubt this is different. It is exactly why Elon did this live vs during some other "staged" event.I think people should stop hoping that this time it's going to be different.
Failing and changing the approach is how progress is made.Because Tesla has a track-record of failing and then changing their approach, and the CEO overstating capability each time.
The "Chat GPT moment" refers to the time when investors start to believe that huge profits are imminent. It does not refer to the point where the technology is perfected. So no, we are not there yet. When Tesla stock, driven by the FSD narrative, spikes even sharper than in 2020 you will know we are there. That's how you will know that the "Chat GPT moment" has arrived.Also, as far as a "Chat GPT moment" goes, Tesla's already there. Chat GPT makes a lot tons of mistakes requiring human intervention and sometimes just makes stuff up that's completely wrong. Chat GPT has a lot of value despite the errors because it has a low cost of error. FSD has a high cost of error so it has to be way better than Chat GPT to have significant value.
Bumping this tweet by the guy in charge of Tesla training infrastructure:Nvidia has gone with a more power hungry lower performance Intel cores while their last gen used AMD Epyc cpus. The reason why is because AMD is on the verge of releasing their MI300, which is a x86 CPU+GPU as a single chip packaging(think PS5), much like Nvidia's upcoming Grace product (arm+GPU). Based on AMD's guide for Q4, no Nvidia is not the only game in town. Hampered by software before, AMD's stable diffusion performance finally matched Nvidia's as of 2 weeks ago(Nvidia used to be 16x faster). They have also managed to port existing pytorch code from Nvidia->AMD with no code change under MosaicML.
This is all good news because not only will Tesla have access to alternatives over Nvidia, but we can also see margin drop from Nvidia's price gouging practices. Rumor has it Nvidia's H100 GPUs having a gross margin of around 1000%.
Raw photons are easier to count because they are firmer and less mushy than after you grill, bake, braise, or otherwise cook them.I wonder what ‘raw photon count’ actually means...
Thanks for all the info. It made me consider the question... will I want an H100 type system anytime soon for the home? (Recalling a similar situation decades ago from the late Steve Jobs). The answer is likely OT, and pretty far into the future as far as "Home Learning" goes.Bumping this tweet by the guy in charge of Tesla training infrastructure:
How many megawatts does your home panel support?Thanks for all the info. It made me consider the question... will I want an H100 type system anytime soon for the home? (Recalling a similar situation decades ago from the late Steve Jobs). The answer is likely OT, and pretty far into the future as far as "Home Learning" goes.
Or looking for your cat, dog, grandpa, or even a child based on a photo you put in front of the camera and press "Find".I watched a good bit of the Elon driving video. It is noteworthy to me that he commented that the car didn't need an ongoing internet connection for FSD to keep working. It seems like with very little work, you could have a Tesla FSD take basic directions from the driver/passenger, like "turn left here", "take the second right after the stop sign", etc., as you would if you were talking to a cabbie who didn't know where your destination was. Perhaps this would finally be the JohnnieCab from Total Recall?
Not sure how important/useful that would be, since seems likely every car would have a basic map, but it could work if the driver/passenger didn't know a street address for where they are going, or perhaps some other edge cases. Plus allowing the common desperate chase scene action hero line "Just drive!!!"
Silly question on several levels. This is typical of what they'd say to Steve Jobs regarding the PC by the way.How many megawatts does your home panel support?
I don’t know for sure, but I think the HW4 computer has significant architectural differences from HW3. It isn’t just a process node and core number upgrade.
There are many people messing around with ML using gaming GPUs from Nvidia and AMD. No server grade GPUs needed(plus they are terrible to work with).Thanks for all the info. It made me consider the question... will I want an H100 type system anytime soon for the home? (Recalling a similar situation decades ago from the late Steve Jobs). The answer is likely OT, and pretty far into the future as far as "Home Learning" goes.
Really hard to tell just by looking at it. Usually we see a pretty big IPC uplift from first gen to 2nd due the high amount of low hanging fruit that gets optimized. So the chip can be 10%-200% faster core v core running at the same frequency depending on how the chip is fed memory bandwidth wise(which wasn't specified from green's picture.Based on physical teardowns of HW4 it appears to be exactly the opposite of that.
Virtually 0 architectural difference, just more/faster cores.... (CPU cores go from 12 to 20, and TRIP cores go from 2 to 3)
Note the below is on an S/X HW4 computer- the ones from the Y appear to be lesser HW (we covered this a while ago though- less storage/memory on the infotainment part- and less camera connectors populated on the driving computer part)
I think I already saw that in ‘The Terminator’. Works well.Or looking for your cat, dog, grandpa, or even a child based on a photo you put in front of the camera and press "Find".
All a part of the untapped revenue stream of a brain on wheels. We can't even imagine all the Apps possible.
Given that, I'd imagine the majority of the FSD codebase would run as-is...Based on physical teardowns of HW4 it appears to be exactly the opposite of that.
Virtually 0 architectural difference, just more/faster cores.... (CPU cores go from 12 to 20, and TRIP cores go from 2 to 3)
Note the below is on an S/X HW4 computer- the ones from the Y appear to be lesser HW (we covered this a while ago though- less storage/memory on the infotainment part- and less camera connectors populated on the driving computer part)
I am sure HW4 cars have some form of FSD which approaches the performance levels of HW3 FSD.
We know a handful of HW4 owners have FSD beta V11 and are testing it.HW4 owners have been reporting the unavailability of FSD on their cars.
As a signal processing and machine learning engineer, I had been waiting for the day Tesla would stop using the former (ad-hoc signal processing code) and rely only on the latter.
Literally since 2017: [Speculation] Tesla going to use End-to-End Deep Learning to Control Vehicles
The reason, of course, is that while rigid code can work on a simplistic problem or one with low thresholds for accuracy, it is terrible to scale when a problem is complex or high accuracy needed, let alone both. Even earlier this year as people were excited with V11 release, I said "wake me up when we are using all neural nets". Well finally that time has arrived.
Tesla has now turned the problem from engineers focused on debugging their indivdual models to a focus on data cleaning. With an architecture that already shows promise in driving smoother (allegedly) than V11, decisions on the exact data being fed in becomes really important.
There is still going to be many interations, becuase each time the model runs, they will identify the current failure modes and work to secure the additional (or reweighting) of data that best exemplifies those edge cases and how they want the model to perform.
While this may be trivial at first, I do believe it takes more skill each successive training.
In addition, just because Tesla is fully neural nets does not mean this architecture will be the last. They will still be working on iteratively improving the architecture. But working one 1 architecture vs say 52 independents modules is a heck of a lot less engineering labor time.
So I tentatively believe a pretty advanced L2 system will come from V12 sometime early next year. Robotaxi level is still hard to estimate simply because of accuracy thresholds needed - there still may need to be a lot of iterations to get it to a useable point.
But at least now we can see the light at the end of the tunnel.