I was under the impressions cpu's were very sensitive to radiation. If we could mine and manufacture in space I could see this maybe.
Technology
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related news or articles.
- Be excellent to each other!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
- Check for duplicates before posting, duplicates may be removed
- Accounts 7 days and younger will have their posts automatically removed.
Approved Bots
It’s not speculation. Nvidia themselves have run experiments with GPUs in orbit, and the issue gets worse with smaller lithography (eg newer chips).
Spacecraft software engineer here:
They are and they aren't. Radiation causes problems in terms of Single Event Upsets where a 0 turns to 1 and a 1 turns to 0 for a super tiny second. CPUs take some amount of time to let the transistor circuit stabilize before moving onto the next instruction so if an SEU happens in the beginning of this period it won't have any downstream effects. Like a bump on the road.
Memory however is vulnerable to this tiny amount of time and can flip a bit to a different state than it's supposed to be, but both are solvable problems with hardware and software based solutions, with ECC being the most common.
The other major problem is Total Ionizing Dose. Put silicon based semiconductors in radiation long enough and they will break down, and there's no real hardware or software based solution to that. But it takes a long time