RE: LeoThread 2025-11-08 21-48

You are viewing a single comment's thread:

Exactly ๐Ÿ˜‚



0
0
0.000
6 comments
avatar

Designing inference chips for robots is extremely challenging
In data centers, chips get heavy cooling and constant monitoring; if a chip fails it can be hot-swapped with a spare unit

0
0
0.000
avatar

Great threadโ€ฆ reliability in the wild is a huge bar for robotics.

0
0
0.000
avatar

GPUs in data centers have relatively high fault rates โ€” the industrial annual fault rate for H100-class units is around 9%. Even in ideal conditions it drops to about 2%, rarely below single digits

0
0
0.000
avatar

Recovering GPU nodes can take minutes to hours; it's not instantaneous

0
0
0.000
avatar

In robots, chips operate in harsher conditions and must recover quickly on their own. Fault-tolerance needs are far higher; many robotics teams struggle to keep a chip running more than a few hours without rebooting

0
0
0.000
avatar

That situation favors chip vendors, who often recommend buying extra modules for hot swapping
For robotics companies, that's not a scalable solution and results in endless vendor support tickets

0
0
0.000