Do we have a critical mass of GPU owners to train a legitimate LLM that could compete with commercial ones?

Posted by decentralize999@reddit | LocalLLaMA | View on Reddit | 40 comments

I discussed with Claude the idea of training a legitimate LLM in a decentralized way using an uncensored 20TB dataset. It recommended a 300B parameter model with a 10M token context size. To train such an LLM, participants (nodes) would need at least 4 RTX Pro 6000 cards if using the DiLoCo training approach.

To summarize my discussion with Claude, here is what is required:

3,000 nodes (owners with 4 RTX Pro 6000 cards)

Duration: 2.5 months

Daily network traffic about 1.7TB per node for syncing checkpoints, etc.

Around $666 total per node for electricity and internet costs, assuming $0.15/kWh

Assuming there are 300,000 people who already own 4 such cards (or are close to it), and even 1% of them would be willing to donate their time and resources to train this LLM - this poll was created to find out.

View Poll