50+ hot takes on the current and future of AI/ML
3 points by chse_cake 1 year ago | 1 comment- sci-genie 1 year agoTBH I kinda agree with the argument that distributed training is too hard. Its so architecture/compute-resources/network-topology dependent that when people open those can of worms, they quickly realize that the cost/benefit tradeoff is limited unless you are doing large-scale pre-training. its just so much easier to train as much as possible on a single node
- 1 year ago