50+ hot takes on the current and future of AI/ML

3 points by chse_cake 1 year ago | 1 comment
  • sci-genie 1 year ago
    TBH I kinda agree with the argument that distributed training is too hard. Its so architecture/compute-resources/network-topology dependent that when people open those can of worms, they quickly realize that the cost/benefit tradeoff is limited unless you are doing large-scale pre-training. its just so much easier to train as much as possible on a single node
    • 1 year ago