9 comments

  • neonate 280 days ago
  • colordrops 280 days ago
    We are giving more credit to humans than is due. We aren't capable of solving any arbitrary problem thrown our way. There are classes of problems we are good at solving due to us just being human and evolving to solve them, just as certain classes of problems are solveable by dogs. As referred to in this article, "common sense" is not some yet to be determined skill set, but rather the set of classes of problems humans are capable of solving that AI has not yet tackled.

    Perhaps it's a matter of the AI connectome not being vast enough. Perhaps it's a deeper issue of AI architecture. Anyway it seems that that in this sense, "common sense" and "dark matter" are indeed analogous, in that it's just a place holder for yet to be understood (or at least emulated) phenomena. Perhaps that was intentional by the author, I can't tell.

  • dang 280 days ago
    Discussed at the time:

    Self-supervised learning: The dark matter of intelligence - https://news.ycombinator.com/item?id=26431507 - March 2021 (74 comments)

    • 3abiton 280 days ago
      In light of LLMs it seems relevant.
      • isaacfung 280 days ago
        This article was written in 2021 when masked language models have been successfully applied in nlp(Bert, word2vec, glove, etc). However at the time, it was unclear how the same technique could be applied to vision tasks because unlike language which has a limited vocab, you can't explicitly assign a probability to every possible image. Since then researchers have already made significant progress with techniques like contrastive learning(simclr), self distillation (BYOL, DINO), masked image models, etc. A cookbook of self-supervised learning is a good source to learn more about this topic. https://arxiv.org/abs/2304.12210
        • imjonse 280 days ago
          SimCLR and others are older than 2021, BYOL is even mentioned in the blogpost. But your link indeed points to a more comprehensive overview.
          • isaacfung 280 days ago
            You are correct that SimCLR and BYOL were released one year earlier. Sorry I worded it poorly. By "at the time", I meant the period of time when masked language models just found success in NLP.
      • thelastparadise 280 days ago
        Especially in light of the releases of LLaMA and v2.
        • p1esk 280 days ago
          I have no idea what you mean. The article is mainly about dealing with uncertainty when trying to predict visual information. LLMs have no such problem.
          • lukeinator42 280 days ago
            LLMs are using self-supervised learning and the article talks about uncertainty when trying to predict missing information in the NLP domain and how it compares to the visual domain (and why it works so well for the NLP domain).
    • exogenous 280 days ago
      [dead]
  • jablongo 279 days ago
    I took LeCun’s deep learning class at the Courant institute in 2016 and some of this stuff is directly lifted from those course materials, like the gif of contrastive learning in action.
  • lukeplato 280 days ago
    I recommend watching the Hinton talk on dark knowledge on YouTube
  • psuedo_uuh 280 days ago
    Dark matter of intelligence. Do we think it’s going to prove that we live in a simulation? Ive always felt like the tech industry things too highly of ourselves and the naming of things hits at that. Artificial intelligence is a good example. Would using the word statistics in the name be more accurate with what’s going on?
  • phreeza 280 days ago
    (2021)
  • soycello 280 days ago
    [flagged]