A) Search past HN comments on hn.algolia.com, or
B) Post a new 'Ask HN'.
LLMs could provide a new way to find answers within a corpus. These have been described elsewhere, e.g.
- https://github.com/openai/openai-cookbook/blob/main/examples...
- https://news.ycombinator.com/item?id=34477543
I keep expecting someone (maybe minimaxir or simonw?) to post a 'Show HN: Get your question answered by the collective wisdom of HN', but I no one has so far (unless I missed the submission?).
Is someone already working on this?
AFAIK there's no GPT-3-like LLM that's easy to run at home, because the number of parameters is so so large. Your gaming PC's GPU won't have enough RAM to hold the model. For example, gpt-neox-20b needs about 40GB of RAM: https://huggingface.co/EleutherAI/gpt-neox-20b/discussions/1...
With optimization, I have it down to 140GB of ram. Trying to get it under 120GB without loosing too much accuracy so it can be ran on standard desktop consumer hardware (who's limits are usually 128GB).
Given the lack of resources I have found I figured the general intrest was low? Maybe I will open source it and do a write up.
Don't get me wrong: the idea could be nice but... ain't it time to think twice about all this before applying the last technological fad ?
To the sibling comment that I asked about doing this locally: there’s really no need for an LLM, much less for GPT-3. All you need is, well, attention. Sentence-transformer embeddings. Perhaps even just fastText.
I'm not sure how we'd pack enough context into a single 'sentence', to get a useful embedding for this purpose.
(I might be wrong of course.)
[0]https://github.com/HackerNews/API
https://github.com/ashish01/hn-data-dumps