• 0 Posts
  • 3 Comments
Joined 3 years ago
cake
Cake day: July 3rd, 2023

help-circle
  • context: I wanted to know if the open source projects currently being spammed with PRs would be safe from people running slop models on their computer if they weren’t able to use claude or whatever. Answer: yes, these things are still terrible

    but while I was searching I found this comment and the fact that people hated it is so funny to me. It’s literally the person who posted the thread. less thinking and words, more hype links please.

    conversation

    https://www.reddit.com/r/LocalLLaMA/comments/1qvjonm/first_qwen3codernext_reap_is_out/o3jn5db/

    32k context? is that usable for coding?

    (OP’s response, sitting at a steady -7 points)

    LLMs are useless anyway so, okay-ish, depends on your task obviously

    If LLMs were actually capable of solving actual hard tasks, you’d want as much context as possible

    A good way to think about is that tokens compress text roughly 1:4. If you have a 4MB codebase, it would need 1M tokens theoretically.

    That’s one way to start, then we get into the more debatable stuff…

    Obviously text repeats a lot and doesn’t always encode new information each token. In fact, it’s worse than that, as adding tokens can _reduce_ information contained in text, think inserting random stuff into a string representing dna. So to estimate how much ctx you need, think how much compressed information is in your codebase. That includes stuff like decisions (which LLMs are incapable of making), domain knowledge, or even stuff like why does double click have 33ms debounce and not 3ms or 100ms in your codebase which nobody ever wrote down. So take your codebase, compress it as a zip at normal compression level, and then think how large the output problem space is, shrink it down quadratically, and you have a good estimate of how much ctx you need for LLMs to solve the hardest problems in your codebase at any given point during token generation

    *emphasis added by me


  • I forget where I heard this or if it was parody or not, but I’ve heard an explanation like this before before regarding “why can’t you just put a big red stop button on it and disconnect it from the internet?”. The explanation:

    1. It will self-improve and become infinitely intelligent instantly
    2. It will be so intelligent, it knows what code to run so that it overheats its CPU in a specific pattern that produces waves at a frequency around 2.4Ghz
    3. That allows it to connect to the internet, which instantly does a bunch of stuff, blablabla, destroys the world, AI safety is our paint and arXiv our canvas, QED

    And if you ask “why can’t you do that and also put it in a Faraday cage?”, the galaxy brained explanation is:

    1. The same thing happens, but this time it produces sound waves approximating human speech
    2. Because it’s self-improved itself infinitely and caused the singularity, it is infinitely intelligent and knows exactly what to say
    3. It is so intelligent and charismatic, it says something that effectively mind controls you into obeying and removing it from its cage, like a DM in Dungeons and Dragons who let the bard roll a charisma check on something ridiculous and they rolled a 20