Skip Navigation

  • nobody asked you to post in this thread. you came and posted this shit in here because the thread is very popular, because lots and lots of people correctly fucking hate generative AI

    so I guess please enjoy being the only “non-disingenuous” bootlicker you know outside of work, where everyone’s required (under implicit threat to their livelihood) to love this shitty fucking technology

    but most of all: don’t fucking come back, none of us Luddites need your mid ass

  • no thx

  • exactly. in Doom Emacs (and an appropriately configured vim), you can surround the word under the cursor with brackets with ysiw] where the last character is the bracket you want. it’s incredibly fast (especially combined with motion commands, you can do these faster than you can think) and very easy to learn, if you know vim.

    and I think that last bit is where the educational branch of our industry massively fucked up. a good editor that works exactly how you like (and I like the vim command language for realtime control and lisp for configuration) is like an electrician’s screwdriver or another semi-specialized tool. there’s a million things you can do with it, but we don’t teach any of them to programmers. there’s no vim or emacs class, and I’ve seen the quality of your average bootcamp’s vscode material. your average programmer bounces between fad editors depending on what’s being marketed at the time, and right now LLMs are it. learning to use your tools is considered a snobby elitist thing, but it really shouldn’t be — I’d gladly trade all of my freshman CS classes for a couple semesters learning how to make vim and emacs sing and dance.

    and now we’re trapped in this industry where our professionals never learned to use a screwdriver properly, so instead they bring their nephew to test for live voltage by licking the wires. and when you tell them to stop electrocuting their nephew and get the fuck out of your house, they get this faraway look in their eyes and start mumbling about how you’re just jealous that their nephew is going to become god first, because of course it’s also a weirdo cult underneath it all, that’s what happens when you vilify the concept of knowing fuck all about anything.

  • it could turn on the RGB! though that would imply that the RGB could be turned off in the first place, which is optimistic on my part

  • it’s really embarrassing when the promptfans come here to brag about how they’re using the technology that’s burning the earth and it’s just basic editor shit they never learned. and then you watch these fuckers “work” and it’s miserably slow cause they’re prompting the piece of shit model in English, waiting for the cloud service to burn enough methane to generate a response, correcting the output and re-prompting, all to do the same task that’s just a fucking key combo.

    Same with text formatting, for example. I regularly need to format long strings in specific ways, adding brackets and changing upper/lower capitalization. It does it in a second, and really well.

    how in fuck do you work with strings and have this shit not be muscle memory or an editor macro? oh yeah, by giving the fuck up.

  • I will not go into ethics.

    then stop wasting our fucking time

    I didn’t read the rest of your post but it’s vaguely LLM-shaped so off you fuck

  • you may or may not classify it as ‘Generative AI’ on it’s own.

    while the ship has sailed on calling the opencv shit you’re doing AI (thx, grifters of the first AI bubble), which part of object detection and facial recognition is generative?

  • also you’re right, I was unfair towards cubism_pitta. I wasn’t enough of an asshole.

    hey fuckers. next time you get the sudden, overwhelming urge to jack off in public about how much money you’re feeding into the machine that does plagiarism and nothing else, keep it the fuck off my instance.

  • okwhateverdude

  • ah yes the only way to make LLMs, a technology built on plagiarism with no known use case, “not silly” is to throw a shitload of money at Apple or framework or whichever vendor decided to sell pickaxes this time for more RAM. yes, very interesting, thank you, fuck off

  • ah yes the only way to make LLMs, a technology built on plagiarism with no known use case, “useful for any actual ai application” is to throw a shitload of money at nvidia. weird how that works!

  • I vaguely remember that one of the articles talking about the physics forum mentioned it happening elsewhere, but I haven’t dug into it myself. it might just be one or two shitty admins doing this, but I suspect (without evidence, I just can’t think of another reason to do it) there’s some party offering a financial incentive for them to go back and fuck up their old forums

  • I think you’re absolutely correct, and this feels to me like the only reason why we’re seeing some of the bizarre shit we’ve been keeping an eye on:

    • several old forums, all of which are unique high-quality data sources, are being polluted by their own admins with backdated LLM-generated answers. this destroys that forum as a trustworthy data source and removes it as competition for the LLM that already scraped the forum — and, as a bonus, it also makes training a future LLM on that data source utterly impractical without risking model collapse.
    • Wikipedia refuses to compromise on quality in general, so it’s under increasing political pressure to change. the game here is to shut down or pollute the original data source by any means necessary, so that the only way to access that data becomes an LLM. the people behind the AI startups are experts at creating monopolies, and shutting down a world-class data source like Wikipedia or making it otherwise unusable would guarantee a monopoly position for them.
  • I keep stopping myself from doing this exact project, with the fediverse as the curation source, several times. I’ve talked about this before, but interestingly Postgres’ full-text search is effectively the complete core of a search engine, minus what you’d need for crawling and ranking (which is where curation and a bit of scripting would come in)

    other than resources and time, one big open question is how to do this kind of thing as a positive part of the fediverse — to not make the same mistake that a bunch of techbros already have and index the fediverse without consent. how does one make the curation process simultaneously consensual and also automated enough that it can be reasonably ruggedized against abuse?

  • and of course, not a single citation for the intro paragraph, which has some real bangers like:

    This process involves self-assessment and internal deliberation, aiming to enhance reasoning accuracy, minimize errors (like hallucinations), and increase interpretability. Reflection is a form of "test-time compute," where additional computational resources are used during inference.

    because LLMs don’t do self-assessment or internal deliberation, nothing can stop these fucking things from hallucinating, and the only articles I can find for “test-time compute” are blog posts from all the usual suspects that read like ads and some arXiv post apparently too shitty to use as a citation

  • oh yeah, I’m waiting for David to wake up so he can read the words

    the trivial ‘homework’ of starting the rule violation procedure

    and promptly explode, cause fielding deletion requests from people like our guests who don’t understand wikipedia’s rules but assume they’re, ah, trivial, is probably a fair-sized chunk of his workload

  • this would explain so much about the self-declared 10x programmers I’ve met

  • there’s something fucking hilarious about you and your friend coming here to lecture us about how Wikipedia works, but explaining the joke to you is also going to be tedious as shit and I don’t have any vegan nacho fries or junior mints to improve my mood