Skip Navigation

  • ah yes the only way to make LLMs, a technology built on plagiarism with no known use case, “useful for any actual ai application” is to throw a shitload of money at nvidia. weird how that works!

  • I vaguely remember that one of the articles talking about the physics forum mentioned it happening elsewhere, but I haven’t dug into it myself. it might just be one or two shitty admins doing this, but I suspect (without evidence, I just can’t think of another reason to do it) there’s some party offering a financial incentive for them to go back and fuck up their old forums

  • I think you’re absolutely correct, and this feels to me like the only reason why we’re seeing some of the bizarre shit we’ve been keeping an eye on:

    • several old forums, all of which are unique high-quality data sources, are being polluted by their own admins with backdated LLM-generated answers. this destroys that forum as a trustworthy data source and removes it as competition for the LLM that already scraped the forum — and, as a bonus, it also makes training a future LLM on that data source utterly impractical without risking model collapse.
    • Wikipedia refuses to compromise on quality in general, so it’s under increasing political pressure to change. the game here is to shut down or pollute the original data source by any means necessary, so that the only way to access that data becomes an LLM. the people behind the AI startups are experts at creating monopolies, and shutting down a world-class data source like Wikipedia or making it otherwise unusable would guarantee a monopoly position for them.
  • I keep stopping myself from doing this exact project, with the fediverse as the curation source, several times. I’ve talked about this before, but interestingly Postgres’ full-text search is effectively the complete core of a search engine, minus what you’d need for crawling and ranking (which is where curation and a bit of scripting would come in)

    other than resources and time, one big open question is how to do this kind of thing as a positive part of the fediverse — to not make the same mistake that a bunch of techbros already have and index the fediverse without consent. how does one make the curation process simultaneously consensual and also automated enough that it can be reasonably ruggedized against abuse?

  • and of course, not a single citation for the intro paragraph, which has some real bangers like:

    This process involves self-assessment and internal deliberation, aiming to enhance reasoning accuracy, minimize errors (like hallucinations), and increase interpretability. Reflection is a form of "test-time compute," where additional computational resources are used during inference.

    because LLMs don’t do self-assessment or internal deliberation, nothing can stop these fucking things from hallucinating, and the only articles I can find for “test-time compute” are blog posts from all the usual suspects that read like ads and some arXiv post apparently too shitty to use as a citation

  • oh yeah, I’m waiting for David to wake up so he can read the words

    the trivial ‘homework’ of starting the rule violation procedure

    and promptly explode, cause fielding deletion requests from people like our guests who don’t understand wikipedia’s rules but assume they’re, ah, trivial, is probably a fair-sized chunk of his workload

  • this would explain so much about the self-declared 10x programmers I’ve met

  • there’s something fucking hilarious about you and your friend coming here to lecture us about how Wikipedia works, but explaining the joke to you is also going to be tedious as shit and I don’t have any vegan nacho fries or junior mints to improve my mood

  • also lol @

    Vibe coding, sometimes spelled vibecoding

    cause I love the kayfabe linguistic drift for a term that’s not even a month old that’s probably seen more use in posts making fun of the original tweet than any of the shit the Wikipedia article says

  • did you know: you too can make your dreams come true with Vibe Coding (tm) thanks to this article’s sponsors:

    Replit Agent, Cursor Composer, Pythagora, Bolt, Lovable, and Cline

    and other shameful assholes with cash to burn trying to astroturf a term from a month old Twitter brainfart into relevance

  • no thx, nobody came here for you to assign them tedious homework

  • it’s turning out the most successful thing about deepseek was whatever they did to trick the worst fossbro reply guys you’ve ever met into going to bat for them

  • standard “fuck off programming.dev” ban with a side of who the fuck cares. deepseek isn’t the good guys, you weird fucks don’t have to go to a nitpick war defending them, there’s no good guys in LLMs and generative AI. all these people are grifters, all of them are gaming the benchmarks they designed to be gamed, nobody’s getting good results out of this fucking mediocre technology.

  • this is utterly pointless and you’ve taken up way too much space in the thread already

    It sounds to me like you have a very clear bias, and you don’t care at all about whether or not what they said is actually true or not, as long as the headlines about AI are negative

    oh no, anti-AI bias in TechTakes? unthinkable

  • also:

    So in that thinking, Wikipedia is not open source, if the editor used a proprietary browser?

    fucking no! how in fuck do you manage to misunderstand LLMs so much that you think the weights not being reproducible is at all comparable to… editing Wikipedia from a proprietary browser???? this shit isn’t even remotely exotic from an open source standpoint — it’s a binary blob loaded by an open source framework, like how binary blob modules taint the Linux kernel (you glided right past this reference when our other poster made it, weird that) or how loading a proprietary ROM in an open source emulator doesn’t make the ROM open source. the weights being permissively licensed doesn’t make them open source (or really make any sense at all) if the source literally isn’t available.

  • my fucking god how have you missed the point this hard. fuck off

  • fuck off promptfan

  • off you fuck

  • what if none of it’s good, all of it’s fraud (especially the benchmarks), and having a favorite grifter in this fuckhead industry is just too precious