buc.ci is a Fediverse instance that uses the ActivityPub protocol. In other words, users at this host can communicate with people that use software like Mastodon, Pleroma, Friendica, etc. all around the world.
This server runs the snac software and there is no automatic sign-up process.
I've recently summed up my thoughts on generative "AI" on my homepage. Here's a screenshot of that section.
#tech #technology #BigTech #IT #AI #ArtificialIntelligence #LLM #LLMs #ML #MachineLearning #GenAI #generativeAI #AIAgent #AISlop #FuckAI #Fuck_AI #enshittification #microslop #microsoft #copilot #meta #google #NVIDIA #gemini #OpenAI #ChatGPT #anthropic #claude
It seems hard to escape the AI virus. It's also infecting the open source world…
https://codeberg.org/small-hack/open-slopware
#FOSS #OpenSource #tech #technology #IT #AI #ArtificialIntelligence #LLM #LLMs #ML #MachineLearning #GenAI #generativeAI #AIAgent #AISlop #Fuck_AI #microslop #microsoft #copilot #meta #google #NVIDIA #gemini #OpenAI #ChatGPT #anthropic #claude #editor #app #apps #tools #software #linux #FreeSoftware #free #BigTech
Loved reading this…
Microslop
https://www.s-config.com/microslop
#tech #technology #BigTech #IT #AI #ArtificialIntelligence #LLM #LLMs #ML #MachineLearning #GenAI #generativeAI #AIAgent #AISlop #Fuck_AI #microslop #Microsoft #copilot #Meta #Google #NVIDIA #gemini #OpenAI #ChatGPT #anthropic #claude #linux #FOSS #OpenSource
PyTorch has such frustrating documentation.
It's very thorough, but the API documents always fall into the trap of merely describing the code, without giving any sort of explanation for how you ought to use it. The tutorials similarly show you how to build a thing, but don't bother explaining why they wrote the code like they did.
What's the point? How is this better than just reading the source code?
Also, so much of PyTorch is just wrapper objects that let you hook up any sort of data with any sort of model and run it on any sort of hardware. This means knowing ML won't help you much, because PyTorch isn't about ML. It's about this system of abstractions the PyTorch devs came up with to glue together all the ML code in the world, then failed to document. 🙄
Backprop is an algorithm I run to optimize an ANN. It needs a top-down view of the network topology and the weights of all synapses. It solves the credit-assignment problem in a clever way, usually based on the error rate compared to a known target. Then it simultaneously updates all the link weights in the network based on how the ANN responded as a whole. First you train your network, then you can use it, but not both at once.
Rather than being tuned by some external actor, brain cells manage their own relationships with their neighbors. They grow, prune, and modulate their synapses, and they decide when and how to do that based on imperfect feedback, limited information, and evolved heuristics. Brains track and minimize errors, but the targets are internally generated. This is happening continuously, with fluid transitions between acting in the real world, imagining, thinking, and learning.
I'd argue what the brain does is much harder, and much more interesting.
(3/3)
There are two big reasons I dislike saying brains do something "backprop-like." The first issue is how work is split between nodes and links.
In ANNs, the nodes themselves are trivial, and they're completely homogeneous across a full layer of the network, if not all the layers. Any deeper computation is about how the nodes are wired together. That is, the program is in the links (synapse weights), not the nodes.
By comparison, brain cells are both complex and diverse. We don't know how much of the computation happens within cells vs. between them. We're just starting to figure out what all the different kinds of cells are, but have little idea of what they're doing. It's clear that individual neurons do a lot, and that ensembles of cells manage each other in complex ways.
I worry saying the brain "does backprop" implies a network of trivial nodes, where tuning weight vectors is the place where learning happens. That's likely wrong, and it obscures other possibilities.
(2/3)
I've been thinking about this article @UlrikeHahn shared with me recently. Apparently, I have strong opinions about why we shouldn't say that the brain is doing something "backprop-like" when we learn!
Before we start, the key thing to know is that computation in a neural network is distributed across many nodes connected by links. To tune the behavior of the network as a whole, you need to tune each of the nodes and links, but how do you know how any one node or link contributed to the final answer? It's complicated, and each one depends on many others. We call that "credit assignment."
I think both brains and artificial neural networks (ANNs) need to solve the credit assignment problem. For ANNs, an algorithm called "back propagation" or just "backprop" is the industry standard solution, and it works very well. I think what brains do is different.
(1/3)
EDIT: Free PDF
Interesting paper that was recently updated today engaging Heidegger's philosophy with contemporary Machine Learning techniques. I hope to take more time to engage with this paper over the next couple of days. Still, it is encouraging to see such consideration connecting AGI concerns with philosophical principles. Exploring how contemporary design principles lead more to tool use, than AGI.
https://arxiv.org/abs/2602.19028
The AI shit show goes on…
Pinterest Is Drowning in a Sea of AI Slop and Auto-Moderation
https://www.404media.co/pinterest-is-drowning-in-a-sea-of-ai-slop-and-auto-moderation
#pinterest #tech #technology #BigTech #AI #ArtificialIntelligence #LLM #LLMs #ML #MachineLearning #GenAI #generativeAI #AISlop #Fuck_AI #Microsoft #copilot #Meta #Google #NVIDIA #gemini #OpenAI #ChatGPT #anthropic #claude
How AI slop is causing a crisis in computer science…
Preprint repositories and conference organizers are having to counter a tide of ‘AI slop’ submissions.
https://www.nature.com/articles/d41586-025-03967-9
( No paywall: https://archive.is/VEh8d )
#research #science #tech #technology #BigTech #AI #ArtificialIntelligence #LLM #LLMs #ML #MachineLearning #GenAI #generativeAI #AISlop #Fuck_AI #Microsoft #copilot #Meta #Google #NVIDIA #gemini #OpenAI #ChatGPT #anthropic #claude
Can someone clarify, in academia and industry are LLM hallucinations the result of overfitting, or simply a false positive?
I'm beginning to think that hallucinations are evidence of overfitting. It seems surprising that there are few attempts to articulate the underlying cause of hallucinations. Also, if the issue is overfitting, then increasing training time and datasets may not be an appropriate solution to the problem of hallucinations.
Is there a particular article you have in mind?
I ask because the term ambiguity on this topic has really begun to fascinate me. As the topic has gone more and more mainstream, it has been difficult to sit back and listen to people talk about #AI with no idea what the underlying algorithms are doing. There seems to be a gap between how general users understand #AI or #ML and how developers understand them. My fear is that this gap is what is causing so many issues right now.
Another kind of dangerous feedback loop in #ML involves the user. #MLsec
https://www.nytimes.com/2026/01/26/us/chatgpt-delusions-psychosis.html
Good solid coverage of ChatGPT-Health. Many of the risks BIML has been warning about are coming into play fast. #MLsec #ML #AI
https://www.darkreading.com/remote-workforce/chatgpt-health-security-safety-concerns
Recursive Pollution and Model Collapse Are Not the Same
#MLsec #ML #AI #security #recursivepollution
https://berryvilleiml.com/2026/01/10/recursive-pollution-and-model-collapse-are-not-the-same/
New blog post, first in a long time.
Why I chose to go towards Machine Learning research
https://nilio.ca/post?title=Why%20I%20chose%20to%20go%20towards%20Machine%20Learning%20research
"To reiterate first principles, my main problem with Artificial Intelligence, as its currently sold, is that it’s a lie."
Seamas O'Reilly in @Tupp_ed's (Guest) Gist, today:
https://www.thegist.ie/guest-gist-2026-our-already-rotting-future/
#TheGist #MastoDaoine #AI #GenAI #ML #LLM #ArtificialIntelligence
Psyched to serve on the conference committee and review board for [un]prompted, a new AI security practitioner conference, happening March 3/4 in SF's Salesforce Tower.
This is a community-focused event with a bead on what actually works in #ML/#AI security, from simple tools that just work, through strategy, all the way to offense and defense.
Submit a talk. Check the conference out.
Let's see some real #MLsec
Meta released their Segment Anything Model for Audio that lets you separate audio with text prompts. It could be guitar, speech, bird sound, etc. #ML #AI https://ai.meta.com/samaudio/
I see a lot of blank, outright rejection of #AI, LLMs general or coding LLMs like #ClaudeCode in special here on the Fediverse.
Often, the actual impact of the AI / #LLM in use is not even understood by those criticizing it, at times leading to tantrums about AI where there is....no AI involved.
The technology (LLM et al) in itself is not likely to go away for a few more years. The smaller #ML variations that aren't being yapped about as much are going to remain here as they have been for the past decades.
I assume that what will indeed happen is a move from centralized cloud models to on-prem hardware as the hardware becomes more powerful and the models more efficient. Think migration from the large mainframes to the desktop PCs. We're seeing a start of this with devices such as the ASUS Ascent #GX10 / #Nvidia #GB10.
Imagine having the power of #Claude under your desk, powered for free by #solar cells on your roof with some nice solar powered AC to go with it.
Would it not be wise to accept the reality of the existence of this technology and find out how this can be used in a good way that would improve lives? And how smart, small regulation can be built and enforced that balances innovation and risks to get closer to #startrek(tm)?
Low-key reminds me of the Maschinenstürmer of past times...
One Open-source Project Daily ML.NET is an open source and cross-platform machine learning framework for .NET.
The present perspective outlines how epistemically baseless and ethically pernicious paradigms are recycled back into the scientific literature via machine learning (ML) and explores connections between these two dimensions of failure. We hold up the renewed emergence of physiognomic methods, facilitated by ML, as a case study in the harmful repercussions of ML-laundered junk science. A summary and analysis of several such studies is delivered, with attention to the means by which unsound research lends itself to social harms. We explore some of the many factors contributing to poor practice in applied ML. In conclusion, we offer resources for research best practices to developers and practitioners.From The reanimation of pseudoscience in machine learning and its ethical repercussions here: https://www.cell.com/patterns/fulltext/S2666-3899(24)00160-0. It's open access.
In other words ML--which includes generative AI--is smuggling long-disgraced pseudoscientific ideas back into "respectable" science, and rejuvenating the harms such ideas cause.
#AI #GenAI #GenerativeAI #LLMs #MachineLearning #ML #AIEthics #science #pseudoscience #JunkScience #eugenics #physiognomy
Some of the pathological beliefs we attribute to techbros were already present in this view of statistics that started forming over a century ago. Our writing is just data; the real, important object is the “hypothetical infinite population” reflected in a large language model, which at base is a random variable. Stable Diffusion, the image generator, is called that because it is based on latent diffusion models, which are a way of representing complicated distribution functions--the hypothetical infinite populations--of things like digital images. Your art is just data; it’s the latent diffusion model that’s the real deal. The entities that are able to identify the distribution functions (in this case tech companies) are the ones who should be rewarded, not the data generators (you and me).
So much of the dysfunction in today’s machine learning and AI points to how problematic it is to give statistical methods a privileged place that they don’t merit. We really ought to be calling out Fisher for his trickery and seeing it as such.
#AI #GenAI #GenerativeAI #LLM #StableDiffusion #statistics #StatisticalMethods #DiffusionModels #MachineLearning #ML
Something I’ve been thinking about a lot in the current battle over the future of (pseudo) AI is the cotton gin.
I live in a country where industrial progress is always considered a positive. It’s such a fundamental concept to the American exceptionalism claim that we are taught never to question it, let alone realize that it’s propaganda.
One such myth, taught early in grade school, is the story of Eli Whitney and the cotton gin. Here was a classic example of a labor-saving device that made millions of lives better. No more overworked people hand cleaning the cotton (slaves, though that was only mentioned much later, if at all). Better clothes and bedding for the world. Capitalism at its best.
But that’s only half the story of this great industrial time saver. Where did those cotton cleaners go? And what was the impact of speeding up the process?
Now that the cleaning bottleneck was gone, the focus was on picking cotton as fast as possible. Those cotton cleaners likely, and millions of other slaves definitely, were sent to the fields to pick cotton. There was an unprecedented explosion in the slave trade. Industrial time management and optimization methods were applied to human beings using elaborate rule-based systems written up in books. How hard to punish to get optimal productivity. How long their lifespans needed to be to get the lost production per dollar. Those techniques, practiced on the backs and lives of slaves, became the basis of how to run the industrial mills in the North. They are the ancestors of the techniques that your manager uses now to improve productivity.
Millions of people were sold into slavery and worked to death *because* of the cotton gin. The advance it provided did not, in fact save labor overall. Nor did it make life better overall. It made a very small set of people much much richer; especially the investors around the world who funded the banks who funded the slave purchases. It made a larger set of consumers more comfortable at the cost of the lives of those poorer. Over a hundred years later this model is still the basis for our society.
Modern “AI” is a cotton gin. It makes a lot of painstaking things much easier and available to everyone. Writing, reading, drawing, summarizing, reviewing medical cases, hiring, firing, tracking productivity, driving, identifying people in a lineup…they all can now be done automatically. Put aside whether it’s actually capable of doing any of those things *well*; the investors don’t care if their products are good, they only care if they can make more money off of them. So long as they work enough to sell, the errors, and the human cost of those errors, are irrelevant. And like the cotton gin, AI has other side effects. When those jobs are gone, are the new jobs better? Or are we all working that much harder, with even more negative consequences to our life if we fall off the treadmill? One more fear to keep us “productive”.
The Luddites learned this lesson the hard way, and history demonizes them for it; because history isn’t written by the losers.
They’ve wrapped “AI” with a shiny ribbon to make it fun and appealing to the masses. How could something so fun to play with be dangerous? But like the story we are told about the cotton gin, the true costs are hidden.
We later expanded this work and landed it as a chapter in a 2008 book Multiobjective Problem Solving from Nature, which is downloadable from https://link.springer.com/book/10.1007/978-3-540-72964-8 . You'll see the chapter starting on page 357 of that PDF (p 361 in the PDF's pagination). We applied a technique from the theory of coevolutionary algorithms to examine small instances of the game of Nim, and were able to make several interesting statements about that game. Arthur Samuel's original papers on checkers were about learning by self-play, a particularly simple form of coevolutionary algorithm, as I argue in the introductory chapter of my PhD dissertation. Our technique is applicable to Samuel's work and any other work in that class--in other words, it's squarely "machine learning" in the sense Samuel meant the term.
Whatever you may think of this particular work of mine, it's bad news when a field forgets and rejects its own historical origins and throws away the early fruitful lines of work that led to its own birth. #GenerativeAI threatens to have a similar wilting effect on artificial intelligence and possibly on computer science more generally. The marketplace of ideas is monopolizing, the ecosystem of ideas collapsing. Not good.
#MachineLearning #ML #AI #ComputerScience #Coevolution #CoevoutionaryAlgorithm #checkers #Nim #BoardGames
https://mastodon.world/@Mer__edith/113197090927589168
Hype, Sustainability, and the Price of the Bigger-is-Better Paradigm in AI
With the growing attention and investment in recent AI approaches such as large language models, the narrative that the larger the AI system the more valuable, powerful and interesting it is is increasingly seen as common sense. But what is this assumption based on, and how are we measuring value, power, and performance? And what are the collateral consequences of this race to ever-increasing scale? Here, we scrutinize the current scaling trends and trade-offs across multiple axes and refute two common assumptions underlying the 'bigger-is-better' AI paradigm: 1) that improved performance is a product of increased scale, and 2) that all interesting problems addressed by AI require large-scale models. Rather, we argue that this approach is not only fragile scientifically, but comes with undesirable consequences. First, it is not sustainable, as its compute demands increase faster than model performance, leading to unreasonable economic requirements and a disproportionate environmental footprint. Second, it implies focusing on certain problems at the expense of others, leaving aside important applications, e.g. health, education, or the climate. Finally, it exacerbates a concentration of power, which centralizes decision-making in the hands of a few actors while threatening to disempower others in the context of shaping both AI research and its applications throughout society.Currently this is on #arXiv which, if you've read any of my critiques, is a dubious source. I'd love to see this article appear in a peer-reviewed or otherwise vetted venue, given the importance of its subject.
I've heard through the grapevine that US federal grantmaking agencies like the #NSF (National Science Foundation) are also consolidating around generative AI. This trend is evident if you follow directorates like CISE (Computer and Information Science and Engineering). A friend told me there are several NSF programs that tacitly demand LLMs of some form be used in project proposals, even when doing so is not obviously appropriate. A friend of a friend, who is a university professor, has said "if you're not doing LLMs you're not doing machine learning".
This is an absolutely devastating mindset. While it might be true at a certain cynical, pragmatic level, it's clearly indefensible at an intellectual, scholarly, scientific, and research level. Willingly throwing away the diversity of your own discipline is bizarre, foolish, and dangerous.
If you accept the viewpoint of ecological rationality, data (about people) is not nearly as useful for predictive purposes as it's made out to be. There is a "less is more" phenomenon in many applications, especially those that claim to predict behaviors or outcomes of some kind. See also this talk: https://www.cs.princeton.edu/news/how-recognize-ai-snake-oil .
There is a "less is more" effect with food, too. People need a baseline amount of food to maintain health, but having significantly more food than that doesn't confer significantly more health. Also food can spoil if one hoards it.
If data is a liquid, it's more like milk than oil.
What you're seeing here is that for most categories, there is a linear increase in the number of submissions to the category year-over-year up until the end of the data series in 2021. Computer science is dramatically different: its increase looks exponential, and it looks like its rate of increase may have accelerated circa 2017. The chart on the right, which is the same data shown proportional instead of as raw counts, suggests computer science might be "eating" mathematics starting around 2017.
2017 is around when generative AI papers started to appear in large quantities. There was a significant advance in machine learning published around 2018 but known before then that made deep learning significantly more effective. Tech companies were already pushing this technology. #OpenAI (the #GPT / #ChatGPT maker) was founded in 2015; GPT-2 was released in early 2019. arXiv's charts don't show this, but I suspect these factors play a role in the seeming phase shift in their CS submissions in 2017.
We don't know what 2022 and 2023 would look like on a chart like this but I expect the exponential increase will have continued and possibly accelerated.
In any case, this trend is extremely concerning. The exponential increase in number of submissions to what is supposed to be an academic pre-print service is not reasonable. There hasn't been an exponential increase in the number of computer scientists, nor in research funding, nor in research labs, nor in the output-per-person of each scientist. Furthermore, these new submissions threaten to completely swamp all other material: before long computer science submissions will dwarf those of all over fields combined; since this chart stops at 2021 they may have already! arXiv's graphs do not break down the CS submissions by subtopic, but I suspect they are in the machine learning/generative AI/LLM space and that submissions on these topics dwarf the other subdisciplines of computer science. Finally, to the extent that arXiv has quality controls in place for its archive, these can't possibly keep up with an exponentially-increasing rate of submissions. They will eventually fail if they haven't already (as I suggested in a previous post I think there are signs that their standards are slipping; perhaps that started circa 2017 and that's partly why the rate of submissions accelerated then?).