I think I’m the type of person who gets into things after everyone. To that regard AI is no different, and for a long time I considered LLMs a toy - this was truer of older models, such as the original chatGPT models that came out in 2022-2023.
The discourse has understandably evolved over time and it’s clear that AI is not going anywhere. It’s like quadcopters in warfare, or so many other new techs before. As much as we’d like them not to be used or exist, they will still be. To refuse to adopt new advancements means to be left behind and giving oneself a disadvantage on purpose.
Ultimately the problems around AI stem from capitalism. Yes, there are excesses. But this is true of humans too.
AI - especially LLMs, which I have more experience with - are great at some tasks and absolutely abysmal at others. Just like some people are good at their job and others don’t know the first thing about it. I used to get an ad on Twitter about some guy’s weird messianic book, and in it he showed two pages. It was the most meaningless AI bullshit, just faffing on and on while saying nothing, written in the most eye-rolling way.
That’s because LLMs currently aren’t great at writing prose for you. Maybe if you prompt them just right they might, but that’s also a skill in itself. So we see that there is bottom-of-the-barrel quality, and better quality, and that exists with or without AI. I think the over-reliance on AI to do everything for them regardless of output will eventually be pushed out, and people who do it will stop finding success (if they even found it in the first place, don’t readily believe people when they boast about their own success).
I use AI to code, for example. It’s mostly simpler stuff, but:
1- I would have to learn entire coding languages to do it myself, which takes years. AI can do it in 30 minutes and better than I could in years, because it knows things I don’t. We can talk about security for example, but would a hobbyist programmer know to write secure web code? I don’t think so.
2- You don’t always have a coder friend available. In fact, the reason I started using AI to code my solutions is because try as we might to find coders to help, we just never could. So it was either don’t implement cool features that people will like, or do it with AI.
And it works great! I’m not saying it’s the top-tier quality I mentioned, but it’s a task that AI is very good at. Recently I even gave deepseek all the JS code it previously wrote for me (or even handwritten code) and asked it to refactor the entire file, and it did. We went from a 40kb file to 20 after refactoring, and 10kb after minifying. It’s not a huge file of course, but it’s something AI can do for you.
There is of course the environmental cost. To that I want to say that everything has an environmental cost. I don’t necessarily deny AI is a water-hog, just that the way we go about it in capitalism, everything is contributing to climate change and droughts. Moreover to be honest I’ve never seen actual numbers and studies, everyone just says “generating this image emptied a whole bottle of water”. It’s just things people repeat idly like so many other things; and without facts, we cannot find truth.
Therefore the problem is not so much with AI but with the mode of production, as expected.
Nowadays it’s possible to run models on consumer hardware that doesn’t need to cost 10,000 dollars (though you might have seen that post of the 2000$ rig that can run the full deepseek model). Deepseek itself is very efficient, and there are even more efficient models being made to the point that soon it will be more costly (and resource-intensive) to meter API usage than give it out for free.
I think the place you have as a user is finding where AI can help you individually. People also like to say AI fries your brain, that it incentivizes you to shut your brain off and just accept the output. I think that’s a mistake, and it’s up to you not to do that. I’ve learned a lot about how linux works, how to manage a VPS, and how to work on mediawiki with AI help. Just like you should eat your vegetables and not so many sweets, you should be able to say “this is wrong for me” and stop yourself from doing it.
If you’re a professional coder and work better with handwritten code, then continue with that! When it comes to students relying on AI for everything, then schools need to find other methods. Right now they’re going backwards to doing pen and paper tests. Maybe we should rethink the entire testing method? When I was in school, years before AI, my schoolmates and I already could tell that rote memorization was torture and a 19th century way of teaching. I think AI is just the nail in the coffin for a very, very outdated method of teaching. Why do kids use AI to do their homework for them? That is a much more important question than how are they using AI.
As a designer I’ve used AI to help get me started on some projects, because this is my weakness. Once I get the ball rolling it becomes very easy for me, but getting it moving in the first place is the hard part. If you’re able to prompt it right (which is definitely something I lament, it feels like you have to say the right magic words and they don’t work), it can help with that, and then I can do my thing.
Personally part of my unwillingness to get into AI initially was from the evangelists who like to say literally every new tech thing is the future. Segways were the future, crypto was the future, VR was the future, NFTs were the future, google glasses were the future… They make money on saying these things so of course they have an incentive to say it. It still bothers me that they exist, if you were wondering (if they bother you too lol), but ultimately you have to ignore them and focus on your own thing.
Another part of it I think is how much mysticism there is around it, with companies and let’s say AI power users who are so unwilling to share their methods or how LLMs actually work. They retain information for themselves, or lead people to think this is magic and does everything.
Is AI coming for your job? Yes, probably. But burying our heads in the sand won’t help. I see a lot of translators talking about the soul of their art - everything has a soul and is art now (even saw a programmer call it that to explain why they don’t use AI in their work), we’ve gone full circle back to base idealism to “explain” how human work is different from AI work. AI already handles some translation work very well, and professionals are already losing work to it. Saying “refuse to use AI” is not materially sound, it is not going to save their client base. In socialism getting your job automated is desirable, but not in capitalism of course. But this is not new either, machines have replaced human workers for centuries now, as far back as the printing press to name just one. Yet nobody today is saying “return to scribing monks”.
I think it would be very useful to have an AI guide written for communists by communists. Something that everyone can understand, written from a proletarian perspective - not the philosophy of it but more like how the tech works, how to use it, etc. I can put it up on the ProleWiki essays space if someone wants to write it, we’ve put up guides before, e.g. if you want to see a nutrition and fitness guide written from a communist perspective.
This is not comprehensive, but here is a draft write-up of sorts for the time being. There are probably others who would know better how to write about best use of chat models like Deepseek, I’m not very familiar with them in detail:
What is AI?: AI stands for Artificial Intelligence, which can have broad connotation and could be applied to a number of forms of automation. More recently, it has become synonymous with generative AI, a specific subset of AI in which an AI can be given a prompt (some input from the user) and it will generate something based on the prompt (some output for the user). For the purposes of this article, I will be using AI primarily to refer to generative AI.
Is AI actually “intelligent”?: What is defined as intelligence and whether any AI falls within that may be up for some debate, but instead I want to address the nature of associations with intelligence, autonomy, and behavior. There is no evidence that AI has a “mind of its own” in the way that human beings have a mind of their own and when you consider that it is some math happening on a GPU trained to impersonate human things, not something with a sensate material form interacting with the physical world and then bouncing that back to an inner world, it makes sense that even if it were to have anything resembling intelligence, there is no reason to think it would look similar to the human experience. The point here is not to weigh in on the whole nature of consciousness and intelligence, but that you don’t want to be fooled by an AI acting convincingly human. That’s what it is trained for, but that doesn’t mean it is like a human on a material level. It is still fundamentally a GPU doing some math and shares no actual material traits with you.
AI can be confidently wrong, even when it has high confidence in the right answer: An AI model effectively has a certain amount of probabilistic confidence in what should come next in any given input fed into it. The way it continues is token by token and a token is “however the tokenizer breaks things up”, which varies based on the tokenizer, but most of them do a lot as whole words. There are, however, components of words sometimes too, such as -ly. Tokenizing aims to maximize the semantic literacy of the AI, i.e. it can better understand why some words follow other words if the components give context. A tokenizer that breaks everything up into letters would probably end up more like the word game Hangman than word association. Further, the base probabilities of the model tokens get sampled from through various statistical algorithms, which can dramatically alter the end probability of a given token. In effect, this means that it could have 95% confidence that “purple” is an accurate continuation of "red and blue make " and still choose “yellow.”
AI has no ideology, but it does have biases: Because AI does not have a mind of its own in the way that we do, I think it’s better not to think of it like it has an ideology or belief system. However, this doesn’t prevent it from having biases. The probabilities it learns are based on what kind of material it was given in training data. If everything in its training is saying “Stalin did nothing wrong,” it probably will too. If everything in its training is saying “commie bad”, it probably will too. Further, some companies may consciously try to tune a model’s biases in a particular direction, or block off certain paths of conversation, for this reason. The same goes for image generation. If most hair in its training is blonde, unspecified hair color will likely be blonde. If most skin is white, well… you get the idea.
AI, like anything, is limited by material constraints: Contrary to how it may feel at times, AI is not magic and its capabilities do have limits. If you ask it about theoretical physics, it might have a very good answer depending on how well it has been trained on material written by experts in theoretical physics. If you ask it about a TV show that came out yesterday, chances are it has never seen anything about that show in training and so it will have no idea what you are talking about. Like a human, if it has never encountered something before, it’s not going to magically be able to know what it is. However, unlike a human, you can’t just explain something to it that it doesn’t know and it will now know going forward. More on that below.
An AI model can’t remember what you said to it: This is tricky wording in a way because there are text services that offer some form of automated simulation of memory, where the AI can draw from some kind of record of certain things that you have said and these are supposed to play into how it responds. But the model itself is still not remembering what you said. Training AI models is expensive and training them to be different for a specific user would mean needing to have tons of variations on a model, which would get absurd fast and might result in an overall worse model for any given user (training can be a sensitive process and easily go wrong). Instead, most of what models are working with is a sort of “short-term memory”, sometimes called context; this is still not really any kind of actual working memory in the human sense, it’s just a way of thinking about input. The AI generates from input and so if part of the input is “I already said that, what are you talking about?”, that hints at a kind of conversation where a misunderstanding has occurred. This doesn’t mean the AI will “know” what you already said, it can only “see” what is given as input, plain and simple, and generate based on that. If what you already said is part of the input, it might be able to reference it. Otherwise, it can’t.
AI doesn’t know there is a “you” and an “it”: Some chat-focused AI can be very convincing at acting like a real person, but this doesn’t mean that the AI really “knows” there are separate entities talking to each other. The standard text model is a continuation model, which means they are continuing the given input, i.e. adding onto it. So to get an AI to simulate being a chat partner, special designing behind the scenes forces it to stop writing before it would start writing your side of the conversation. Without these rails, you could watch an AI simulate a conversation between two people, doing both sides of it.
Most AI is not private: This is not to say that people are necessarily interested in reading your specific chats with your username attached, but most AI companies use your chat data in some form, probably anonymized but nevertheless plaintext data they can read, in order to further tune their models. There are rare exceptions, such as NovelAI (which uses a special system of encryption) and ChubAI (possibly, based on wording of their policy), but you need to read carefully through Privacy Policies and how things are worded. A policy that simply says “your stuff is private” doesn’t necessarily mean it can’t be seen, it might just mean that there is no explicit policy in place for members of the company to be looking at what is on your profile.
Steering AI and avoiding unwanted outcomes: Most AI can go in a wide variety of different directions. This opens up a lot of possibility, but it also means a lot of ways in which the AI can not know what you want and go down a path you weren’t expecting and would rather it didn’t go down. Getting it to head toward the realm you want can be subtle or overt, and with the models that are tuned for instruction, overt may be easier, but either way, expect AI will not meet your expectations sometimes. And if it goes down a path you really don’t want, it’s usually better to move on as fast as possible, rather than getting bogged down in telling it how wrong it is. Remember that it’s just continuing based on input, so if it sees “argument”, it may figure “more arguing fits here”. It doesn’t know when to quit, it doesn’t get weary, and the more there is of something in input, the more likely it will be to double down on that.
P.S. These are the points to cover I could think of at the moment based on what I know about generative AI. If there are other points you are thinking of when you talk about writing a guide, I’m curious to know. Can’t guarantee I’d be able to cover it all myself, but yeah. Also am curious if this comes across clear enough for layperson. I’m not an ML researcher or the like, but I’ve picked up a fair bit on it over time.
I think this is a great start, for a guide I would personally see it divided into sections/chapters and subsections, and the concepts broken down even more, with examples and down to their basic elements. Could also imagine including diagrams. Think of the stuff you would have liked to know starting out. I also thought prior to that that it could include example prompts with good and bad practices. Also a part on self-hosting a model.
To give you an idea the nutrition and fitness guide took me an entire weekend to write and then several more days of refining and fine-tuning x)
If you put it up on a google doc (or similar) we can more easily copy and paste to a prolewiki page afterwards!
Thanks for the input. I will see if I can work on it more at some point in a google doc or the like, and maybe see about sharing editing capabilities on such a doc if there’s anybody here who knows more about things like prompting an instruct model? Especially for different specialized purposes like coding. I’m not sure how well I’d do writing that part, since my familiarity with text AI is more so with co-writing fiction. This was mostly an “inspiration struck” thing, so wanted to do some writing while it was there, or else I would have looked more into prolewiki and contributing first lol. But yeah, makes sense to want to section it out more formally and in more detail.
I don’t mean to steal your thunder haha, just brainstorming with half-awake brain (or is it half-asleep?)
Another thing I think of when writing guides is would I read it myself if I didn’t write it? I was thinking of including a glossary but personally I would not just read a glossary, so if I wrote it I wouldn’t include one but include the terms naturally in the flow.
No worries, I appreciate it, really. I have a bit of a push and pull sometimes where I both want to make sure something is actually useful and meets what people expect of it, but also don’t want to overthink it to the point I don’t ever progress on it. Which can be especially tricky with a thing like this, where there’s no one right way to do it. Or rather, there are several right ways, but we want to make sure to do it in the left ways. :P But more seriously, getting feedback, ideas, a feel for what people expect or might want, is an important part of that.