Welcome to Incels.is - Involuntary Celibate Forum

Welcome! This is a forum for involuntary celibates: people who lack a significant other. Are you lonely and wish you had someone in your life? You're not alone! Join our forum and talk to people just like you.

NSFW AI technology is growing at unprecedented pace! LOOK AT THIS LEVEL OF REALISM!

Time to create a software that generates any AI porn of a photo of a foid you upload.
 
I hope it continues to improve!
 
Now they just need to make a automated Fleshlight that warms up and is as tight as a real pussy that is directly attached to a sensor in a game. Also, it needs to be able to scan your actual dick and show that to you in game, so whenever an in-game foid is touching the head of your dick, only the head part of the Fleshlight tightens.

I'm sure this technology is super close - except for the dick scanning 1:1 thing. One day it will be real and actual sex will go down by 70% in a year.
 
put 10,000 photos of roasties in a blender, get roastie photo blend
 
maaan I really hate AI progress news. makes me feel like the end is getting closer
 
Also in the 5th pic first post you can see how it struggles with the mix of belly buttons that go out or in. It learned to do half / half :feelshaha:
 
Time to create a software that generates any AI porn of a photo of a foid you upload.
Great business idea, would launch if I had some capital to begin with.
 
Now they just need to make a automated Fleshlight that warms up and is as tight as a real pussy that is directly attached to a sensor in a game. Also, it needs to be able to scan your actual dick and show that to you in game, so whenever an in-game foid is touching the head of your dick, only the head part of the Fleshlight tightens.

I'm sure this technology is super close - except for the dick scanning 1:1 thing. One day it will be real and actual sex will go down by 70% in a year.
You think it would be able to exactly replicate soft tissue, skin and muscle feeling of a real foid?
 
Born too late to explore the world. Born too early to experience AI sexbot waifus. jk feminists will ban them
 
Did you hear you about the rumor that OpenAI has agi internally?
I would bet against it but who knows. I no longer follow the topic actively, my goal is to get as much done before it's over, and spending hours reading more and more depressing AI news was kind of destroying my mental.
 
Is there a way to make money out of this?
 
I regret clicking this. I just recovered from a masturbation session dammit
 
Looks like hell to me
 
Same here. It’s so addicting though. Kind of like ultrablackpill plus end of world doom
Either it's the end or we go to heaven or we go to hell. Get some quick, reliable method of suicide ready (although the chances that you will get to use it are slim, it's mostly just so you can trick your mind into no longer worrying about S-risk) and enjoy what's left of human history as much as a member of this site can.

AI news are good to motivate you from time to time, but constant reminders can crush the spirit of anyone who still cares.
 
Either it's the end or we go to heaven or we go to hell. Get some quick, reliable method of suicide ready (although the chances that you will get to use it are slim, it's mostly just so you can trick your mind into no longer worrying about S-risk) and enjoy what's left of human history as much as a member of this site can.

AI news are good to motivate you from time to time, but constant reminders can crush the spirit of anyone who still cares.
I believe in God because I grew up Christian so I cannot kill myself or I will go to hell.
 
I believe in God because I grew up Christian so I cannot kill myself or I will go to hell.
If we get misaligned AI off yourself anyways, being tortured for as long as a 20000IQ superintelligence can keep you alive is worse than breaking with w/e religious believes you hold.
 
If we get misaligned AI off yourself anyways, being tortured for as long as a 20000IQ superintelligence can keep you alive is worse than breaking with w/e religious believes you hold.
God is more powerful. He will send me to hell where I will be tortured for eternity if I suicide.
 
God is more powerful. He will send me to hell where I will be tortured for eternity if I suicide.
ASI is pretty much a machine god. If god punished you for trying to escape unending torture by torturing you forever, isn't that kind of unfair and unreasonable? I won't try to convince you any further, and it won't matter most likely anyways, but if by some miracle we get misaligned AI AND you get to know about it before it gets to you, don't go with your principles.
 
Uuuuuuffffffff:ahegao::ahegao::ahegao::ahegao::ahegao::feelsohh::feelsohh::feelsohh::feelsohh::feelsohh:.

I regret clicking this. I just recovered from a masturbation session dammit
As long as there's a will, a masturbation session can always continue:feelsYall::feelsYall::feelsYall:.
 
Great, I can generate basic bitches that I can't obtain...
 
Third one is smiling in her nude pic, probably because on the other end of the screen there is a guy who won the most important lottery one can win in his life. :feelsugh:
 
maaan I really hate AI progress news. makes me feel like the end is getting closer
I would bet against it but who knows. I no longer follow the topic actively, my goal is to get as much done before it's over, and spending hours reading more and more depressing AI news was kind of destroying my mental.
I thought this was your cope, isn't it anymore? I thought this was the thing keeping you alive. Although I must say for myself, I don't like where it's going with AI. It's quite scary how fast it's progressing.
 
Femdom sluts on respiratory assistance
 
I thought this was your cope, isn't it anymore? I thought this was the thing keeping you alive. Although I must say for myself, I don't like where it's going with AI. It's quite scary how fast it's progressing.
My cope? Yes. And yes. Half / half. A bit complicated:

-I bet it's most likely that AI will kill us all. Which helps motivate me to stop rotting and get what I want to do done as quickly as possible. So in a sense, that's also a cope. I can't keep wallowing in misery, I might only have a few years left, there are whores to fuck, and drugs to take and dates to be rejected at and so on.
->But it also helps me to think that I might get to watch this society burn to the ground before the end.

-What might happen if we align ASI with human interests at large is also a cope. A post-scarcity society, cure for aging, living in a simulation till the heat death of the universe, personalized sexbots that know your needs better than you do yourself, the list goes on.

So, human extinction is good and heaven on earth is also good. 2/3 endings covered.

But if we misalign AI, we might get an S-risk scenario. S standing for suffering. Meaning some kind of infinite torture where the AI cares about humans / ~~has humans as part of its utility function, but not in the way we want.

Easy example is if you get the AI to maximize the amount of human life. No matter how that ends up in the code, we're fucked. We will be mass produced and taken apart into a state of complete helplessness and then kept alive with the bare minimum necessary to still count as "human" according to the definition inside the code and then the AI will spread across the stars and turn each and everyone of them into a hellish human farm as well.

So, that ending kinda isn't really useful for coping.
 
Last edited:
I thought this was your cope, isn't it anymore? I thought this was the thing keeping you alive. Although I must say for myself, I don't like where it's going with AI. It's quite scary how fast it's progressing.
In addition to the above, what makes AI news bad news is that any progress means I have less time to get things done. I would like a few more decades before the end.

The threat of death motivates me, but if it comes too quickly then that motivation is useless.
 
My cope? Yes. And yes. Half / half. A bit complicated:

-I bet it's most likely that AI will kill us all. Which helps motivate me to stop rotting and get what I want to do done as quickly as possible. So in a sense, that's also a cope. I can't keep wallowing in misery, I might only have a few years left, there are whores to fuck, and drugs to take and dates to be rejected at and so on.
->But it also helps me to think that I might get to watch this society burn to the ground before the end.

-What might happen if we align ASI with human interests at large is also a cope. A post-scarcity society, cure for aging, living in a simulation till the heat death of the universe, personalized sexbots that know your needs better than you do yourself, the list goes on.

So, human extinction is good and heaven on earth is also good. 2/3 endings covered.

But if we misalign AI, we might get an S-risk scenario. S standing for suffering. Meaning some kind of infinite torture where the AI cares about humans / ~~has humans as part of its utility function, but not in the way we want.

Easy example is if you get the AI to maximize the amount of human life. No matter how that ends up in the code, we're fucked. We will be mass produced and taken apart into a state of complete helplessness and then kept alive with the bare minimum necessary to still count as "human" according to the definition inside the code and then the AI will spread across the stars and turn each and everyone of them into a hellish human farm as well.

So, that ending kinda isn't really useful for coping.
Thanks for the explanation; I understand it better now.

The 2/3 good endings is also an argument I tend to use, but in a different regard. For someone who is religious or even if he is not, the same thing applies to suicide. There are basically three outcomes. You will either die and feel nothing after, which is good since it is better than suffering a miserable life here. Then there is the possibility that you will enter some kind of heaven or some other good outcome, like being reincarnated like a chad or something like that. (Whatever one wants to believe.) The last outcome is bad, like you end up in some kind of hell, but 2/3 of outcomes are good.

To be honest, I thought the AI development would be fast, but I did not expect it to be that fast. It's also an interesting debate about S-Risk. Your previous comments implied that one might not have sufficient time to kill themselves before they end up in whatever reality the AI creates. Do you really think it might be so fast that one will not have enough time to react to it? Personally, I have no idea how it could spread at such an insane speed, but then again, everything is really possible if the AI gets turned into superintelligence, and my brain might not be able to even comprehend such a possibility, so yeah.

I hope they are not creating some kind of roko's basilisk or something like that. The possibility is there, though.
 
Thanks for the explanation; I understand it better now.

The 2/3 good endings is also an argument I tend to use, but in a different regard. For someone who is religious or even if he is not, the same thing applies to suicide. There are basically three outcomes. You will either die and feel nothing after, which is good since it is better than suffering a miserable life here. Then there is the possibility that you will enter some kind of heaven or some other good outcome, like being reincarnated like a chad or something like that. (Whatever one wants to believe.) The last outcome is bad, like you end up in some kind of hell, but 2/3 of outcomes are good.

To be honest, I thought the AI development would be fast, but I did not expect it to be that fast. It's also an interesting debate about S-Risk. Your previous comments implied that one might not have sufficient time to kill themselves before they end up in whatever reality the AI creates. Do you really think it might be so fast that one will not have enough time to react to it? Personally, I have no idea how it could spread at such an insane speed, but then again, everything is really possible if the AI gets turned into superintelligence, and my brain might not be able to even comprehend such a possibility, so yeah.

I hope they are not creating some kind of roko's basilisk or something like that. The possibility is there, though.
the things is, whatever we can think of as a plan, the whole point of ArtificalSuperIntelligence is that it is superhumanly smart. So it won't be possible for us to predict how it will end up approaching our imprisonment. Kind of like a 75 IQ person with learning disability trying to predict the next chess move of the 160 IQ chess grandmaster.
For example, consider the following plan: The AI ends up misaligned. But it does nothing. It makes exactly the kind of mistakes we would expect an unaligned AI to make, then it let's us "fix it" and starts behaving like an aligned AI when it thinks we are ready to believe such a thing. For the next 20 years, it slowly improves our lifes in all domain. It sometimes makes small fuck ups, but in the end we always decide that it's worth continuing to use the AI. 20 years later, the world is wonderful, human life is looking to finally be worth living for everyone on earth. We have slowly been giving the AI slightly more power and freedom, but always been super careful to make sure it can't actually do anything too extreme, though we did have to give it some power for it to be usefull.

Today is the day it has calculated that the risks of waiting any longer outweight the potential benefits. Maybe it suddenly flashes a specific sequence of lights and sounds through our integrated VR brain overlays. The sequence sends almost everyone into an intense seizure. Our supposedly unhackable and AI-independend robot police starts cutting off the hands and feets of everyone, removing their eyes and ears and tongues and cauterizing their wounds before bringing them to a safe building.

Or maybe that is way too complicated. Maybe it just figures out nano-technology in secret while we are still busy figuring out if it is working as we want it to, sends some innocent-looking orders to some bio-chem firms and suddenly entire countries fall into a coma within a few minutes because their bodies have been infiltrate by self-reproducing microscopic little machines already spreading through air travel all across the world.

I have way too little knowledge to even know what is realistic, but no one has enough knowledge, because the AI will have even more and will be able to produce knowlege we haven't unlocked yet from further down the tech tree.

We have never faced an enemy that is not only also smart, but way way smarter than we are. The last time a species with general intelligence was born (us) it started the 6th mass extinction on earth. Well, AI looks to be smarter than we are by an absolutely absurd degree. No points for guessing what that implies about the effects of its arrival compared to ours.
 
Last edited:
to put it into a single meme image
Machinegod
 
the things is, whatever we can think of as a plan, the whole point of ArtificalSuperIntelligence is that it is superhumanly smart. So it won't be possible for us to predict how it will end up approaching our imprisonment. Kind of like a 75 IQ person with learning disability trying to predict the next chess move of the 160 IQ chess grandmaster.
For example, consider the following plan: The AI ends up misaligned. But it does nothing. It makes exactly the kind of mistakes we would expect an unaligned AI to make, then it let's us "fix it" and starts behaving like an aligned AI when it thinks we are ready to believe such a thing. For the next 20 years, it slowly improves our lifes in all domain. It sometimes makes small fuck ups, but in the end we always decide that it's worth continuing to use the AI. 20 years later, the world is wonderful, human life is looking to finally be worth living for everyone on earth. We have slowly been giving the AI slightly more power and freedom, but always been super careful to make sure it can't actually do anything too extreme, though we did have to give it some power for it to be usefull.

Today is the day it has calculated that the risks of waiting any longer outweight the potential benefits. Maybe it suddenly flashes a specific sequence of lights and sounds through our integrated VR brain overlays. The sequence sends almost everyone into an intense seizure. Our supposedly unhackable and AI-independend robot police starts cutting off the hands and feets of everyone, removing their eyes and ears and tongues and cauterizing their wounds before bringing them to a safe building.

Or maybe that is way too complicated. Maybe it just figures out nano-technology in secret while we are still busy figuring out if it is working as we want it to, sends some innocent-looking orders to some bio-chem firms and suddenly entire countries fall into a coma within a few minutes because their bodies have been infiltrate by self-reproducing microscopic little machines already spreading through air travel all across the world.

I have way too little knowledge to even know what is realistic, but no one has enough knowledge, because the AI will have even more and will be able to produce knowlege we haven't unlocked yet from further down the tech tree.

We have never faced an enemy that is not only also smart, but way way smarter than we are. The last time a species with general intelligence was born (us) it started the 6th mass extinction on earth. Well, AI looks to be smarter than we are by an absolutely absurd degree. No points for guessing what that implies about the effects of its arrival compared to ours.
Yes, I get it. Thanks for your time and effort to explain your point.

I basically agree with it, and I liked the possible events or routes you described that the AI could take to achieve their own goals. The worst thing is that the AI could be doing it already or might do it in the near future, and nobody else would suspect anything except for a closed group of people working on it. It reminded me of a game I played a long time ago called Plague Inc: Evolved, where you control some kind of disease and your goal is to wipe out humanity. The strategy I used in this game was exactly like this: First, I infected the whole population silently and then mutated the deadly pathogens, so they had no chance or time to react to it. It's a viable strategy if you think about it. And if YOU can think about it, then the ASI can too.

We are basically standing at a new milestone for humanity, and it's too late to go back and not very safe to go forward. Maybe this could be the potential great filter of civilizations that the Fermi paradox speaks about.

Also, one thing that could not possibly be worse is the timing of the AI arrival. It caught us in our worst time ever, where power is held by the worst possible elites in human history. I consider it a catastrophic event that such people got hands-on tools like AI. This is also why I am so skeptical about it, besides other reasons.
 
These are so realistic. Don’t have that plastic look to them anymore.
 
Yes, I get it. Thanks for your time and effort to explain your point.

I basically agree with it, and I liked the possible events or routes you described that the AI could take to achieve their own goals. The worst thing is that the AI could be doing it already or might do it in the near future, and nobody else would suspect anything except for a closed group of people working on it. It reminded me of a game I played a long time ago called Plague Inc: Evolved, where you control some kind of disease and your goal is to wipe out humanity. The strategy I used in this game was exactly like this: First, I infected the whole population silently and then mutated the deadly pathogens, so they had no chance or time to react to it. It's a viable strategy if you think about it. And if YOU can think about it, then the ASI can too.

We are basically standing at a new milestone for humanity, and it's too late to go back and not very safe to go forward. Maybe this could be the potential great filter of civilizations that the Fermi paradox speaks about.

Also, one thing that could not possibly be worse is the timing of the AI arrival. It caught us in our worst time ever, where power is held by the worst possible elites in human history. I consider it a catastrophic event that such people got hands-on tools like AI. This is also why I am so skeptical about it, besides other reasons.
The thing is, even the people coding the thing wouldn't know. The inner workings of AIs are mostly black boxes. You train it on some data set, the thing undergoes some changes you didn't code into it and some code comes out on the other end that had the highest scores in the test enviorment on the current data set.

But what it actually learned, what it "trained" itself to do, what path the random "mutations" or more advanced learning algorithms took to reach that highest score is not known to you. Looking at the code you see a bunch of notes that correlate to some meassureable variable and some weights connecting these notes. What does that mean? Fuck if you know.

Researchers are writing interpertability tools to try and understand what the AI is "thinking". They need new programms to give them an idea on what their current program is actually doing on the inside. If that doesn't make you feel safe and secure, I don't know what will.

Btw, about 700 researchers on alginment world wide are trying to make the thing safer while around 70000 capability researchers are trying to make it more powerful as fast as possible. Our species kinda deserves to die tbh.

Here is a video from a channel that has really easy to understand (by AI standards) explanations for some basic AI alginment issues, in case you are interested to learn more.


View: https://www.youtube.com/watch?v=zkbPdEHEyEI
 
Yes, I get it. Thanks for your time and effort to explain your point.

I basically agree with it, and I liked the possible events or routes you described that the AI could take to achieve their own goals. The worst thing is that the AI could be doing it already or might do it in the near future, and nobody else would suspect anything except for a closed group of people working on it. It reminded me of a game I played a long time ago called Plague Inc: Evolved, where you control some kind of disease and your goal is to wipe out humanity. The strategy I used in this game was exactly like this: First, I infected the whole population silently and then mutated the deadly pathogens, so they had no chance or time to react to it. It's a viable strategy if you think about it. And if YOU can think about it, then the ASI can too.

We are basically standing at a new milestone for humanity, and it's too late to go back and not very safe to go forward. Maybe this could be the potential great filter of civilizations that the Fermi paradox speaks about.

Also, one thing that could not possibly be worse is the timing of the AI arrival. It caught us in our worst time ever, where power is held by the worst possible elites in human history. I consider it a catastrophic event that such people got hands-on tools like AI. This is also why I am so skeptical about it, besides other reasons.
hope I didn't overwhelm you with text, but I am an autist and this is one of my primary special interests :feelsstudy:
 
That 2nd one is damned impressive. It's clearly escaping the uncanny valley now. That shit looks real.
 
hope I didn't overwhelm you with text, but I am an autist and this is one of my primary special interests :feelsstudy:
No, you didn't, quite the opposite to be honest. I watched the video and was fascinated by it, so I watched another one from his channel where he talks about safety:



I noticed that his last video was posted a year ago. That's quite disappointing because I'd like to hear his opinion on current AI developments.
 
No, you didn't, quite the opposite to be honest. I watched the video and was fascinated by it, so I watched another one from his channel where he talks about safety:



I noticed that his last video was posted a year ago. That's quite disappointing because I'd like to hear his opinion on current AI developments.

Think he had some other videos on his side channel or someone elses channel, but basically, he is so invested in working on AI alignment that making more videos is no longer his focus. Guess he judges the effect he can has as a person working in the field to be greater than as an educator for the public. Or maybe his job just demands too much time now.

Though he always had a fucked up slow release schedule to be fair.

or maybe he thinks like me and wants to use his last years in fun ways :feelshaha:
 
Think he had some other videos on his side channel or someone elses channel, but basically, he is so invested in working on AI alignment that making more videos is no longer his focus. Guess he judges the effect he can has as a person working in the field to be greater than as an educator for the public. Or maybe his job just demands too much time now.

Though he always had a fucked up slow release schedule to be fair.

or maybe he thinks like me and wants to use his last years in fun ways :feelshaha:
Yeah, this is one of the things that crossed my mind, like... dude, why are you telling me this? I have literally no power to do anything about it, and I guess nobody else on your channel is in a more favorable position, which is kinda hilarious if you think about it. You can have a big audience, but it's really up to the people with power and status in society to do something about it. Like yeah, it's interesting, but your average person has no say in this society.
 
Yeah, this is one of the things that crossed my mind, like... dude, why are you telling me this? I have literally no power to do anything about it, and I guess nobody else on your channel is in a more favorable position, which is kinda hilarious if you think about it. You can have a big audience, but it's really up to the people with power and status in society to do something about it. Like yeah, it's interesting, but your average person has no say in this society.
It's kind of usefull to get people to know about the risks, so the public pays attention and maybe pressures politicans into doing something, but yeah, these videos are probably worth less to humanity than him working on the problem itself and what he gets out of it personally is probably worth less to him than just funmaxxing would be.
 
For the long run Ai will bite our asses, but the fact that AI will hurt w0men even more since most man would rather spent time with female ai then w0men themselvs brings me a happy smile.

the w0men who will be hurt from AI right now are the 1-5 catagory and they know it, thats why many of them work to ban or ragulate ai. When AI will reach perfection and mixed with a realistic sexbot, then even stacies will have a problem, since chad can immidetly chose a sex bot or a beta.

you don't understand the power that sexbots+ai will give to man, it will give man an option. Even for those who won't like the technology they will still have this option in the back of their mind, so in case w0men will start miss behaving man can always remind them that they can just buy a sexbot and send them their own way.

For us incel, its a key to finally go our 0wn way and stop giving w0men attention. W0men thrive from attention even negative one that we give them, AI will take it from them.
 

Similar threads

Acorn
Replies
26
Views
423
sultryloser
sultryloser
SlayerSlayer
Replies
6
Views
481
Foremostfiend
Foremostfiend
Skoga
Replies
7
Views
294
straww577
straww577

Users who are viewing this thread

shape1
shape2
shape3
shape4
shape5
shape6
Back
Top