

This person is just posting slop. I’m not sure if trolling/farming, but any interaction will just feed their weird delusions or bait. Just ignore would be my suggestion.


This person is just posting slop. I’m not sure if trolling/farming, but any interaction will just feed their weird delusions or bait. Just ignore would be my suggestion.


This person is using AI and posting slop. I’m not clicking on their stuff but their post history says they made a video about this stupidity previously, so I don’t know that I’d trust them.


I don’t necessarily support this, but I feel like this at least adds a barrier to a single person setting up hundreds of bot accounts easily. Similar to locking your doors, it just adds friction to those that would otherwise not put a lot of effort in. People can still break into your house/run a bot farm, but this makes it just that much more difficult. Again, I don’t really support doing this, and it’s definitely not being done for user’s benefit, but it will likely have some impact on botting.


Guy work in IT and spent 100k to pay devs to make an app so people can talk to his tuned ChatGPT? I hope anyone who has hired him checks his work. That does not bode well for his work product.
Another case from the article:
“I still use AI, but very carefully,” he says. “I’ve written in some core rules that cannot be overwritten. It now monitors drift and pays attention to overexcitement. There are no more philosophical discussions. It’s just: ‘I want to make a lasagne, give me a recipe.’ The AI has actually stopped me several times from spiralling. It will say: ‘This has activated my core rule set and this conversation must stop.’
What’s weird to me is they now recognize AI will lie to you but somehow think they can prompt it not to? Your rules can be “overwritten” because they do not exist to ChatGPT. It does not know what words mean.


This was only done because the editors pushed to minimize AI involvement. There’s a comment here already mentioning that: https://lemmy.world/comment/22826863


The images I am referring to are likely distinct from the ones in the title as they are from his iPhone and Google is who reported him. Regardless in the article it says the detective looked at one of the Google reported images. Whether they just referenced a known hash I don’t know for sure, but I think it’s pretty well known that FAANG scan basically all images for CSAM nowadays.


The detective alleges that that photograph and others she examined appeared to be stored in a folder on the iPhone titled “Girls I Drugged And Raped.”


This link was shared on Lemmy just recently about this book. I watched some of the persons other videos, and though we have different tastes it seems like they can identify “bad” books.


Not sure how I didn’t hear of this already. Apparently it’s not necessarily a breathalyzer, but the proposals include a camera facing the driver to monitor them and passive monitoring of the air in the car.
I don’t drunk drive and barely even drink, but that’s horrifying. I can’t believe this went under the radar for me.
More garbage that is going to break and cost thousands of dollars to fix in addition to all the violations of privacy. Cars are already advertising to people. Can you imagine if they put a camera inside the vehicle? Why not invest in public transit? That’s a great way to decrease impaired drivers of all stripes as well as help people in general. All this does is funnel more money into auto makers. I am so upset that this is the first I’m hearing of it.


Wow. What a terrible idea. There was a woman who was sent to jail in a different state for several months and lost her house, car, and dog because AI misidentified her and cops didn’t give a fuck. Cops should need a warrant for facial recognition at the very least, if it’s allowed at all. Can’t wait for “give me a smile” to be codified into law.
What a completely accurate description. The nuance of the issues being subtle yet catastrophic is always the part that I find the funniest, because how are they so incapable of seeing how that might be a universal issue? Thank you for the chuckle.
Ridiculous that Grammarly even attempted to do this. The article was good, but at the end, though they hedged, they fell into the same trap everyone seems to. AI is not better at coding than it is at writing and their tinkering with this does not suggest that. Grammarly had a bad product, but realistically, there was likely just no effort put into this aspect of the software. Maybe I’m way off base, and I don’t support AI either way, but I just think it was a poor way to end the article. Programmers think it’s good for art, artists think it’s good for programming, it’s almost like it’s easier to see flaws in a field you’re familiar with.


If you sandbox anything it’ll be safer than otherwise. Not really sure what you’re suggesting. I would still want the code reviewed regardless of the safety measures in place.
I wrote a program that basically auto organizes my files for me. Even if an AI was sandboxed and only had access to the relevant files and had no delete privileges, I would still want the code reviewed. Otherwise it could move a file into a nonsensical location and I would have to go through all possible folders to find it. Someone would have to make the interfaces/gateways and also review the code. There’s no way to know how it’s working, so there’s no way to know IF it’s working, until the code is reviewed. Regardless of how detailed you prompt, AI will generate something that possibly (currently very likely) needs to be adjusted. I’m not going to take an AIs raw output and run it assuming the AI did it properly, regardless of the safety measures.


While I personally don’t like AI, I do think it is changing things. I don’t think it’s ever safe to run code without oversight from an actual programmer, but AI will likely affect the number of programmers being hired in a non negligible way.


I keep seeing this, but I think people forget what things were like before we had a formalized education system. They were not good. The modern system can certainly be improved, but it has overall improved upon itself essentially since its inception. I’m not sure if there’s some kind of golden age of US education people are imagining, or they’re just pointing out current flaws, but it really is (unfortunately or not) the best it’s ever been.


That’s really not fair to universities or the inventors. Knowledge is not inherently evil, and things that have far reaching positive impacts can be used for nefarious purposes. Modern society has perverse incentives, but individuals adding to a corpus of humanity’s knowledge are not the ones at fault.


Often times these purchases are not for the product itself, but how it can be incorporated into an existing product. I imagine if Meta makes bot accounts for people to follow/engage with, they can increase user retention and therefore ad revenue.


Yang is a grifter and no one should listen to him. Companies will happily use any excuse to fire employees and create a perception of job scarcity so that they can rehire workers who are scared and desperate and willing to take less compensation for more work.
All of that said, AI is definitely being incorporated quite heavily into a lot of products. It’s already caused issues with services we all rely on, and I hope we are able to hold companies accountable and stop patronizing them wherever possible. AI cannot do a lot of the things they are pretending it can and we are paying the price, not the companies responsible.


That behaviour would probably have the opposite effect that the people who created this rule would want.
Why are you suggesting that? Ignoring capitalistic incentives, the rule is theoretically in place to increase safety. Your decision would have no impact on safety so I’m not sure why you think it would have the opposite effect.
Devastating that Apple did this and still manages to have the worst AI out there. Was really hoping it was so far behind because it only used “ethical” or licensed data. I guess it’s bad just because. That’s extra embarrassing.