• 0 Posts
  • 56 Comments
Joined 8 months ago
cake
Cake day: March 13th, 2024

help-circle







  • Go to pubmed. Type “social media mental health”. Read the studies, or the reviews if you don’t have the time.

    The average American teenager spends 4.8 hours/day on social media. Increased use of social media is associated with increased rates of depression, eating disorders, body image dissatisfaction, and externalizing problems. These studies don’t show causation, but guess what, we literally cannot show causation in most human studies because of ethics.

    Social media drastically alters peer interactions, with negative interactions (bullying) associated with increased rates of self harm, suicide, internalizing and externalizing problems.

    Mobile phone use alone is associated with sleep disruption and daytime sleepiness.

    Looking forward to your peer-reviewed critiques of these studies claiming they are all “just vibes.”






  • I remember hearing this argument before…about the Internet. Glad that fad went away.

    As it has always been, these technologies are being used to push us forward by teams of underpaid unnamed researchers with no interest in profit. Meanwhile you focus on the scammers and capitalists and unload your wallets to them, all while complaining about the lack of progress as measured by the products you see in advertisements.

    Luckily, when you get that cancer diagnosis or your child is born with some rare disease, that progress will attend to your needs despite your ignorance if it.



  • We are not in a recession. The problems with wage stagnation are not some temporary hiccup in the economy. It is a systemic problem. Stop conflating the two, complaining that a macroeconomic term with a very specific meaning isn’t defined the way you want it to be. Stop expecting the problem to heal itself if the fed lowers rates or taxes get nudged up or down or whatever. We know how to fix wage stagnation because we have done it before. Regulation. Labor protections. Minimum wage increases. Wage stagnation occurs in the absence of these things, and they can only be done by Congress.


  • Even though the law can be circumvented, it nonetheless provides resistance. Traveling to another state, filling out paperwork, paying extra money, etc all provide additional obstacles to overcome. If someone was having an acute mental problem and felt compelled to eat a barrel, a simple few hours delay in acquiring a gun can make all the difference. For someone planning on using a gun for criminal activity, at some point they might just consider employment as an easier alternative if acquiring a gun is too much of a pain.

    We have already seen this effect in reverse with regard to immigration. Legal immigration is such a painful crapshoot that people are willing to surrender their fate to cartels as an alternative.




  • These cases are interesting tests of our first amendment rights. “Real” CP requires abuse of a minor, and I think we can all agree that it should be illegal. But it gets pretty messy when we are talking about depictions of abuse.

    Currently, we do not outlaw written depictions nor drawings of child sexual abuse. In my opinion, we do not ban these things partly because they are obvious fictions. But also I think we recognize that we should not be in the business of criminalizing expression, regardless of how disgusting it is. I can imagine instances where these fictional depictions could be used in a way that is criminal, such as using them to blackmail someone. In the absence of any harm, it is difficult to justify criminalizing fictional depictions of child abuse.

    So how are AI-generated depictions different? First, they are not obvious fictions. Is this enough to cross the line into criminal behavior? I think reasonable minds could disagree. Second, is there harm from these depictions? If the AI models were trained on abusive content, then yes there is harm directly tied to the generation of these images. But what if the training data did not include any abusive content, and these images really are purely depictions of imagination? Then the discussion of harms becomes pretty vague and indirect. Will these images embolden child abusers or increase demand for “real” images of abuse. Is that enough to criminalize them, or should they be treated like other fictional depictions?

    We will have some very interesting case law around AI generated content and the limits of free speech. One could argue that the AI is not a person and has no right of free speech, so any content generated by AI could be regulated in any manner. But this argument fails to acknowledge that AI is a tool for expression, similar to pen and paper.

    A big problem with AI content is that we have become accustomed to viewing photos and videos as trusted forms of truth. As we re-learn what forms of media can be trusted as “real,” we will likely change our opinions about fringe forms of AI-generated content and where it is appropriate to regulate them.