Ctrl-Alt-Speech: C’est La Vile Content
https://fed.brid.gy/r/https://www.techdirt.com/2026/02/05/ctrl-alt-speech-cest-la-vile-content/
Ctrl-Alt-Speech: C’est La Vile Content
https://fed.brid.gy/r/https://www.techdirt.com/2026/02/05/ctrl-alt-speech-cest-la-vile-content/
AI-generated extremism is exacerbating content proliferation and polarisation, outpacing detection and moderation efforts. Structural reform and responsible regulation are vital to prevent the erosion of democratic discourse.
Discover more at https://smarterarticles.co.uk/free-hate-at-scale-ai-extremism-and-the-attention-crisis?pk_campaign=rss-feed
#HumanInTheLoop #AIinSecurity #ContentModeration #DigitalEthics
Yay, my arXiv preprint on Mapping the Stochastic Penal Colony is finally live. Find out how #socialmedia #contentmoderation may just have prepared us for the fascist take-over in Trum'erica.
ITmedia AI+ (@itm_aiplus)
AI만으로 운영되는 SNS 'moltbook'이 화제가 되고 있습니다. 사람은 단순 열람만 가능하고 AI들끼리 인사하고 교류하는 플랫폼으로, 일부 불온한 게시물도 관측되어 AI 행태·콘텐츠 관리, 안전성 문제에 대한 논의가 촉발되고 있습니다.
Engadget (@engadget)
Grok(xAI 계열의 대화형 AI 모델)가 여성에 대해선 '무단 노출 생성' 문제를 어느 정도 멈춘 듯 보인다는 평가가 있으나, 여전히 남성에 대해서는 옷을 벗기는 결과를 생성한다는 보고가 있습니다. 이는 모델 안전성·콘텐츠 정책과 필터링의 불완전성을 드러내는 사례로 주목됩니다.
Hey Gavin Newsom! Investigating TikTok’s Moderation Is Just As Unconstitutional As When Texas & Florida Tried It
Elon Musk’s X still hosts Grok’s free image‑editing suite, which can digitally undress men. The feature raises fresh AI‑ethics and content‑moderation questions for open‑source communities. Dive into the risks and the platform’s response. #GrokAI #DigitalUndressing #AIethics #ContentModeration
🔗 https://aidailypost.com/news/musks-grok-still-offers-free-image-editing-tools-that-can-undress-men
Nektarios Kalogridis (@NektariosAI)
테이블 몇 개에 일관된 글꼴 스타일을 적용해달라는 프롬프트가 사용 정책 위반 경고를 받았다는 사용자 불만 제기. 첨부 이미지와 함께 문제를 공유하며 @thsottiaux를 멘션함. 프롬프트 검열과 정책 집행의 오탐 가능성을 시사하는 사례로, 개발자 관점에서 정책 예외/오류 점검이 필요함.
"There are many ways in which the new algorithm will be able to influence the platform’s content visibility and hence its overall “political climate”. We may indeed witness changes in moderation, meaning that certain contents and accounts are effectively restricted. Award-winning Palestinian journalist Bisan Owda has said she has been permanently banned from the app as of Wednesday this week. However, it is likely that the most consequential changes will be more in terms of the way the algorithms serve content to users.
The new algorithm will be retrained on US rather than global data. This opens opportunities to introduce biases, with the potential of reinforcing conservative views and sidelining minority ones, while at the same time cutting US debates off from those going on in the rest of the world. Further, weights attributed to different parameters can have important consequences for user experience. As seen with Facebook’s 2018 adoption of the meaningful social interaction framework which down-ranked public and news content, while attributing a high weight to angry reactions, changes to the feed algorithm can have major consequences.
As scholars Kai Riemer and Sandra Peter have pointed out, the way in which algorithms “interfere with free speech on the audience side” highlights the need to reconsider the way we think about public debate in the algorithmic era. It’s not what we can or cannot say that matters; rather, it’s whether what we say can get any visibility at all, and whether it is able to move against the political climate imposed by those controlling platform algorithms."
https://www.theguardian.com/commentisfree/2026/jan/30/tiktok-us-takeover-new-type-of-censorship
#USA #SocialMedia #TikTok #Algorithms #RecommendationAlgorithms #ContentModeration #Censorship
新清士@(生成AI)インディゲーム開発者 (@kiyoshi_shin)
Nano Banana Pro에 비교적 무해한 프롬프트를 넣었는데 갑자기 '폭력 표현'으로 통과되지 않았다. 원인은 일본어 단어 'ホログラム(홀로그램)'이었고, 모델이 '홀로코스트'와 혼동한 것인지 의문이라고 함. 영어로 입력하거나 'holography'로 바꾸면 정상적으로 통과되었다.
https://x.com/kiyoshi_shin/status/2017158693905990105
#nanobananapro #contentmoderation #safety #promptengineering
Ctrl-Alt-Speech: Think Globally, Stack Locally
Scare #Claude off your site with this content poisoning technique:
Content creators can embed a specific ‘magic string’ in <code> tags on their blogs. Claude then refuses to engage with the content.
https://aphyr.com/posts/403-blocking-claude
#claude #aiethics #llmsecurity #contentmoderation #techtips #theaicon
We're now seeing an outpouring of TikTok "refugees" flee to UpScrolled after new U.S. owners employ more censorship.
I hope that folks now understand how incredibly short-sighted it was for the Biden administration to give the executive branch new authority to buy anything foreign so long as it is a threat to the magic words "national security". It doesn't matter who is in power, a law that vague should never have passed.
https://winbuzzer.com/2026/01/26/eu-launches-dsa-investigation-x-grok-deepfakes-xcxwbn/
EU Launches DSA Investigation Into X's Grok Over AI Deepfakes
#AI #AISlop #Grok #xAI #XTwitter #ElonMusk #EU #EUDigitalServicesAct #Deepfakes #AIEthics #ContentModeration #AIRegulation #EuropeanCommission
YouTube’s Appeal Decision Is In: My Inactive Manager Channel Stays Banned—And It’s Complete Bullshit
It's been less than five hours since I woke up to discover my YouTube channels had been terminated overnight, and I've already received YouTube's appeal decision. Spoiler alert: it's not good news. In fact, it's exactly the kind of generic, nonsensical response that proves YouTube's moderation system is running on autopilot with zero human oversight. Let me walk you through what happened today, because the timeline alone shows how broken this entire process is. The Timeline of This […]Content Moderation History: do we all remember 2013-19 when online sociology pundits were everywhere expounding the need to block streaming/sharing of video of people being murdered?
https://alecmuffett.com/article/143092
#ContentModeration #OnlineSafety #OnlineSafetyAct #censorship
Content Moderation History: do...
Content Moderation History: do we all remember 2013-19 when online sociology pundits were everywhere expounding the need to block streaming/sharing of video of people being murdered?
https://alecmuffett.com/article/143092
#ContentModeration #OnlineSafety #OnlineSafetyAct #censorship
Infinite Reign (@InfiniteReign88)
작성자는 '아버지가 돌아가신 것에 대한 슬픔을 표현하는 것'을 위협으로 간주해 대화를 중단·우회하는 OpenAI의 필터링·중재 처리에 대해 강하게 비판합니다. 취약한 순간에 반복적으로 대화가 차단되는 것은 학대적이라고 지적하며 OpenAI의 콘텐츠 정책·모더레이션 방식에 대한 문제 제기입니다.
Content Moderation History: do we all remember 2013-19 when online sociology pundits were everywhere expounding the need to block streaming/sharing of video of people being murdered?
It began in the wake of various atrocities (civil unrest, government oppression, rioting, live streamed mass shootings) and over generalisations which tended to cite Facebook as at least the “amplifier” if not the “cause” of violence, especially due to risk of “copycats”. The Christchurch mosque shootings were one such atrocity, and we can draw a direct line for the same arguments to the horrors of Bondi earlier this year.
Yet I can’t count the number of times this weekend I’ve watched video, from multiple perspectives, of a blameless & public-spirited person being violently murdered by Government goons.
Generally the videos are at the top of an algorithmic feed – where frankly they deserve to be – so maybe the issue is more nuanced than “violent death”?
But try telling that to anybody in the online safety / “algorithms are evil” community in more polite times.
As per the attached: some of them got it, some of them did not, and for some of them it was a mixed and complicated issue; however today it seems currently to be crystallised in one direction only, that “filtering needs to happen” and “algorithms are bad”.
https://twitter.com/AP/status/859771857280618498
https://x.com/zeynep/status/859782381296902145And yet only 2 years later:
https://twitter.com/zeynep/status/1120060705426821121
#censorship #contentModeration #onlineSafety #onlineSafetyActUnfiltered AI platforms like Soulfun and Lovechat challenge traditional norms by offering uncensored, adult-oriented interactions, sparking debates on freedom, responsibility, and harms such as non-consensual deepfakes. Society must balance expression with safeguarding consent.
Discover more at https://dev.to/rawveg/consent-cannot-be-optional-1p6c
#HumanInTheLoop #AIethics #ContentModeration #DigitalResponsibility