@matthewmaybe@sigmoid.social avatar

matthewmaybe

@matthewmaybe@sigmoid.social

Developer of the first "AI art" detection model.

I'm interested in content moderation tooling, data dignity, and fair, fully disclosed, non-commercial uses of generative AI.

My profile pic was generated using Mitsua Diffusion, a unique text-to-image model that was trained from scratch on public domain and Creative Commons CC0-licensed data only.

This profile is from a federated server and may be incomplete. Browse more on the original instance.

NanoRaptor, to random
@NanoRaptor@bitbang.social avatar

Reinventing the potato.

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@NanoRaptor you laugh, but my grandfather built a career on exactly this

matthewmaybe, to random
@matthewmaybe@sigmoid.social avatar

Huge Proportion of Internet Is AI-Generated Slime, Researchers Find

https://futurism.com/the-byte/internet-ai-generated-slime

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

From the article: "...the prevalence of AI-spun gibberish might make effectively training AI models in lower-resource languages nearly impossible in the long run. To train an advanced LLM, AI scientists need large amounts of high-quality data, which they generally get by scraping the web. If a given area of the internet is already overrun by nonsensical AI translations, the possibility of training advanced models in rarer languages could be stunted before it even starts."

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

I'd be surprised if many AI practitioners found the preponderance of AI-generated content on the web to be a serious problem for training new models. Synthesizing training data using GPT-4 is an extremely popular approach for fine-tuning base LLMs. If people are intentionally poisoning training datasets with AI output, will they really care about data contamination?

trochee, (edited ) to random
@trochee@dair-community.social avatar

The boom in LLMs is going to hollow out a number of knowledge-worker industries — for example, writing boilerplate code or technical documentation

Not because it does it well but because the flacks can sell upper management on the idea that it can do it at all, as @pluralistic recently pointed out

This sale is a pig-in-a-poke, and the winning move is to not be holding the bag when the actual code or documentation is found to be terrible

1/

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@luke @trochee another one is the seemingly underrated A24 movie "After Yang", which I just watched last night. Not perfect, but worth checking out.

carnage4life, to random
@carnage4life@mas.to avatar

OpenAI has updated its policy and removed the ban on using its technology for warfare or military uses.

OpenAI is just an AI division of Microsoft at this point masquerading as a non-profit trying to develop AGI “for humanity”. It’s for Microsoft shareholders

https://theintercept.com/2024/01/12/open-ai-military-ban-chatgpt/

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@carnage4life I watched Collossus last night, an old sci-fi movie about (basically) an LLM being granted control of all national security during the Cold War. There are a lot of pretty well-trodden tropes but several aspects are on point, such as the hubris of the system's designer who believes he can tame it with prompt engineering. Oddly, it was recently taken down from all streaming services, but you can still watch it (for now) at the Internet Archive: https://archive.org/details/colossus-the-forbin-project-1970

JohannesKleiner, to random German

Are there feeds yet for Mastodon?

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@roaldarboel @JohannesKleiner @jonny I've been working on this as well; happy to collaborate or share ideas

404mediaco, to random
@404mediaco@mastodon.social avatar

NEW: LAION, the gigantic dataset powering Stable Diffusion, several Google products, and other prominent AI tools has been taken down from the internet because it contains thousands of images of child sexual abuse material

https://www.404media.co/laion-datasets-removed-stanford-csam-child-abuse/

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@404mediaco “It's not that the technology is necessarily bad... it's not that AI is bad. It's the fact that a bunch of things were blindly stolen, and now we're trying to put all these Band-aids to fix something that really never should have happened in the first place.”

Amen. It's increasingly clear that researchers, companies and others operating in this space need to kick their addiction to Common Crawl and all the models and data based upon it (or similar to it).

timnitGebru, to random
@timnitGebru@dair-community.social avatar

"What If Sam Altman Were A Black Woman? Tech Twitter Weighs In On The OpenAI Debacle"

I mean the dude who fired me was waxing poetic about the "small research community" and empathy for his colleagues and stuff so we know the answer.

https://peopleofcolorintech.com/articles/what-if-sam-altman-were-a-black-woman-tech-twitter-weighs-in-on-the-openai-debacle/

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@timnitGebru yes, and also I think it's significant that the OpenAI saga is another example where the white, male management of a tech company sought to stifle AI safety research published by a woman. It says a lot about where we're at with women in tech that he thought it was OK to try to force Helen Toner off the board for doing her job when technically as CEO he reported to her! And now he appears to have won that battle--even if there will be an "investigation..."

timnitGebru, to random
@timnitGebru@dair-community.social avatar

Sigh. Good thing US legislators are meeting with all these leaders trying to hear from them on how they should be regulated 🙄
https://www.404media.co/ai-generated-mushroom-foraging-books-amazon/

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@timnitGebru a very good example of why AI detection tools are needed. yes they could be better but the author of this article has demonstrated an excellent use case for existing ones

timnitGebru, to random
@timnitGebru@dair-community.social avatar

And another message. There was absolutely no need for this and we got the companies wanting to create an AGI god, putting these students in this awful situation. These "AI detection" tools need to be banned.

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@timnitGebru what should be done instead? and why ban the detectors but not the generators?

themarkup, to random
@themarkup@mastodon.themarkup.org avatar

The consequences of being wrongfully accused by an AI detector don't fall on groups evenly. We explain: https://themarkup.org/hello-world/2023/08/19/ai-detector-bias-and-higher-ed-coverage-at-the-markup

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@themarkup "AI detectors work by searching for predictable phrasing, simple vocabulary, and less complex grammar" Where is the source for this? The AI detectors I am aware of use BERT and similar language models as classifier neural networks; we have no more of a clear idea what they are "searching for" than what GPT generators are "trying to say". Bias in AI detection is absolutely an issue, but not one researchers (such as Unitary) haven't been aware of for a long time.

trochee, to random
@trochee@dair-community.social avatar

At the library in the SF section and I found a book that looked promising and checks boxes for me (appears it isn't Heinlein wannabe pastiche; covers questions of autonomy, sentience and who counts as "people"; author is not a white straight cis anglophone dude*)

…but then it has a glowing blurb from Ray Kurzweil, and I just… put it back on the shelf.

No thank you.

I do read a few things by authors in this category, but I'm much less willing to give them* a try without recommendation

**us

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@timnitGebru @trochee I feel like I hear relatively little said about Ray Kurzweil despite his apparent role as the godfather of #TESCREAL. Am I getting that wrong? Didn't he basically invent this new religion, and the rest are just preaching different versions of his gospel? Or did it exist before him in some form?

mekkaokereke, to random
@mekkaokereke@hachyderm.io avatar

Fantastic accomplishment!♥️👍🏿

https://19thnews.org/2023/07/chanda-prescod-weinstein-physicist-tenure-rare-feat/

And when she tried to join the Fediverse, she was greeted with a barrage of hate, sexism, racism, and anti-semitism that should have never been allowed to happen.

So now no one on Fediverse gets to interact with her directly about her work on here. Our loss. 😢

Which is why we'll make it so that this type of terrible welcome is unlikely to happen again. Allowing it to happen to her was a choice. We will make better ones.

#BlackMastodon

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@b3n @techghoul @mekkaokereke @marjolica agreed that no human can tackle this alone, but in addition to shared action I think this has to involve bots of some sort. I've been working on one that scans a feed for toxic status updates (detected using a machine learning model) and reports them to admins. had gotten as far as thinking that after X toxic toots it might recommend banning a user but this could extend to instances.

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@techghoul @mekkaokereke @b3n @marjolica I'm definitely aware of this problem. That's part of why I chose the Detoxify library for my proof-of-concept, as the authors put some thought and effort into countering such false positives with "unbiased" models. But I'm still seeing similar issues, and frankly, looking for a different approach. One thing that would really help is a dataset of reports and admin actions with context.

botwiki, to random
@botwiki@mastodon.social avatar

Do you keep a list of all your bot projects somewhere online?

#bots #CreativeBots

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@botwiki sort of. i started one at https://umm-maybe.com/Bots.html but haven't kept it updated...

chrismessina, to random
@chrismessina@mastodon.xyz avatar

The AI arms race is playing out in real time on the Product Hunt leaderboard.

Undetectio thwarts AI content detection:
https://www.producthunt.com/posts/undetectio

GptSafe detects AI content:
https://www.producthunt.com/posts/gptsafe

Judging by the upvotes, you can tell which side is winning. 😈

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@chrismessina Time to make an Undetectio detector I guess

mmitchell_ai, to random
@mmitchell_ai@mastodon.social avatar

CHECK IT OUT! You can now exclude unconsented images from your models. Next step: Norms+standards for opt-in (convos @huggingface is having with multiple orgs to create!) Data on right column of the Dataset Card!
https://huggingface.co/datasets/conceptual_captions

matthewmaybe,
@matthewmaybe@sigmoid.social avatar

@huggingface @mmitchell_ai great idea. 👏 if the model card templates could encourage disclosure of the licensing status of training material that would also be helpful. I see many new models touted as "open" without discussing whether any steps were taken to filter out copyrighted works.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • JUstTest
  • InstantRegret
  • mdbf
  • ethstaker
  • magazineikmin
  • cubers
  • rosin
  • thenastyranch
  • Youngstown
  • osvaldo12
  • slotface
  • khanakhh
  • kavyap
  • DreamBathrooms
  • provamag3
  • Durango
  • everett
  • tacticalgear
  • modclub
  • anitta
  • cisconetworking
  • tester
  • ngwrru68w68
  • GTA5RPClips
  • normalnudes
  • megavids
  • Leos
  • lostlight
  • All magazines