A few absolute shockers in the list of websites the Washington Post has revealed are used to train Google's generative AI tools. Apparently including the likes of 4Chan, Breitbart, and RT.
From WaPo:
"Meanwhile, we found several media outlets that rank low on NewsGuard’s independent scale for trustworthiness: RT.com No. 65, the Russian state-backed propaganda site; breitbart.com No. 159, a well-known source for far-right news and opinion; and vdare.com No. 993, an anti-immigration site that has been associated with white supremacy.
"The top Christian site, Grace to You (gty.org No. 164), belongs to Grace Community Church, an evangelical megachurch in California. Christianity Today recently reported that the church counseled women to 'continue to submit' to abusive fathers and husbands and to avoid reporting them to authorities."
Anyone interested in services like #ChatGPT should check out the new #GenerativeAI service released to the public today. It appears to take a different conversational approach to the same chat concept as we've seen before. It's called PI and you can find it at the address below. It has a voice that responds back if you want to turn it on. Four different voices to choose from. As worried as I am, this stuff fascinates me. This thing is free for now.
This excellent comic on the history of Luddism by Tom Humberstone https://thenib.com/im-a-luddite led me to this site with folks developing 'Glaze' which is a thing artists are using to mess with AI trying to scrape their art. Check it out here:
𝟮𝟬𝟭𝟬𝘀…
I worked hard on some 3D artwork. Now I'm going to publish it on social media and enjoy the likes and responses. 😊
𝟮𝟬𝟮𝟬𝘀…
I worked hard on some 3D artwork. Before it will be ready to publish, I need to make a compilation of work-in-progress screenshots and wireframes, to prove that it's not AI-generated, and to avoid people asking what AI prompt I used to generate the image. 😖😖
Here’s an illustration of a rudimentary mental model that I use when thinking about advances in computing tools. Computers are already (essentially) Turing-complete, so anything that can be done can in theory already be done using any existing technology. But what these leaps do is that they bring down the SKILL and RESOURCES needed to accomplish a certain task. The red arrow in the image is what such a leap does.
#GenerativeAI is such a leap. The red dot represents not only something genuinely useful that people used to have to learn a skill or pay someone to do, but can now do with ease (removing a distracting object from a photograph), but also vectors of abuse that similarly gain the same level of ease (crafting misinformation, generating fake revenge porn). With every leap, thousands of these red dots move toward the origin, some good for humanity, and others bad.
PKD writing presciently about our current information age?
"It is like information theory; it is noise driving out signal. But it is noise posing as signal so you do not even recognize it as noise...If you can float enough disinformation into circulation you will totally abolish everyone’s contact with reality, probably your own included."
Philip K. Dick, The Transmigration of Timothy Archer
Much as I dislike the theft of human labor that feeds many of the #generativeAI products we see today, I have to agree with @pluralistic that #copyright law is the wrong way to address the problem.
To frame the issue concretely: think of whom copyright law has benefited in the past, and then explain how it would benefit the individual creator when it is applied to #AI. (Hint: it won’t.)
Copyright law is already abused and extended to an absurd degree today. It already overreaches. It impoverishes society by putting up barriers to creation and allowing toll-collectors to exist between citizen artists and their audience.
Labor law is likely what we need to lean on. #unions and #guilds protect creators in a way that copyright cannot. Inequality and unequal bargaining power that lead to exploitation of artists and workers is what we need to address head-on.
#Netlify's new #Drop tool allows you to create a new website using #AI and have it automatically deployed to #theCloud.
"Simply describe the website you want to build not only will #chatGPT author the markup scripts and style sheets but it will deploy it to Netlify I for you too!"
Verify then trust. Goes for pretty much everything that Meta releases.
#Meta releases FACET, an #AI benchmark tool to evaluate the "fairness" of AI models that classify and detect things in photos and videos, including people.
"There was a whole chain of separate departments dealing with proletarian literature, music, drama and entertainment generally. Here were produced rubbishy newspapers containing almost nothing except sport, crime and astrology, sensational five-cent novelettes, films oozing with sex, and sentimental songs which were composed entirely by mechanical means on a special kind of kaleidoscope known as a versificator."
So now we can all create endless AI generated pictures with Bing chat, an example made with "Create a picture of a modern house on an exposed cliff" , quite nice. #AI#GenerativeAI#Bing
I downloaded and tried Dolphin 2.7 Mixtral 8x7b today. So far, it's performing well. It's not too slow, averaging around 7.5 tokens per second. Tomorrow, I plan to test its performance in different topics. Perhaps I should also compare it to the other models I've tried. #LLM#LocalLLaMA#AI#GenerativeAI
#Microsoft, #OpenAI, Cohere, and others are testing the use of "synthetic data", as they find generic data from the web is no longer good enough for training #LLMs.
It seems to me that the main problem with #ChatGPT and other #LLMs is context. Each new conversation with them is a clean slate and the longer a conversation goes on the slower and more confused they seem to get. I presume taking the context into account means extra processing time, and storage on their part, but moreover they just don't provide a very good interface for communicating with the #AI about a long-lived project. This is critical for #softwareDevelopment.