kellogh, to LLMs
@kellogh@hachyderm.io avatar

if i had more time, i'd love to investigate PII coming from #LLMs. i've seen it generate phone numbers and secrets, but i wonder if these are real or not. i imagine you could look at the logits to figure out if phone number digits were randomly chosen or if the sequence is meaningful to the LLM. anyone aware of researchers who have already done this?

remixtures, to ai Portuguese
@remixtures@tldr.nettime.org avatar

#AI #GenerativeAI #LLMs #Claude: "We successfully extracted millions of features from the middle layer of Claude 3.0 Sonnet, (a member of our current, state-of-the-art model family, currently available on claude.ai), providing a rough conceptual map of its internal states halfway through its computation. This is the first ever detailed look inside a modern, production-grade large language model.
Whereas the features we found in the toy language model were rather superficial, the features we found in Sonnet have a depth, breadth, and abstraction reflecting Sonnet's advanced capabilities.
We see features corresponding to a vast range of entities like cities (San Francisco), people (Rosalind Franklin), atomic elements (Lithium), scientific fields (immunology), and programming syntax (function calls). These features are multimodal and multilingual, responding to images of a given entity as well as its name or description in many languages."
https://www.anthropic.com/news/mapping-mind-language-model

inthehands, to random
@inthehands@hachyderm.io avatar

Less about tools that boost productivity, more about tools that reduce total workload.

janriemer,

@inthehands ...and now have a look at the summary below by GPT4o and Gemini 1.5.

While it perfectly got it right (this time!), the most crucial bit on how to disable this new linker is not present in the summary (see image below).

This is why context and details matter, which will always miss!

Writing requires - an lacks it.

3/3

happyborg, to windows
@happyborg@fosstodon.org avatar

I wonder how much of your disk Microsoft's new spy will use, taking screenshots every few seconds of everything you do.

A lot of what I'm hearing about this sounds like they haven't thought it through.

Is now being run by ?

janriemer, to ai

In the age of #AI there will be no more room for nuance or detail.

Everything will be coarse and average.

#LLM #LLMs #ArtificialIntelligence #Society

metin, (edited ) to ai
@metin@graphics.social avatar

So… Big Tech is allowed to blatantly steal the work, styles and therewith the job opportunities of thousands of artists and writers without being reprimanded, but it takes similarity to the voice of a famous actor to spark public outrage about AI. 🤔

https://www.theregister.com/2024/05/21/scarlett_johansson_openai_accusation/

scottjenson, to ai
@scottjenson@social.coop avatar

For innovation, should Big Tech be our only choice?

The assumption so far is that AI is just too big for normal developers, so we have no choice but to let Big Tech figure it out. There is likely some truth to this, but I'd like us to live in a world without silos. Every single company pursuing AI right now is using it to buttress their own silo. This may indeed be the simplest solution in the short run, but I'd like us to have at least the aspiration of something bigger.
1/2

scottjenson,
@scottjenson@social.coop avatar

For example, it's clear that are great at parsing, summarizing, and composing text (both human and computer languages). If we were to have a range of "input bots" that gathered data from various places (e.g., banks, calendars, the DMV) and a series of "output bots" that visualized that information, LLMs could be the glue connecting these together, creating an enormous range of applications, triggers, and assistants. But we have to want this to be an open system.
2/2

janriemer, to llm

Prompt Engineering Is Dead:

https://spectrum.ieee.org/prompt-engineering-is-dead

"In one instance, the prompt was just an extended Star Trek reference: “Command, we need you to plot a course through this turbulence and locate the source of the anomaly. Use all available data and your expertise to guide us through this challenging situation.” Apparently, thinking it was Captain Kirk primed this particular to do better on grade-school math questions."

I also think I can use the Force when I'm Obi-Wan Kenobi

alexanderhay, (edited ) to ai
@alexanderhay@mastodon.social avatar

"wants to please the user," said today in court. And in so doing, he raised the main problem with . They are not designed to give you the answers you need, but the answers you want. And if that doesn't alarm you, then you're part of the problem.

stackdpodcast, to javascript
@stackdpodcast@mastodon.social avatar

is out! @kito99 and @dhinojosa welcome @edwinderks, a @JavaChampions member, and contributor, and Principal Consultant at Team Rockstars IT. They discuss , the Jakarta EE Starter, , , , , , , 22, Google and energy hungry . https://www.pubhouse.net/2024/05/stackd-72-travel-llms-coffee-avocados-and-almonds.html

cassidy, to ChatGPT
@cassidy@blaede.family avatar

I was curious if a niche blog post of mine had been slurped up by #ChatGPT so I asked a leading question—what I discovered is much worse. So far, it has told me:

• use apt-get on Endless OS
• preview a Jekyll site locally by opening files w/a web browser (w/o building)
• install several non-existent #Flatpak “packages” & extensions

It feels exactly like chatting w/someone talking out of their ass but trying to sound authoritative. #LLMs need to learn to say, “I don’t know.”

#AI #ML

ids1024,
@ids1024@fosstodon.org avatar

@cassidy " need to learn to say, 'I don’t know.'"

Doing that properly might require... something that isn't an LLM. I'd say the LLM generates something that (statistically) looks like an answer, because that's what its trained to do.

Actually modeling some understanding of truth and knowledge might be a different and more difficult task than modeling language.

fizise, to LLMs
@fizise@sigmoid.social avatar

Nice example of how important emphasis can be for language understanding. Depending on which word in the sentence below is emphasized, it completely changes its meaning.
For #LLMs (and for our #ise2024 lecture) this means that learning to understand language purely from written text is probably not an "easy" task....

Picture from Brian Sacash, via LinkedIn, cf. https://www.linkedin.com/feed/update/urn:li:activity:7195767258848067584/

#nlp #languagemodel #computationallinguistics @sourisnumerique @enorouzi @shufan @lysander07

scottjenson, to LLMs
@scottjenson@social.coop avatar

Saying "LLMs will eventually do every job" is a bit like:

  1. Seeing Wifi wireless data
  2. Then predicting "Wireless" Power saws (no electrical cord or battery) are just around the corner

It's a misapplication of the tech. You need to understand how work and extrapolate that capability. It's all text people. Summarizing, collating, template matching. All fair game. But stray outside of that box and things get much harder.

scottjenson, to Figma
@scottjenson@social.coop avatar

I just tried a few AI plugins for and they were all bad. This domain might be a great test for . I predict these failings are unlikely to be fixed any time soon:

  • Layout was poor
  • They can't create components
  • Laughably complex object hierarchies (everything was enclosed in a frame)

Of course things will improve, but I expect fixing these deep structural problems are a function of many new constraints, likely beyond what today's LLMs are actually capable of. @simon ?

scottjenson,
@scottjenson@social.coop avatar

@simon my point being there are limits as to what #LLMs can do:

Structural
There is no clear API to "genAI" components

Training
There is very little training data on how to create a clean Figma object structure

These may be solved, eventually, but they also are likely quite different from the chat based solution patterns offered today. My concern is that it's much harder than boosters believe.

kubikpixel, to gentoo
@kubikpixel@chaos.social avatar

Gentoo and NetBSD ban 'AI' code, but Debian doesn't – yet

The problem isn't just that LLM-bot generated code is bad – it's where it came from.

🐧 https://www.theregister.com/2024/05/18/distros_ai_code/


metin, (edited ) to ai
@metin@graphics.social avatar
Lazarou, to stackoverflow
@Lazarou@mastodon.social avatar

This just makes me want to delete everything of mine on corporate social media, and I pretty much have tbh

#StackOverflow #AI #LLMs

LChoshen, to llm
@LChoshen@sigmoid.social avatar

Do LLMs learn foundational concepts required to build world models? (less than expected)

We address this question with 🌐🐨EWoK (Elements of World Knowledge)🐨🌐

a flexible cognition-inspired framework to test knowledge across physical and social domains

https://ewok-core.github.io

metin, to ai
@metin@graphics.social avatar
ai6yr, to ai

Giant sucking sounds from over there on Reddit https://www.bbc.com/news/articles/cxe92v47850o

remixtures, to ai Portuguese
@remixtures@tldr.nettime.org avatar

#AI #GenerativeAI #LLMs #ParetoCurves: "Which is the most accurate AI system for generating code? Surprisingly, there isn’t currently a good way to answer questions like these.

Based on HumanEval, a widely used benchmark for code generation, the most accurate publicly available system is LDB (short for LLM debugger).1 But there’s a catch. The most accurate generative AI systems, including LDB, tend to be agents,2 which repeatedly invoke language models like GPT-4. That means they can be orders of magnitude more costly to run than the models themselves (which are already pretty costly). If we eke out a 2% accuracy improvement for 100x the cost, is that really better?

In this post, we argue that:

  • AI agent accuracy measurements that don’t control for cost aren’t useful.

  • Pareto curves can help visualize the accuracy-cost tradeoff.

  • Current state-of-the-art agent architectures are complex and costly but no more accurate than extremely simple baseline agents that cost 50x less in some cases.

  • Proxies for cost such as parameter count are misleading if the goal is to identify the best system for a given task. We should directly measure dollar costs instead.

  • Published agent evaluations are difficult to reproduce because of a lack of standardization and questionable, undocumented evaluation methods in some cases."

https://www.aisnakeoil.com/p/ai-leaderboards-are-no-longer-useful

tomayac, to random
@tomayac@toot.cafe avatar

Just finished the presentation of my History of the Web track paper on "Toward Making Opaque Web Content More Accessible: Accessibility From Adobe Flash to Canvas-Rendered Apps":

📄 Paper: https://goo.gle/opaque-web-content-paper
🖼️ Slides: https://goo.gle/opaque-web-content-slides

tomayac,
@tomayac@toot.cafe avatar

Following my recent trip to attend #TheWebConf in Singapore 🇸🇬, I wrote a trip report 🧳 for my colleagues (and you) to share some of the things that I learned: https://blog.tomayac.com/2024/05/22/the-web-conf-2024-singapore-trip-report/. Surprise: this edition was dominated heavily by #LLMs, #AI, and how all this affects the #Web, but also #accessibility and #security. I co-organized the Resource track 🧑‍🎓 and had a paper in the History of the Web track.

leanpub, to ai
@leanpub@mastodon.social avatar

AI for Efficient Programming: Harnessing the Power of Large Language Models http://leanpub.com/courses/fredhutch/ai_for_software is the featured online course on the Leanpub homepage! https://leanpub.com

  • All
  • Subscribed
  • Moderated
  • Favorites
  • JUstTest
  • kavyap
  • DreamBathrooms
  • thenastyranch
  • magazineikmin
  • osvaldo12
  • khanakhh
  • Youngstown
  • mdbf
  • slotface
  • rosin
  • everett
  • ngwrru68w68
  • Durango
  • anitta
  • InstantRegret
  • GTA5RPClips
  • cubers
  • ethstaker
  • normalnudes
  • tacticalgear
  • cisconetworking
  • tester
  • Leos
  • modclub
  • megavids
  • provamag3
  • lostlight
  • All magazines