Large Language Models

mauve,
@mauve@mastodon.mauve.moe avatar

This post by @maggie has some great ideas on how tech can help enable applications for regular folks. I've been wanting to do something similar within @agregore some day with local LLMs helping people author p2p web apps.

https://maggieappleton.com/home-cooked-software

happyborg,
@happyborg@fosstodon.org avatar

@mauve
I've yet to get a better response from a local LLM to a code question than I get from a web search or going to StackExchange etc. Are you finding good uses yet?

I confess I haven't tried too hard, but then most people won't and that's the point really anyway. 🤷‍♂️

I expect they should be good for accessibility, such as speech in/out but an not seeing those apps. Why not?! 🤦‍♂️

Although I see Mozilla have put a local LLM in Firefox to generate alt text for images.

@maggie @agregore

lutindiscret,
@lutindiscret@mastodon.libre-entreprise.com avatar

@maggie nice reading. I'm a bit skeptical about LLM but you might be right 🤔 future will tell. As a dev, I'm glad to read non-techie people getting the point of local first app: you did well to introduce the concept

@mauve thanks for sharing

#BarefootDevelopment

rstockm, German
@rstockm@openbiblio.social avatar

Pünktlich zur starten wir im VÖBB einen neuen, experimentellen Dienst: den VÖBB-Chatbot. Als meines Wissens erste (?) deutsche Bibliothek kombinieren wir hier Sprachtalent und "Wissen" eines Large Language Models () mit den vollständigen Metadaten unseres Kataloges (als sog. Embedding).

https://www.voebb.de

Ein thread: 🧵
1/6

aronow,
@aronow@hachyderm.io avatar

Question for my friends:

I have a newly graduated SW Eng (BS in CS) who is struggling to find a job and getting advice to go back and get a Master’s Degree in in order to be more marketable.

I’ve always heard that grad degrees aren’t strictly necessary in SWE to start but is this changing? Are there other time investments that make more sense (open source contributions, certifications, personal projects, etc?)?

What would you give a newly degreed ?

BenjaminHan,
@BenjaminHan@sigmoid.social avatar

1/

With applications more abundant, have researchers been using them to assist their writing? We know they have when writing peer reviews [1], but how about doing so in writing their published papers?

Liang et al comes back to answer this question in [3]. They applied the same corpus-based methodology proposed in [2] on 950k papers published between 2020 to 2024, and the answer is a resounding YES, esp. in CS (up to 17.5%) (screenshot 1).

herr_sander, German
@herr_sander@bildung.social avatar

Im kommenden Schuljahr wollen ein Kollege und ich ein zum Thema anbieten. Das erste Modul - ein Lernpfad zu den Grundfunktionen von - ist als erster Aufschlag fertig. Wir freuen uns über Feedback aus dem : https://www.taskcards.de/#/board/846c4790-d371-4abd-a63d-acae11adbba8?token=df9f302f-2ab2-4b35-a591-ef24d80c7ea0 (mit Texten von @mpblkclp und @isotopp . Der @Linkshaender hat da ja vielleicht auch etwas zu zu sagen :))

isotopp,
@isotopp@chaos.social avatar

@herr_sander @mpblkclp @Linkshaender

https://blog.koehntopp.info/2024/05/31/llms-daheim-mit-ollama.html

In welchem ich als Ergänzung zu meinem vorherigen Artikel einmal die Installation und den Gebrauch von Ollama demonstriere.

Wir installieren Ollama, laden mistral:instruct und verwenden den Ollama Prompt auf einem Mac mini oder einem Windows-Rechner mit Nvidia, um einen Text zusammenfassen zu lassen.

ALTAnlp,
@ALTAnlp@sigmoid.social avatar

While you consider submitting to the Call for Problems for the Shared Task (see link below), we'd like to share with you the winner of the Shared Task, which involved distinguishing -generated from human-generated text.

Here, Rinaldo Gagiano and Lin Tian from use a fine-tuned model with label smoothing, yielding an accuracy of 99.91%. Well done!

🔗 Call for Problems for Shared Task: https://alta2024.alta.asn.au/calls

🔗 Paper: https://aclanthology.org/2023.alta-1.18/

bortzmeyer, French
@bortzmeyer@mastodon.gougere.fr avatar

À première vue, que du vide ronflant et des clichés enfilés à la queue-leu-leu. Mais il y a peut-être du sérieux derrière. Quelqu'un a un avis ? https://www.sorbonne-universite.fr/presse/lancement-des-communs-democratiques-une-initiative-de-recherche-francaise-prend-le-lead

#IA #LLM

breizh,
@breizh@pleroma.breizh.pm avatar

@ScriptFanix @aeris @bortzmeyer @shaft Si on te jettes la balle à main nue au lieu d’utiliser le fusil et sa détonation, t’as une chance.

Bon par contre la situation en plus d’être grotesque me semble peu probable.

shaft,
@shaft@piaille.fr avatar

@breizh Ah le jeté de balles sur les méchants : Hot Shots Part Deux :) @aeris @ScriptFanix @bortzmeyer

sebsauvage, French
@sebsauvage@framapiaf.org avatar


Essayons de résumer où on en est sur ces IA de type LLM (+ une nouvelle faille) : https://sebsauvage.net/links/?0aif1Q

tradjincal,
@tradjincal@ludosphere.fr avatar

@sebsauvage j'ai l'impression aussi qu'economiquement parlant, il y a que Nvidia qui fait sont beurre et que les autres boîtes sur les investissements des banques.

sebsauvage,
@sebsauvage@framapiaf.org avatar

@tradjincal
tout à fait !
J'avais oublié, je l'ajoute.

drahardja,
@drahardja@sfba.social avatar

There was a paper shared recently about the exponential amount of training data to get incremental performance gains in #llm #ai, but I seem to have misplaced it. Do you know what I’m referring to? Mind sharing the link if you have it?

anmey,
@anmey@social.anoxinon.de avatar

I think one of the biggest fears people have about AI is that it isn't perfect as assumed, but that, like us humans, it takes the given information, assumes the most likely outcome, and presents it plausibly.

#ki #LLM #AI

kellogh,
@kellogh@hachyderm.io avatar

@anmey yeah, there’s this paradox — we kinda want computers to think like humans, but when they get plausibly good at it, we complain that they don’t think like computers anymore

grumpybozo,
@grumpybozo@toad.social avatar

I’d like to trust this story, but it fails to link to its supposed source or provide enough info to find it elsewise. A few clicks around the site makes me think that it may well be nothing but a #LLM-composed content farm. https://cosocial.ca/@kgw/112498693958537559

feld,
@feld@bikeshed.party avatar
kornel,
@kornel@mastodon.social avatar

There's an economic curse on Large Language Models — the crappiest ones will be the most widely used ones.

The highest-quality models are exponentially more expensive to run, and currently are too slow for instant answers or processing large amounts of data.

Only the older/smaller/cut-down models are cheap enough to run at scale, so the biggest deployments are also the sloppiest ones.

dvshkn,
@dvshkn@fosstodon.org avatar

I gave the pizza question to Golden Gate Claude. This is truly the people's LLM. Google btfo.

tripplehelix,
@tripplehelix@fosstodon.org avatar

@dvshkn What bridge?

chikim,
@chikim@mastodon.social avatar

Llama.cpp now supports the distributed inference, meaning you can use multiple computers to speed up the response time! Network is the main bottleneck, so all machines need to be hard wired, not connected through wifi. # https://github.com/ggerganov/llama.cpp/tree/master/examples/rpc

cheukting_ho,
@cheukting_ho@fosstodon.org avatar

#PyConIT2024 opening keynote by @t_redactyl - #LLM and illusions

pauleveritt,
@pauleveritt@fosstodon.org avatar

@cheukting_ho @t_redactyl Sunday night: saying bye to Jodie, thinking "whew, glad I’m not getting on a plane tomorrow direct to next conference.”

Today: sad I wasn't there to see Jodie.

  • All
  • Subscribed
  • Moderated
  • Favorites
  • llm
  • DreamBathrooms
  • everett
  • InstantRegret
  • magazineikmin
  • thenastyranch
  • rosin
  • GTA5RPClips
  • Durango
  • Youngstown
  • slotface
  • khanakhh
  • kavyap
  • ngwrru68w68
  • cubers
  • JUstTest
  • ethstaker
  • osvaldo12
  • modclub
  • cisconetworking
  • mdbf
  • tester
  • tacticalgear
  • Leos
  • normalnudes
  • provamag3
  • anitta
  • megavids
  • lostlight
  • All magazines