@major and like in those other hype cycles, it’s so important to find the people actually doing the work to bring something new into the world. So much noise from others who are simply standing near the people doing the work.
Andrej Karpathy just released a new repo with an implementation of training LLM with pure C/Cude with a few lines of code 🚀. This repo, according to Andrej Karpathy, is still WIP, and the first working example is of GPT-2 (or the grand-daddy of LLMS 😅) 👇🏼
This weekend, on Sunday afternoon non-important facts - the PyTorch name👇🏼
The #PyTorch library was built as a wrapper for the Torch open-source project. Therefore, it was named #Python Torch and, in short, PyTorch.
Torch is a machine-learning library and scientific computing framework based on Lua (I just learned about this language). The Torch library has a deep learning implementation in C, which PyTorch was primarily based on.
Hey, any Fedora folks out there ( also could include future Fedora people!) with involvement in upstream Pytorch and related development communities? #pytorch
@mattdm Depends on what you mean by involvement, but I do data science stuff in my dayjob and contribute fixes and reports for pytorch related things when our CI identifies issues.
Now again #LLMs: if you do not want to train your own #ai foundation model, you can patch it with so-called #adapters. Benjamin Trim talked about their own open-source adapter micro framework: #Refiners work on top of #PyTorch and use declarative layers to patch models, context API to store state. #fosdem2024
A new crash course for getting started with #CUDA with #Python by Jeremy Howard 🚀. CUDA is NVIDIA's programming model for parallel computing on GPUs. CUDE is being used by tools such as #PyTorch#tensorflow and other #deeplearning and LLMs frameworks to speed up calculations. The course covers the following topics:
✅ Setting up CUDA
✅ CUDA foundation
✅ Working with Kernel
✅ CUDA with PyTorch