Robert Sinclair

ZeroWw

AI & ML interests

LLMs optimization (model quantization and back-end optimizations) so that LLMs can run on computers of people with both kidneys. Discord: https://discord.com/channels/@robert_46007

Recent Activity

Organizations

ZeroWw's activity

replied to TuringsSolutions's post 27 days ago
view reply

hence my idea of the SILLY versions... ;)

replied to TuringsSolutions's post 27 days ago
view reply

I am pretty sure that the actual models "AS THEY ARE" could perform 10 times better using chain of thought and some algorithms like these. Without needing a different training. And I think that's probably what CLAUDE does,

Reacted to TuringsSolutions's post with ā¤ļø 27 days ago
view post
Post
2103
Transformers are not all we need, that is being proven repeatedly now as more alternative frameworks emerge. Another such framework is Kolmogorov Arnold Network based Transformers. I break down exactly how these differ from Perceptron based Transformers and give you the link to my Colab where I create a model based on the research paper that absolutely destroys a standard Transformers based model. Check out the video here: https://www.youtube.com/watch?v=Sw0euxNZCc4
Reacted to TuringsSolutions's post with ā¤ļø 27 days ago
view post
Post
1408
I think Reinforcement Learning is the future, for a lot of reasons. I spell them out for you in this video, and also provide you with the basic code to get up and running with Atari and OpenAI Gym. If you want to get into RL, this is your ticket. Link to a cool training montage of the model in the description of the video as well. Step 2 from here would be the full-on training and certification that HuggingFace offers for RL.

https://youtu.be/ueZl3A36ZQk
New activity in TuringsSolutions/Phi3Unlocked 27 days ago

My quants and silly expriment.

2
#1 opened 28 days ago by ZeroWw
New activity in CohereForAI/aya-expanse-8b 28 days ago
Reacted to TuringsSolutions's post with šŸ‘ 29 days ago
view post
Post
1380
Ever wondered how neural networks actually work under the hood?

In my latest video, I break down the core mathematical concepts behind neural networks in a way that's easy for IT professionals to understand. We'll explore:

- Neurons as logic gates
- Weighted sums and activation functions
- Gradient descent and backpropagation

No complex equations or jargon, just clear explanations and helpful visuals!

āž”ļø Watch now and unlock the mysteries of neural networks: https://youtu.be/L5_I1ZHoGnM