#nanogpt — Public Fediverse posts
Live and recent posts from across the Fediverse tagged #nanogpt, aggregated by home.social.
-
🎉 Behold, the #NanoGPT Slowrun: a marvel of 10x data efficiency that nobody asked for, achieved by throwing infinite compute at the problem like a toddler with a tantrum. 🤦♂️ Who knew the solution to our data bottleneck woes was to just ignore them and hope they go away with more compute power? 🚀 True #innovation at its finest!
https://qlabs.sh/10x #Slowrun #DataEfficiency #ComputePower #TechHumor #HackerNews #ngated -
NanoGPT Slowrun: 10x Data Efficiency with Infinite Compute
#HackerNews #NanoGPT #Slowrun #DataEfficiency #InfiniteCompute #AI
-
Ah, the 2026 #visionaries have graced us with #NanoGPT Slowrun! 🚀 An open letter to the future, where our data droughts are hilariously overshadowed by compute floods. It's like trying to fill a swimming pool with a teaspoon and wondering why you're not Olympic-ready yet. 🏊♂️🤡
https://qlabs.sh/slowrun #Slowrun #DataDroughts #ComputeFloods #FutureTech #HackerNews #ngated -
NanoGPT Slowrun: Language Modeling with Limited Data, Infinite Compute
#HackerNews #NanoGPT #Slowrun #LanguageModeling #LimitedData #InfiniteCompute
-
NanoGPT Slowrun: Language Modeling with Limited Data, Infinite Compute
#HackerNews #NanoGPT #Slowrun #LanguageModeling #LimitedData #InfiniteCompute
-
NanoGPT Slowrun: Language Modeling with Limited Data, Infinite Compute
#HackerNews #NanoGPT #Slowrun #LanguageModeling #LimitedData #InfiniteCompute
-
NanoGPT Slowrun: Language Modeling with Limited Data, Infinite Compute
#HackerNews #NanoGPT #Slowrun #LanguageModeling #LimitedData #InfiniteCompute
-
NanoGPT Slowrun: Language Modeling with Limited Data, Infinite Compute
#HackerNews #NanoGPT #Slowrun #LanguageModeling #LimitedData #InfiniteCompute
-
Đào tạo mô hình NanoGPT 124m từ đầu chỉ trong 115 phút với card đồ họa 4090 và 1 tỷ token Fineweb! #NanoGPT #AI #MachineLearning #4090 #Fineweb #GPT2 #Training #Model #ArtificialIntelligence #DeepLearning #VietAI #MáyHọc #TríTuệNhânTạo
https://www.reddit.com/r/LocalLLaMA/comments/1ozre2i/nanogpt_124m_from_scratch_using_a_4090_and_a/
-
Today I tried out #AMD #Instinct #MI300a for my existing Deep Learning pipeline. Good news: It worked out of the box. Bad news: For some reason it could not beat my local #Nvidia #1080ti...
After trying all sorts of #ROCM installation methods via prebuild wheels, #apptainer images etc I tried #nanogpt by @karpathy and sure enought: The gpt code ran approx 2x faster than on a #a100 ... I hope that this is due to my programming skills. Not AMD prefering #transformers over #CNNs ... -
Aaaaand the winner iiiis: 😊
1️⃣ How does ChatGPT work, actually? (2023-01)
A relatively easy-to-grasp explanation how #ChatGPT works, yet still accurate on a high level. Starting with an explanation of ChatGPT's little cousin, #NanoGPT, and then describing the differences.Essentially it is a dual translation of #AndrejKarpathy's #NanoGPT video: Format from video to text; target audience from programmer to general public.
Enjoy!
https://netfuture.ch/2023/01/how-does-chatgpt-work-actually/