#amldgenai — Public Fediverse posts
Live and recent posts from across the Fediverse tagged #amldgenai, aggregated by home.social.
-
According to him the main OpenAI ChatGPT advance was UX + fine-tuning.
And that's it for AMLD Conference Generative Learning!
FIN
#AMLD #AMLDGenAI40/🧵
-
Flywheel of test-train-correct-retrain-... => Will RLHF scale (small tuning) or will keep it rolling?
Question: experimentation in high-risk area (healthcare).
Tay is getting mentionned! (fun story, back in 2015 I initially thought @SwiftOnSecurity actually was the actual Tay, because the impression was so good).
More suggestions about human-in the loop (that makes me once again want to cite one of the last @pluralistic 's articles).
39/🧵
-
Pretraining vs Fine-tuning?
OpenAI 2022 - Instruct; (but it was alredy stated in the 2018 GPT-1 paper and done by someone else back in 2020-ish).
Switch to conversational implicature.
Foundational Models fail with it, but not conversationally fine-tuned. Larger size does not help, more training date neither, nor multi-talsk ones.
(Hypothesis: destructive pruning of the neurons during fine-tuning is very powerful: maybe he should talk to Evelina Fedorenko about that)
38/🧵
-
Edward Grefenstette, director of research at Google Deep Mind.
Talk: 4 waves of computational revolution (funny that google search and Web 2.0 are not mentioned).
General formula : inference of joint probaility by optimizing a set of parameters.
=> General translators/autocompelters (T5)
=> joint code + execution (aka the compiler)So why LLMs revolution is happening now?
Nice Chinchilla excerpt.
37/🧵
-
Good robot: qualitative description to metrics to imrpove the robot description.
36/🧵
-
Generalization:
- LLMs being plugged instead of GA algoritms over the "DNA" of machines.
- LLMs for fabrication of designed forms
- Use LLMs to simulate the likely falure modes of physically simulated solutions(in my opinion still does not solve the predifinite vocabulary of DNA in robotics problem).
How to discribe a good robot?
(Evo: reproduction, but it is not acceptable for human assistants).35/🧵
-
Next up - Josie Hughes, Prof at EPFL, robotics design.
=> Embodied intelligence (#ALIFE themes - yay!)
Problem; depends on the human factor, which needs to be multi-disciplinary (I see this is an emergent space where LLMs are supposed to be a solution. I am afraid this will lead them to become a "Google university 2.0" instead...)
GPT3 was used to support iteration (Nature Machine Intelligence 2023 (why am I not surprised...) - so not yay)
34/🧵
-
(I disagree - things such as generalization in rare cases matter).
Nvidia HPARAM kicking vs examples of Paretto Curves + stacking recipes.
Sweeps of hyperparams
> mosaicML github acceleration library link.
33/🧵
-
Goal: time and $ necessary to train networks
- Same cost of inference
- Same hardware and software
- Have to optimize baselinesRecipes examples :
- Selective backprop. Drop examples that has low loss. Don't work raw, because of torch autograd, but selection with forwardpasses can work well (eg low-resolution losses)Pb:
- What is just as good?
- What is acceptable loss of good?=> Paretto frounteers
32/🧵