ML Collective

ML Collective

ML Collective (MLC) is an independent, nonprofit organization, with a mission to make research opportunities accessible and free, by supporting open collaboration in machine learning (ML) research.

Subscribe our channel to support our effort, and follow our events, especially the "Request for Plot" events where people pitch their project ideas and recruit collaborators!

More on our website: mlcollective.org/

Banner Image: "A Violet and Light Pink Tapestry representing the Collective Researcher Brain. Tessellation by M.C. Escher", generated by Nicholas Bardy.

Research Jam #21

Research Jam #21

Synthetic Data @ DLCT

Synthetic Data @ DLCT

Research Jam #20

Research Jam #20

LLM Reasoning @ DLCT

LLM Reasoning @ DLCT

Research Jam #19

Research Jam #19

Training dynamics @ DLCT

Training dynamics @ DLCT

Research Jam #18

Research Jam #18

Research Jam #17

Research Jam #17

Research Jam #16

Research Jam #16

Research Jam #14

Research Jam #14

Research Jam #15

Research Jam #15

RLSbench @ DLCT

RLSbench @ DLCT

Research Jam #13

Research Jam #13

Research Jam #12

Research Jam #12

Reincarnating RL @ DLCT

Reincarnating RL @ DLCT

Research Jam #11

Research Jam #11

Git Re-Basin @ DLCT

Git Re-Basin @ DLCT

ASH, explained

ASH, explained

OptFormer @ DLCT

OptFormer @ DLCT

DALL-E 2 (3/3) @ DLCT

DALL-E 2 (3/3) @ DLCT

Minecraft team jam @ DLCT

Minecraft team jam @ DLCT

Research Jam #10

Research Jam #10

Пікірлер

  • @AdrianOligarch
    @AdrianOligarch2 ай бұрын

    Love this and want to get involved. but looks like discord link is broken!

  • @tihami4008
    @tihami40084 ай бұрын

    Loved the guy's explanation of DNA

  • @ElanRosenfeld
    @ElanRosenfeld5 ай бұрын

    Quick note: @3:10 when I discuss the step size stability threshold, I mistakenly say that the maximum stable step size is 2/η. I meant to say 2/sharpness! Equivalently, if the step size is fixed at η then the stability requirement is sharpness <= 2/η.

  • @enzocaputodevos
    @enzocaputodevos5 ай бұрын

    Discovering this channel is a source of joy for me as I delve into the fundamentals and connect with a supportive community that will offer insights for my projects. It feels like a dream come true!

  • @rezamahmoudi163
    @rezamahmoudi1636 ай бұрын

    please share slide

  • @sahil-vz8or
    @sahil-vz8or10 ай бұрын

    You should provide GitHub link for this work.

  • @AG-sj6ll
    @AG-sj6ll11 ай бұрын

    How can I join these meetups ?

  • @captainamerica6031
    @captainamerica6031 Жыл бұрын

    These video is really helpful. I expect more videos to come.

  • @melon6771
    @melon6771 Жыл бұрын

    I wanna know about Aditya's education background. What he studied and where?

  • @dfmrrd
    @dfmrrd Жыл бұрын

    Undergrad at NYU (CAS/Courant School)

  • @melon6771
    @melon677111 ай бұрын

    @@dfmrrd Source? I didn't find any of his social media except Instagram and he is not even on linkedin ig.

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w Жыл бұрын

    At a startup, would a generalist have greater value?

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w Жыл бұрын

    These are great insights.

  • @JeremyDWilliamsOfficial
    @JeremyDWilliamsOfficial Жыл бұрын

    Thank you for sharing this! One's personal schedule can often make opportunities like this slip out of reach. Having it made available as a recording is most appreciated. I would suggest the viewer also check out Keerthana's website link above. A person to watch and someone who will go very far indeed! :)

  • @user-wr4yl7tx3w
    @user-wr4yl7tx3w Жыл бұрын

    Thanks for posting these as videos

  • @swyxTV
    @swyxTV Жыл бұрын

    very frank and insightful talk, i wish all top industry performers analyzed themselves in public like this. thank you!

  • @boyhappy4125
    @boyhappy4125 Жыл бұрын

    interesting work, and helpful to democratize large RL model~

  • @kazz811
    @kazz811 Жыл бұрын

    Great talk! One point is that the argument for why the lambda is seemingly at 0.5 doesn't seem right. Because these cases are chosen with random seeds, all you can expect is that the distribution of lambda is peaked at 0.5 (for lots and lots of seeds) but it doesn't follow by symmetry that it would be exactly 0.5. That seems to warrant an explanation.

  • @ibraheemmoosa
    @ibraheemmoosa Жыл бұрын

    This was a great talk! I missed the live talk. Thanks for recording this one.

  • @user-of6hv6ws3o
    @user-of6hv6ws3o Жыл бұрын

    Excellent

  • @tahabsn
    @tahabsn Жыл бұрын

    Thank you

  • @abdulmajeedmarek
    @abdulmajeedmarek Жыл бұрын

    Great video!

  • @shawnlee6633
    @shawnlee6633 Жыл бұрын

    47:00

  • @ankitmars
    @ankitmars Жыл бұрын

    Amazing discussion

  • @Janamejaya.Channegowda
    @Janamejaya.Channegowda2 жыл бұрын

    It was wonderful to present our work in this workshop, keep up the great work.

  • @vishalahuja2502
    @vishalahuja25022 жыл бұрын

    Is the book available for free?

  • @charles_irl
    @charles_irl2 жыл бұрын

    It is not. We had limited-time access to drafts for the purposes of the reading group. The link to preorder is here: twitter.com/chipro/status/1526049559540944897?s=20&t=MC7VnVXF0evyvIwDdK0kbA

  • @jonathansum9084
    @jonathansum90842 жыл бұрын

    Support it! I believe ur job is meaningful.

  • @Mutual_Information
    @Mutual_Information2 жыл бұрын

    Nice to see ML collective has a KZread channel. Didn’t watch the whole vid but I know Rosanne is top notch from Twitter :)

  • @charles_irl
    @charles_irl2 жыл бұрын

    Followup on the "Overfitting a Single Batch" discussion from 31:49 -- I did some experiments to follow up on my claim about Transformers not being able to overfit single batches, and I actually want to weaken it a lot. I spent some time with HF Transformers and I've been able to get them to consistently overfit single batches for simple tasks like sequence classification. The other transformer problem I was working had a more difficult task -- image-to-text -- and the implementation was not as well-tested. Results are here: wandb.ai/cfrye59/hf-transformers-overfit-glue-mrpc/sweeps/soi1gyw5?workspace=user-cfrye59 Code is here: colab.research.google.com/drive/1pAWd6MsY4yJrjoqknIbPGxW0usiTTAOJ?usp=sharing The issues with the initialization, normalization, and gradient stability of the TF architectures are real. I've seen them in real-world models, e.g. from BigScience @ HF huggingface.co/bigscience/tr11-176B-ml-logs/tensorboard and in Dall-E mini from Boris Dayma twitter.com/charles_irl/status/1506487785783365633?s=20&t=qcNiNoQ9OF6uJmmqFx20SQ. They may still be related to the failure of the other model+task combo, but they're not as bad as I thought.

  • @mertbozkir
    @mertbozkir2 жыл бұрын

    Will be session 4 uploaded? Or you leave chapter to the participants :)

  • @charles_irl
    @charles_irl2 жыл бұрын

    Actually, Session 4 covers Chapter 5! The book is still being edited, and the numbering of the chapters changed mid-stream. So the next session is this one: kzread.info/dash/bejne/k5iTqtyIldjdoNI.html

  • @mertbozkir
    @mertbozkir2 жыл бұрын

    @@charles_irl Thank you Charles, you're my best teacher ever in ML. 🔥

  • @mdbayazid6837
    @mdbayazid68372 жыл бұрын

    Your glasses remind me about adversarial attack on images. But its really very much colorful and nice @Charles

  • @vinciardovangoughci7775
    @vinciardovangoughci77752 жыл бұрын

    Bummed I missed this one. I’ll have to come do a quick share on progress

  • @bhalchandranaik3514
    @bhalchandranaik35142 жыл бұрын

    The area of the circular cross section perpendicular to the white pole black pole axis reduces when you get closer to the poles. This means that you have fewer shades to chose from. Isn't this invalid and shouldn't the number of shades remain the same?

  • @Janamejaya.Channegowda
    @Janamejaya.Channegowda2 жыл бұрын

    Good initiative, keep up the great work.

  • @DrOsbert
    @DrOsbert2 жыл бұрын

    This is one genuine talk.

  • @leodu561
    @leodu5612 жыл бұрын

    Regarding a minor point around 8:45 mark -- I don't think that conference paper decisions are *that* correlated. Sure, strong papers get in, terrible papers get rejected. But for the mid-tier papers, re-submitting to different conferences is the action based on the belief that the reviewing process from one to the other is more independent (in a probabilistic sense) than correlated. Otherwise, if the reviewing processes are extremely correlated, a rejection from one conference is enough evidence that you shouldn't submit to somewhere else because they are all correlated.

  • @pw7225
    @pw72252 жыл бұрын

    Being open about personal experiences and vulnerabilities is still much too rare in tech. Thank you, Rosanne.

  • @Marcos10PT
    @Marcos10PT2 жыл бұрын

    Hearing one of the ML community's rockstars share such an honest perspective on the struggles we likely all recognize is refreshing and motivating. Thank you for sharing this!!

  • @swyxTV
    @swyxTV Жыл бұрын

    i’m new to her work and need a bit of context - what are you referencing when saying she is a rockstar? (ie what must we know about her?)

  • @golabidoon381
    @golabidoon3812 жыл бұрын

    With due respect, I do not buy the generalist argument for hiring. isn't there already so many people who know a little about everything (like RL, vision, gradient descent, conv nets, etc)? Even any fresh school graduate worked on ML should know a bit about these. Isn't it that, as a research community, we want to understand why deep learning works at the fundamental level rather than treating it as a black box, and that is where we need depth more than ever?

  • @manncodes
    @manncodes2 жыл бұрын

    I think she meant being jack of all trades, master of one. BUT your 'jack' being equivalent to others 'master'. Also, I do agree to your point on interpretability of AI!

  • @sarahjamal86
    @sarahjamal862 жыл бұрын

    Realistic, open, and brave! Thanks a lot for this brilliant talk.

  • @Bianchi77
    @Bianchi772 жыл бұрын

    Nice video, thanks :)

  • @Anonymous-lw1zy
    @Anonymous-lw1zy2 жыл бұрын

    Simply fabulous presentation! I love the thematic connection between the career advice of changing approach to alter outcomes, and the clever tweaking of the model to significantly change its output!

  • @lennymaxmusic9945
    @lennymaxmusic99452 жыл бұрын

    Here fully watching from Jamaica 🇯🇲👍

  • @priyamehta9699
    @priyamehta96992 жыл бұрын

    Incredibly brave and intelligent points to make. I hope it starts a lasting conversation, thanks for starting it.

  • @MewadaDeepak
    @MewadaDeepak2 жыл бұрын

    Fantastic !!! Quiet relatable, inspiring, and very helpful. Thanks a lot, Rosanne :)

  • @dwightzz4449
    @dwightzz44492 жыл бұрын

    It is narrow when ... all of them are trying to hire the same kind of people, with the same rigid rubric. Can not agree more on this, we call this "内卷" in chinese.

  • @ahmedrehab6571
    @ahmedrehab65712 жыл бұрын

    I’m glad that you are an extremely petty person because I am just the same. Thanks for bringing up this topic.

  • @ly1052
    @ly10522 жыл бұрын

    Great talk! Your story almost brings tears to my eyes. 一定要成功呀!

  • @nikre
    @nikre2 жыл бұрын

    great topic.

  • @freedom_fighter
    @freedom_fighter2 жыл бұрын

    Amazing work, everyone!

  • @txwonderwokxc
    @txwonderwokxc3 жыл бұрын

    This is an important message

  • @carsonl941
    @carsonl9413 жыл бұрын

    Thanks Rosanne and Jason!