Demo: Rapid prototyping with Gemma and Llama.cpp

Ғылым және технология

Learn how to run Gemma locally on your laptop using Llama.cpp and quantized models.
Watch more videos of Gemma Developer Day 2024 → goo.gle/440EAIV
Subscribe to Google for Developers → goo.gle/developers
#Gemma #GemmaDeveloperDay

Пікірлер: 57

  • @banzai316
    @banzai3163 ай бұрын

    Very cool, thank you! I like this format with demos. We are developers!

  • @ayoubachak2154
    @ayoubachak21542 ай бұрын

    I've used gemma for a benchmark in a research project I'm working on, where I compared human results against AI, gemma was the closest after bloom 176B, followed by models like mistral instruct 7Band llama 34B, even the 2b version did pretty well, great work team 👏🏻

  • @polish4932

    @polish4932

    2 ай бұрын

    Hi mate, if you'd like to compare diff models for the same question, you do so on Wordware. Highly recommending it! ;)

  • @ayoubachak2154

    @ayoubachak2154

    2 ай бұрын

    @@polish4932 thank you

  • @ser1ification
    @ser1ification3 ай бұрын

    Thanks for the demo!

  • @arpitkumar4525
    @arpitkumar45253 ай бұрын

    Really cool and simple to understand

  • @flynnmc9748
    @flynnmc97483 ай бұрын

    This is a fantastic format for a talk, insightful and engaging for a viewer!!!

  • @GoogleDevelopers

    @GoogleDevelopers

    2 ай бұрын

    Glad you enjoyed this video! 😎

  • @user-eh7uo8hw2v

    @user-eh7uo8hw2v

    2 ай бұрын

    0:21 🎉🎉🎉🎉🎉🎉🎉🎉🎉🎉

  • @judevector
    @judevector3 ай бұрын

    Wow this is so cool 😎, developers changing the world

  • @kevinkawchak
    @kevinkawchakАй бұрын

    Thank you for the discussion.

  • @thesimplicitylifestyle
    @thesimplicitylifestyleАй бұрын

    I was looking for this! Thanks! 😎🤖

  • @zencephalon
    @zencephalon3 ай бұрын

    Good demo, nice tooling suggestions out of this

  • @cho7official55
    @cho7official553 ай бұрын

    Cool demo, I'll try it

  • @arpitkumar4525
    @arpitkumar45253 ай бұрын

    Minimum System Requirements for running a model locally?

  • @forrestfeng1098
    @forrestfeng109818 күн бұрын

    Like it very good sharing.

  • @TheOrator_Ese
    @TheOrator_Ese2 ай бұрын

    Very nice 👌 cool Demo

  • @voidan
    @voidan3 ай бұрын

    how do you connect the LM Studio to llama.cpp? you used a preset which was probably custom.

  • @takudzwamakusha5941
    @takudzwamakusha59413 ай бұрын

    This is so cool.

  • @tonydevelopingstuff
    @tonydevelopingstuff2 ай бұрын

    Very nice!!!!

  • @zoomatic293
    @zoomatic2932 ай бұрын

    This is so cool :)

  • @johnkost2514
    @johnkost25143 ай бұрын

    Wrapped in the llamafile runtime it is an even better single file .. oh yes!

  • @parisneto
    @parisneto3 ай бұрын

    CODE would be awesome, as well as knowing the SPEC of the notebook as it’s easy to buy a sub1k or 5K+ at apple store depending on so many factors…

  • @ChrisTrotter-oj9du
    @ChrisTrotter-oj9du2 ай бұрын

    good, thank you

  • @digisignD
    @digisignD3 ай бұрын

    Cool. Will definitely use this soon

  • @KuldeepSingh-in6js
    @KuldeepSingh-in6js3 ай бұрын

    cool demo

  • @MacGuffin1
    @MacGuffin12 ай бұрын

    Great choice of demo app!!

  • @Daniel-zl7wf
    @Daniel-zl7wf2 ай бұрын

    At 9:03, Gemma shows some solid satire

  • @MyEthan1998
    @MyEthan19982 ай бұрын

    If anyone faces an error on Mac about "network error: self signed certificate", close the app and use terminal, run "NODE_TLS_REJECT_UNAUTHORIZED=0 open -a "LM Studio" " This reopens the app and the error should go away. I have no idea where to put this info sooooo...

  • @indylawi5021
    @indylawi50213 ай бұрын

    Very cool demo 👍. Any chance we can get the source code 😀

  • @airhead2741
    @airhead27413 ай бұрын

    Is this meant to be super accessible? If I have an APU, on a laptop with no GPU or NPU(?), that means I can expect it to run fairly well? Also considerations for a lighter yet usable model?

  • @erickcarrasco1938

    @erickcarrasco1938

    3 ай бұрын

    I tried that in an old APU, very slow generations but the same result.

  • @user-vq8on7dh1y

    @user-vq8on7dh1y

    3 ай бұрын

    Nah, Gemma is just a parot. It is released for fine-tuning, aka research purpose.

  • @monamibob
    @monamibob2 ай бұрын

    Very interesting demo! What kind of extra work would be required to run this without LM Studio? Does Llama.cpp contain the necessarry functions to load models as servers you can interrogate?

  • @A032798
    @A0327983 ай бұрын

    How about windows environment? Is LMstudio/Ollama a better choice?

  • @awakenwithoutcoffee
    @awakenwithoutcoffee2 ай бұрын

    where can we learn to set this up ?

  • @some1rational
    @some1rationalАй бұрын

    Has anyone else tried doing this? I tried following this exactly with LM Studio using the exact model and prompt but I am consistently getting atrocious outputs; the gemma model is just outputting gibberish or incorrectly formatted JSON. I wish there were more details on the presets used.

  • @dtmdota6181
    @dtmdota61812 ай бұрын

    Anyone notice ram usage of 16.68 GB? What was that?

  • @bonadio60
    @bonadio603 ай бұрын

    Very nice, but what is your computer spec? Memory and chip?

  • @darthvader4899

    @darthvader4899

    3 ай бұрын

    Probably m3 max 128gb

  • @JJN631

    @JJN631

    3 ай бұрын

    Gemma 7b can run on a rtx 4060 laptop

  • @svenkoesling
    @svenkoesling3 ай бұрын

    Just my two cents: 1. No explanation on how to connect LM Studio to the Llama.cpp, 2. newest hardware required - at least it doesn't work on my M1 with eight performance cores and 32 GB Ram

  • @nayzawminnaing2562
    @nayzawminnaing25623 ай бұрын

    That's a lot of RAM to run this for me.

  • @devagarwal3250
    @devagarwal32503 ай бұрын

    Pls provide source code also

  • @learnwithdmitri
    @learnwithdmitri3 ай бұрын

    damnn its using 15 gb of ram i have an 8gb m1 i dont think it will work for me..

  • @lorenzo9196

    @lorenzo9196

    3 ай бұрын

    You can download a quantized version 8 maybe 4-5 bits

  • @learnwithdmitri

    @learnwithdmitri

    3 ай бұрын

    @@lorenzo9196 okay i will try

  • @yubrshen
    @yubrshen3 ай бұрын

    What’s the required hardware specs?

  • @deeplearningpartnership
    @deeplearningpartnership2 ай бұрын

    Awesome.

  • @andreawijayakusuma6008
    @andreawijayakusuma6008Күн бұрын

    did gemma should use GPU ? so I wanna try to learn this model, but I didn't want to use GPU

  • @tandaramandaraba
    @tandaramandaraba3 ай бұрын

    wow

  • @AIPeter-dd9hr
    @AIPeter-dd9hr2 ай бұрын

    game using lm studio, interesting.

  • @emmanuelokorafor1705
    @emmanuelokorafor17053 ай бұрын

    It's cool now, but what if each application starts deploying local models. It'll turn our phones into what data centers were meant for thereby reducing costs for large corporations. Trading a few megabytes for faster and more expensive chips.

  • @cmoncmona959

    @cmoncmona959

    2 ай бұрын

    Please Elaborate. What were data centres meant for? Asides hardware to run inference of worldwide requests. If it’s done locally, surely it’s better for redundant tasks. Also, the data centres use a lot of megabytes and expensive chips.

  • @savire.ergheiz
    @savire.ergheiz3 ай бұрын

    Just focus on your existing products Google. Which are a mess 😂

  • @Killputin777
    @Killputin7772 ай бұрын

    never trust google products.

  • @f00kwhiteblackracismwarsh07
    @f00kwhiteblackracismwarsh073 ай бұрын

    Google seems to be trying out too many new things. to me thats a turn off and red flag. everyone is different 🙂

Келесі