Kolmogorov-Arnold Networks (KANs) and Lennard Jones

Ғылым және технология

KANs have been a hot topic of discussion recently (arxiv.org/abs/2404.19756). Here I explore using them as an alternative to a neural network for a simple atomistic potential using Lennard Jones data.
See kitchingroup.cheme.cmu.edu/bl....

Пікірлер: 16

  • @dennyloevlie768
    @dennyloevlie7682 ай бұрын

    I was just about to reach out to you and ask what you thought of the KAN architecture. So glad to see you already made a video on it! Great video and explanation.

  • @scottotterson3978
    @scottotterson3978Ай бұрын

    I guess the magic is that you can get an analytic expression from a trained KAN -- no idea how much better that expression would be than what you'd get from, say, PySR, but I can imagine that it could be better at extrapolation than an MLP.

  • @bithigh8301
    @bithigh83012 ай бұрын

    nice! can you make a tutorial about using emacs like that? One can go full focus on this env

  • @JasonMitchellofcompsci
    @JasonMitchellofcompsciАй бұрын

    It's basically just a different activation function with more tunable parameters than ReLU, yeah? I actually consider ReLU to be tunable because it's one way to took at your biases. Biases just set the threshold for when a signal goes through or not.

  • @kanalarchis
    @kanalarchis2 ай бұрын

    Thanks, that was a very good video. I don't understand the point of this. We already knew how to interpolate data with splines.

  • @PhysBrain
    @PhysBrain2 ай бұрын

    Failure to extrapolate beyond the training set is a well know consequence of polynomial interpolation/regression. However, as polynomial regressors go, splines tend to be much better behaved outside of distribution than say, power series or Taylor series. Splines are highly flexible polynomial functions, but the basis functions are typically defined such that they have limited support (region over which they are non-zero). That means they do not go off to infinity or contribute other bad behavior to the function when evaluated far from their defined domain. So, at least with spline functions, there is an explicit acknowledgement that they will not produce useful extrapolations beyond the range of the provided data. However, as the paper describes, the limited (local) support is actually what allows KANs to store additional information without forgetting previously learned data. During training, the only spline parameters that are modified are the ones for which their basis functions evaluate to non-zero values (i.e. the parameters that are most closely associated with the new input data). Contrast this with weights on ReLU activation functions which are always modified when the the weighted sum of its inputs (multivariate) are above some threshold. The spline basis functions are only non-zero when their one input parameter (univariate) is within some well-defined range. So only the parameters local to the new data will be modified.

  • @jasmeetsingh4688
    @jasmeetsingh46882 ай бұрын

    how can i install 'kan' package or did you write any script for it?

  • @aaronbelikoff8605

    @aaronbelikoff8605

    2 ай бұрын

    The package is called “pykan” but you import it as kan.

  • @user-rl8to5nc2q
    @user-rl8to5nc2q2 ай бұрын

    I don’t see the advantage. Someone enlighten me?

  • @antonpershin998

    @antonpershin998

    2 ай бұрын

    More imterpretable

  • @user-rl8to5nc2q

    @user-rl8to5nc2q

    2 ай бұрын

    @@antonpershin998 thamks

  • @MaJetiGizzle

    @MaJetiGizzle

    2 ай бұрын

    No catastrophic forgetting.

  • @user-rl8to5nc2q

    @user-rl8to5nc2q

    2 ай бұрын

    @@MaJetiGizzle hm how so?

  • @Chidorin

    @Chidorin

    2 ай бұрын

    less resources needed for final model usage, may be 🤔

  • @starship9629
    @starship96292 ай бұрын

    It is pronounced KolMOgoROV

Келесі