• nifty@lemmy.world
    link
    fedilink
    English
    arrow-up
    0
    ·
    2 months ago

    In my experience, papers which propose numerical solutions cover in great detail the methodology (which relates to some underlying physical phenomena), and also explain boundary conditions to their solutions. In ML/DL papers, they tend to go over the network architecture in great detail as the network construction is the methodology. But the problem I think is that there’s a disconnect going from raw data to features to outputs. I think physics informed ML models are trying to close this gap somewhat.

    • model_tar_gz@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      2 months ago

      As I was reading your comment I was thinking Physics Informed NN’s and then you went there. Nice. I agree.

      I’ve built some models that had a solution constrained loss functions—featureA must be between these values, etc. Not quite the same as defining boundary conditions for ODE/PDE solutions but in a way gets to a similar space. Also, ODE/PDE solutions tend to find local minima and short of changing the initial conditions there aren’t very many good ways of overcoming that. Deep learning approaches offer more stochasticity so converge to global solutions more readily (at the risk of overfitting).

      The convergence of these fields is exciting to watch.

      • nifty@lemmy.world
        link
        fedilink
        English
        arrow-up
        0
        ·
        2 months ago

        Deep learning approaches offer more stochasticity so converge to global solutions more readily (at the risk of overfitting).

        Yeah, thats a fair point and another appealing reason for DL based methods