• 16 Posts
  • 35 Comments
Joined 7 months ago
cake
Cake day: April 4th, 2024

help-circle


  • merari42@lemmy.worldtoScience Memes@mander.xyzAh yes, regression
    link
    fedilink
    English
    arrow-up
    26
    arrow-down
    3
    ·
    edit-2
    14 days ago

    Machine Learning enthusiasts: Why settle for linear regression when you can deploy a Gradient-Boosted Random Deep Neural Net Surface Vector Cluster that consumes the entire power of Iceland to trace a perfect ∞-dimensional hypersphere around those blue points? Overparameterization is the future!

















  • merari42@lemmy.worldtoScience Memes@mander.xyzPandas
    link
    fedilink
    English
    arrow-up
    4
    ·
    3 months ago

    Have you heard that there are great serialised file formats like .parquet from appache arrow, that can easily be used in typical data science packages like duckdb or polars. Perhaps it even works with pandas (although do not know it that well. I avoid pandas as much as possible as someone who comes from the R tidyverse and try to use polars more when I work in python, because it often feels more intuitive to work with for me.)






  • merari42@lemmy.worldtoScience Memes@mander.xyzanswer = sum(n) / len(n)
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    4 months ago

    This is actually why I find a lot of arguments about AI’s limitations as stochastic parrots very shortsighted. Language, picture or video models are indeed good at memorizing some reasonable features from their respective domains and building a simplistic (but often inaccurate) world model where some features of the world are generalized. They don’t reason per se but have really good ways to look up how typical reasoning would look like.

    To get actual reasoning, you need to do what all AI labs are currently working on and add a neuro-symbolic spin to model outputs. In these approaches, a model generates ideas for what to do next, and the solution space is searched with more traditional methods. This introduces a dynamic element that’s more akin to human problem-solving, where the system can adapt and learn within the context of a specific task, even if it doesn’t permanently update the knowledge base of the idea-generating model.

    A notable example is AlphaGeometry, a system that solves complex geometry problems without human demonstrations and insufficient training data that is based on an LLM and structured search. Similar approaches are also used for coding or for a recent strong improvement in reasoning to solve example from the ARC challenge..