• jj4211@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 month ago

      Certainly his use of LLM was stupidly egregious, but he found that even by those standards the math results underpinning the LLM were way off.

  • Pommes_für_dein_Balg@feddit.org
    link
    fedilink
    English
    arrow-up
    20
    ·
    1 month ago

    The core functionality is simple:

    Automatically, upon each payment, add the expense to my app
    Update an Apple Watch complication with the % of my monthly budget spent
    Categorize the purchase for later analysis

    Can someone enlighten me? I don’t understand why you need AI for this in the first place.

  • Coolcoder360@lemmy.world
    link
    fedilink
    English
    arrow-up
    5
    ·
    1 month ago

    I went with quantized Gemma

    Well, was it quantized in a way that iphone 16 supports?

    Often it’s the quantization where things break down, and the hardware needs to support the quantization, can’t run FP16 on int8 hardware… And sometimes the act of quantization can cause problems too.

    And yeah, LLMs are likely going to be very hit or miss anyway.

  • Treczoks@lemmy.world
    link
    fedilink
    English
    arrow-up
    3
    arrow-down
    3
    ·
    1 month ago

    He combines LLMs with numbers and wonders why this does not work? Under which rock does he live?

    • festus@lemmy.ca
      link
      fedilink
      English
      arrow-up
      4
      ·
      1 month ago

      I think you missed the point of his post. His issue is that the numeric operations the phone executes to run the LLM is producing garbage. Arguably this could break all kinds of neural networks, such as voice transcription. He’s not complaining that the LLMs are themselves unable to properly perform math.

    • partial_accumen@lemmy.world
      link
      fedilink
      English
      arrow-up
      3
      ·
      edit-2
      1 month ago

      Under which rock does he live?

      Under the rock where reading comprehension exists apparently.

      Where he was prompting for “What is 2+2?” to the LLMs, the accuracy of the answer was immaterial. At that step he was comparing two systems and simply needed a static question to give both system to compare the internal processes to determine why they arrived at different outputs (or a what appeared to be race condition/infinite loop for one) when the result should be identical to both irrespective of how right or wrong the answer is to the prompt. The LLM answer from the LLM could have been “ham sandwich” and it still would have served his purposes.