Replies: 6 comments 10 replies
-
In my test, this bug occurs when the model surpasses the hardware capacity (out of memory). It seems that Termux utilizes less RAM compared to Flutter apps. You can attempt to utilize Q2_K, but the token speed of the 7B model is still relatively slow on medium-spec smartphone |
Beta Was this translation helpful? Give feedback.
-
Yeah certain 7B models crash the app because they use too much memory. In my testing if you're going to attempt to use 7B models you should try to use smaller Quant's. |
Beta Was this translation helpful? Give feedback.
-
thanks for the update, the new release(maid 1.0.5 apk) it's very stable now, I can run pygmalion2 7b 3ks slowly and without any crashes, thanks again |
Beta Was this translation helpful? Give feedback.
-
Vali is developing a front-end like ST for android https://github.com/Vali-98/ChatterUI |
Beta Was this translation helpful? Give feedback.
-
can you please update the llama.cpp of the app to support stablelm models? |
Beta Was this translation helpful? Give feedback.
-
You need to use the latest actions release
https://github.com/MaidFoundation/Maid/actions/workflows/main.yml
Until I do another fixed point release.
With regards,
Dane Madsen
…________________________________
From: Ar57m ***@***.***>
Sent: Monday, November 20, 2023 3:01:17 AM
To: MaidFoundation/Maid ***@***.***>
Cc: Dane Madsen ***@***.***>; State change ***@***.***>
Subject: Re: [MaidFoundation/Maid] feedback - 7b models crashing (Discussion #28)
I've notice you have updated llama, but tried to run a stablelm3b model on the 1.1.2 version and is crashing
—
Reply to this email directly, view it on GitHub<#28 (reply in thread)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/ACYA2IZO76R7WWKIJZRJIXTYFI3N3AVCNFSM6AAAAAA6RTHZTOVHI2DSMVQWIX3LMV43SRDJONRXK43TNFXW4Q3PNVWWK3TUHM3TMMJSGE3DO>.
You are receiving this because you modified the open/close state.Message ID: ***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
hey, I am running your app on a Galaxy M52, 3B models runs fine on it, but 7b as pygmalion2 7b(3ks) and mystral pygmalion 7b(3km) it crashes and closes the app, on llama.cpp through termux it runs around 0.1~0.2 tokens/second.
Beta Was this translation helpful? Give feedback.
All reactions