Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[General]: Very little RAM is used #154

Closed
hardWorker254 opened this issue Dec 30, 2024 · 1 comment
Closed

[General]: Very little RAM is used #154

hardWorker254 opened this issue Dec 30, 2024 · 1 comment
Labels
question Further information is requested

Comments

@hardWorker254
Copy link

Why do LLMs like Llama 10B use very little RAM?
Before loading
Screenshot_2024-12-30-16-40-51-076_com finalwire aida64

After loading
Screenshot_2024-12-30-16-41-15-999_com finalwire aida64

What about adding the ability to attach files to text? Maybe use file=path/to/file after promt or add special button? And use the view:
[begin of something.txt file]
Content of something.txt
...
...
[end of something.txt file]

@hardWorker254 hardWorker254 added the question Further information is requested label Dec 30, 2024
@a-ghorbani
Copy link
Owner

Please ensure that auto offload/load is disabled in the app (on the settings page). Otherwise, the model will be offloaded when the app goes into the background. I am going to close this issue as this is not a direct app issue, but if the models doesn't work feel free to open the issue again.

As for file attach it is not planed yet. RAG is challenging to implement for on-device, unless we fit the entire text in the context.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

2 participants