Skip to content
Discussion options

You must be logged in to vote

Hi @rubenvde,

Thank you for reaching out and for exploring SpeziLLM! We’re excited to hear you’re integrating it into your application 🚀

You are absolutely right in your observation: memory usage remains high after the model is loaded and a request is dispatched. Once the associated ChatView (as you described) is dismissed, memory usage should return to normal levels.

If you require more granular control over how the model is managed in memory, feel free to take a look at this branch and pull request. It introduces the ability to explicitly offload the model used in a LLMLocalSession from memory using custom logic, and to reload it again when needed.

Please note that this branch is not ye…

Replies: 2 comments

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
0 replies
Answer selected by rubenvde
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
3 participants