• 9 Posts
  • 77 Comments
Joined 2 years ago
cake
Cake day: June 21st, 2023

help-circle



  • Yeah I tested with lower numbers and it works, I just wanted to offload the whole model thinking it will work, 2GB it’s a lot. With other models it prints about 250MB when fails and if you sum up the model size it’s still well below the iGPU free memory so I dont get it… anyway, I was thinking about upgrading the memory to 32GB or may be 64GB but I hesitate because with models around 7GB and CPU only I get around 5 t/s and with 14GB 2-3 t/s, so I run one of around 30GB I guess it will get around 1 t/s? My supposition is that increasing RAM doesn’t increase performance per se, just let’s you upload bigger models to memory, so performance is approximately linear on model size… what do you think?


  • I get an error when offloading the whole model to GPU

    ./build/bin/llama-cli -m ~/software/ai/models/deepseek-math-7b-instruct.Q8_0.gguf -n 200 -t 10 -ngl 31 -if

    The relevant output is:

    llama_model_load_from_file_impl: using device Vulkan0 (Intel® Iris® Xe Graphics (RPL-U)) - 7759 MiB free

    print_info: file size = 6.84 GiB (8.50 BPW)

    load_tensors: loading model tensors, this can take a while… (mmap = true) load_tensors: offloading 30 repeating layers to GPU load_tensors: offloading output layer to GPU load_tensors: offloaded 31/31 layers to GPU load_tensors: Vulkan0 model buffer size = 6577.83 MiB load_tensors: CPU_Mapped model buffer size = 425.00 MiB

    ggml_vulkan: Device memory allocation of size 2013265920 failed ggml_vulkan: vk::Device::allocateMemory: ErrorOutOfDeviceMemory llama_kv_cache_init: failed to allocate buffer for kv cache llama_init_from_model: llama_kv_cache_init() failed for self-attention cache common_init_from_params: failed to create context with model ‘~/software/ai/models/deepseek-math-7b-instruct.Q8_0.gguf’ main: error: unable to load model

    It seems to me that there is enough room for the model, but I don’t know what “Device memory allocation of size 2013265920” means.





  • corvus@lemmy.mlOPtoLocalLLaMA@sh.itjust.worksModels not loading into RAM
    link
    fedilink
    English
    arrow-up
    1
    ·
    edit-2
    26 days ago

    Yes, gpt4all runs it in cpu mode, the gpu option does not appear in the drop-down menu, which means the gpu it’s not supported or there is an error. I’m trying to run the models with the SyCL backend implemented in llama.cpp that performs specific optimizations for cpu+gpu with the Intel DPC++/C++ Compiler and the OneAPI Toolkit.

    Also try Deepseek 14b. It will be much faster.

    ok, I’ll test it out.


  • corvus@lemmy.mlOPtoLocalLLaMA@sh.itjust.worksModels not loading into RAM
    link
    fedilink
    English
    arrow-up
    2
    ·
    edit-2
    27 days ago

    I tried llama.cpp but I was having some errors about not finding some library so I tried gpt4all and it worked. I’ll try to recompilte and test it again. I have a thinkbook with Intel i5-1335u and integrated Xe graphics. I installed the Intel OneAPI toolkit so llama.cpp could take advantage of the SYCL backend for Intel GPUs, but I had an execution error that I was unable to solve after many days. I installed the Vulkan SDK needed to compile gpt4all with the hope to being able to use the GPU but gpt4all-chat doesn’t show the option to run from it, so from what I read it means that it’s not supported, but from some posts that I read I should not expect a big performance boost from that GPU.














  • This video has seemingly no sources for its claims.

    It’s just an introductory video. The references are in her book. I counted around 300.

    Lyn Alden is part of “Ego Death Capital”, a venture capital company around cryptocurrencies (https://egodeath.capital/team)

    Only Bitcoin. Bitcoin is not any crypto or altcoin, which I agree most are scams.

    Lyn Alden is the Board Director of Swan Bitcoin - a Bitcoin investment platform (https://www.swanbitcoin.com/alden/) Lyn Alden is not an economist (https://www.lynalden.com/about-lyn-alden/)

    Who are you expecting to make a video about the failure of the current system? A banker?

    Bitcoin cannot be diluted (~27:25) REALITY CHECK: Bitcoin is always being diluted until it reaches its hard limit.

    What she obviously means is that nobody can delute it. It creates new money at a mathematically determined rate.

    The value of Bitcoin has only increased over time (~27:50) REALITY CHECK: The log scale is playing tricks. A linear graph would show how volatile Bitcoin has truly been.

    She doesn’t say that. She says bigger highs and bigger lows, which is true. That doesn’t mean it always increases.

    Bitcoin’s hard limit is likely very dangerous for the network (~29:00): Once the hard limit is reached, it is unclear if people will keep >pumping computing power at it. If the creation of new Bitcoin is no longer allowed, it is possible that transaction fees will need to >be raised to compensate miners.

    Dont worry, it will happen in 2140.

    And of course is Bitcoin propaganda, and more of this quality is needed.