Tag: llama.cpp

  • Hacker News: >8 token/s DeepSeek R1 671B Q4_K_M with 1~2 Arc A770 on Xeon

    Source URL: https://github.com/intel/ipex-llm/blob/main/docs/mddocs/Quickstart/llamacpp_portable_zip_gpu_quickstart.md Source: Hacker News Title: >8 token/s DeepSeek R1 671B Q4_K_M with 1~2 Arc A770 on Xeon Feedly Summary: Comments AI Summary and Description: Yes Summary: The text provides a comprehensive guide on using the llama.cpp portable zip to run AI models on Intel GPUs with IPEX-LLM, detailing setup requirements and configuration steps.…

  • Hacker News: Ollama-Swift

    Source URL: https://nshipster.com/ollama/ Source: Hacker News Title: Ollama-Swift Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text discusses Apple Intelligence introduced at WWDC 2024 and highlights Ollama, a tool that allows users to run large language models (LLMs) locally on their Macs. It emphasizes the advantages of local AI computation, including enhanced privacy,…