Tag: memory

  • The Register: AMD secure VM tech undone by DRAM meddling

    Source URL: https://www.theregister.com/2024/12/10/amd_secure_vm_tech_undone/ Source: The Register Title: AMD secure VM tech undone by DRAM meddling Feedly Summary: Boffins devise BadRAM attack to pilfer secrets from SEV-SNP encrypted memory Researchers have found that the security mechanism AMD uses to protect virtual machine memory can be bypassed with $10 of hardware – and perhaps not even that.……

  • Hacker News: I can now run a GPT-4 class model on my laptop

    Source URL: https://simonwillison.net/2024/Dec/9/llama-33-70b/ Source: Hacker News Title: I can now run a GPT-4 class model on my laptop Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses the advances in consumer-grade hardware capable of running powerful Large Language Models (LLMs), specifically highlighting Meta’s Llama 3.3 model’s performance on a MacBook Pro M2.…

  • Simon Willison’s Weblog: I can now run a GPT-4 class model on my laptop

    Source URL: https://simonwillison.net/2024/Dec/9/llama-33-70b/ Source: Simon Willison’s Weblog Title: I can now run a GPT-4 class model on my laptop Feedly Summary: Meta’s new Llama 3.3 70B is a genuinely GPT-4 class Large Language Model that runs on my laptop. Just 20 months ago I was amazed to see something that felt GPT-3 class run on…

  • Hacker News: Buffer Overflow Risk in Curl_inet_ntop and Inet_ntop4

    Source URL: https://hackerone.com/reports/2887487 Source: Hacker News Title: Buffer Overflow Risk in Curl_inet_ntop and Inet_ntop4 Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text addresses vulnerabilities in the Curl and inet_ntop functions relating to buffer overflow risks due to inadequate buffer size validation. This discussion is particularly relevant for professionals involved in software security,…

  • Simon Willison’s Weblog: Meta AI release Llama 3.3

    Source URL: https://simonwillison.net/2024/Dec/6/llama-33/#atom-everything Source: Simon Willison’s Weblog Title: Meta AI release Llama 3.3 Feedly Summary: Meta AI release Llama 3.3 This new Llama-3.3-70B-Instruct model from Meta AI makes some bold claims: This model delivers similar performance to Llama 3.1 405B with cost effective inference that’s feasible to run locally on common developer workstations. I have…

  • Cloud Blog: Locking down Cloud Run: Inside Commerzbank’s adoption of custom org policies

    Source URL: https://cloud.google.com/blog/topics/financial-services/commerzbank-cloud-run-custom-org-policies/ Source: Cloud Blog Title: Locking down Cloud Run: Inside Commerzbank’s adoption of custom org policies Feedly Summary: Usually, financial institutions process multiple millions of transactions daily. Obviously, when running on cloud technology, any security lapse in their cloud infrastructure might have catastrophic consequences. In serverless setups for compute workloads Cloud Run on…

  • Hacker News: Exploring inference memory saturation effect: H100 vs. MI300x

    Source URL: https://dstack.ai/blog/h100-mi300x-inference-benchmark/ Source: Hacker News Title: Exploring inference memory saturation effect: H100 vs. MI300x Feedly Summary: Comments AI Summary and Description: Yes **Summary:** The text provides a detailed benchmarking analysis comparing NVIDIA’s H100 GPU and AMD’s MI300x, with a focus on their memory capabilities and implications for LLM (Large Language Model) inference performance. It…

  • Hacker News: Bringing K/V context quantisation to Ollama

    Source URL: https://smcleod.net/2024/12/bringing-k/v-context-quantisation-to-ollama/ Source: Hacker News Title: Bringing K/V context quantisation to Ollama Feedly Summary: Comments AI Summary and Description: Yes Summary: The text discusses K/V context cache quantisation in the Ollama platform, a significant enhancement that allows for the use of larger AI models with reduced VRAM requirements. This innovation is valuable for professionals…

  • AWS News Blog: New Amazon EC2 P5en instances with NVIDIA H200 Tensor Core GPUs and EFAv3 networking

    Source URL: https://aws.amazon.com/blogs/aws/new-amazon-ec2-p5en-instances-with-nvidia-h200-tensor-core-gpus-and-efav3-networking/ Source: AWS News Blog Title: New Amazon EC2 P5en instances with NVIDIA H200 Tensor Core GPUs and EFAv3 networking Feedly Summary: Amazon EC2 P5en instances deliver up to 3,200 Gbps network bandwidth with EFAv3 for accelerating deep learning, generative AI, and HPC workloads with unmatched efficiency. AI Summary and Description: Yes **Summary:**…