swe

swe

503 bookmarks
Newest
System Design in a Hurry
System Design in a Hurry
Everything you need to quickly get prepared for FAANG system design interviews. Written by former Meta and Amazon interviewers, this guide breaks down the core concepts, patterns, frameworks, and technologies needed to ace your system design interviews. It also breaks down some of the most commonly asked system design questions and provides detailed answers.
·hellointerview.com·
System Design in a Hurry
Handling billions of invocations – best practices from AWS Lambda | AWS Compute Blog
Handling billions of invocations – best practices from AWS Lambda | AWS Compute Blog
This post is written by Anton Aleksandrov, Principal Solution Architect, AWS Serverless and Rajesh Kumar Pandey, Principal Engineer, AWS Lambda AWS Lambda is a highly scalable and resilient serverless compute service. With over 1.5 million monthly active customers and tens of trillions of invocations processed, scalability and reliability are two of the most important service […]
·aws.amazon.com·
Handling billions of invocations – best practices from AWS Lambda | AWS Compute Blog
Just because you’re getting an index scan, doesn't mean you can’t do better! - pgMustard
Just because you’re getting an index scan, doesn't mean you can’t do better! - pgMustard
An issue I often see folks missing when reviewing query plans, is that they see that all of their scans involve indexes and they think that the query is likely already as fast (or efficient) as it can be. In this post we’ll look through several examples, some things to look out for, and how to then
·pgmustard.com·
Just because you’re getting an index scan, doesn't mean you can’t do better! - pgMustard
How To Scale Your Model
How To Scale Your Model
Training LLMs often feels like alchemy, but understanding and optimizing the performance of your models doesn't have to. This book aims to demystify the science of scaling language models on TPUs: how TPUs work and how they communicate with each other, how LLMs run on real hardware, and how to parallelize your models during training and inference so they run efficiently at massive scale. If you've ever wondered “how expensive should this LLM be to train or “how much memory do I need to serve this model myself” or “what's an AllGather”, we hope this will be useful to you.
·jax-ml.github.io·
How To Scale Your Model
Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog
Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog
A publicly accessible database belonging to DeepSeek allowed full control over database operations, including the ability to access internal data. The exposure includes over a million lines of log streams with highly sensitive information.
·wiz.io·
Wiz Research Uncovers Exposed DeepSeek Database Leaking Sensitive Information, Including Chat History | Wiz Blog