
Happy Monday! ☀️
Welcome to the 199 new hungry minds who have joined us since last Monday!
If you aren't subscribed yet, join smart, curious, and hungry folks by subscribing here.

📚 Software Engineering Articles
Defeating deepfakes by stopping laptop farms and insider threats
Google quantum-proofs HTTPS with clever math tricks
How I use Claude code for productive development workflows
Balyasny built AI research engine for smarter investing decisions
System design interview framework that actually works
🗞️ Tech and AI Trends
Meta's AI smart glasses raise serious privacy concerns
Something brewing with Qwen — OpenAI's new challenger emerges
OpenAI launches GPT-5.4, pushing AI capabilities further
👨🏻💻 Coding Tip
Use ripgrep's multiline mode with type filters for blazing-fast code search across massive codebases
Time-to-digest: 5 minutes

Discord needed visibility into why their Elixir services sometimes lag when handling millions of concurrent users. Their metrics and logs told them what was slow, but not why or how it cascaded through the system. Enter distributed tracing—the observability tool that shows you the complete journey of a user action from API to client.
The challenge: Elixir's message-passing architecture doesn't have built-in metadata layers like HTTP headers, so propagating trace context between processes required building a custom solution without downtime.
Implementation highlights:
Transport: A thin wrapper library: Created an
Envelopeprimitive that automatically wraps messages with trace context, making integration ergonomic and zero-copyGenServer drop-in replacements: Provided wrapped versions of
callandcastfunctions so developers didn't need to think about trace propagationGradual rollout with runtime config: Mixed old and new communication patterns, allowing zero-downtime deployment across thousands of nodes
Dynamic sampling based on fanout size: Detected massive fanouts (messages to millions of sessions) and reduced sampling rates adaptively—from 100% for single recipients to 0.1% for 10k+ recipients
Surgical performance optimizations: Only propagate sampled trace contexts, skip root span creation post-fanout, and fast-path the sampling flag check to dodge expensive unpacking
Results and learnings:
Complete visibility: Track 16-minute session connection delays and quantify user impact during outages with end-to-end traces
Scalable observability: Dropped CPU overhead from tracing by 10+ percentage points through aggressive optimization and smart sampling
Production-ready infrastructure: Deployed across Discord's entire Elixir stack without a single incident or service restart
Discord proved you can bolt sophisticated distributed tracing onto a message-passing architecture without sacrificing performance. Their Transport library is a masterclass in pragmatic observability—simple primitives, massive impact, zero friction.
The lesson here? Sometimes the best infrastructure doesn't solve the problem elegantly. It solves it cheaply. And at Discord's scale, cheap wins you the entire game.

ARTICLE (satellites-go-brrr)
Scenario Modeling and Array Design for Non-Terrestrial Networks (NTNs)
GITHUB REPO (google-do-the-thing)
Google Workspace CLI
ARTICLE (boss-without-the-manual)
Leading Beyond the Framework 🧭 — with Richard Hughes-Jones
ARTICLE (oops-crashed-computers)
How a 12-Word Issue Title Owned 4,000 Developer Machines
ARTICLE (chrome-zoom-zoom)
Amid new competition, Chrome speeds up its release schedule
ARTICLE (keyboard-go-click)
Making keyboard navigation effortless
ARTICLE (robot-brain-time)
Build Your AI Agent the Right Way
ARTICLE (change-makes-scared)
Stability is a Trap
Want to reach 200,000+ engineers?
Let’s work together! Whether it’s your product, service, or event, we’d love to help you connect with this awesome community.

🔒 Meta's AI Smart Glasses Spark Major Privacy Backlash as Workers Report "We See Everything" (6 min)
Brief: Meta's AI-powered Ray-Ban glasses are collecting intimate user data including bathroom visits and naked bodies, which data annotators in Kenya process without users' knowledge; the company's opaque privacy policies and subcontractor practices leave Swedish users unaware that their personal videos are shared globally for AI training, raising serious GDPR compliance concerns.
Brief: Alibaba's lead Qwen researcher Junyang Lin suddenly resigned, triggering a wave of departures from the AI team's core members just as the company released Qwen 3.5, an exceptionally powerful family of open-weight models ranging from 2B to 397B parameters that achieve remarkable performance even at smaller sizes.
Brief: OpenAI released GPT-5.4, its most capable frontier model combining advanced reasoning, coding, and agentic workflows with native computer-use capabilities, achieving state-of-the-art performance on professional knowledge work, achieving 83% parity with industry professionals, and introducing tool search to reduce token usage by 47% while enabling agents to operate across larger tool ecosystems.
Brief: Apple unveils MacBook Neo, its most affordable laptop ever, featuring a durable aluminum design in four colors, a 13-inch Liquid Retina display, A18 Pro chip for 50% faster everyday performance, and up to 16 hours of battery life — all starting at just $599 with availability beginning March 11.
Brief: ETH Zurich researchers found that LLM-generated context files hurt AI coding performance by 3% while increasing costs over 20%, though human-written files provide marginal 4% gains at the cost of higher inference expenses, challenging industry recommendations to include them.

This week’s tip:
Leverage ripgrep's --type and --glob patterns with FPAT-like regex for multi-line matching without loading files into memory. ripgrep is orders of magnitude faster than grep/awk for code search, especially when combined with --multiline (-U) and --max-columns to handle generated or minified code safely.

Wen?
Searching across large TypeScript/Rust codebases for async/await patterns: ripgrep's memory-mapped I/O and parallel search by default beats grep even with xargs -P.
Finding multi-line regex in JSON config or log files: Use --multiline with lookahead to avoid OOM on huge files.
Narrow search scope before slow operations: Combine with --files-with-matches -c to get counts first, then feed results to parallel processing tools like GNU parallel.
Take interest and even delight in doing the small things well.
Jim Rohn


That’s it for today! ☀️
Enjoyed this issue? Send it to your friends here to sign up, or share it on Twitter!
If you want to submit a section to the newsletter or tell us what you think about today’s issue, reply to this email or DM me on Twitter! 🐦
Thanks for spending part of your Monday morning with Hungry Minds.
See you in a week — Alex.
Icons by Icons8.
*I may earn a commission if you get a subscription through the links marked with “aff.” (at no extra cost to you).







