The hottest Performance Tuning Substack posts right now

And their main takeaways
Category
Top Technology Topics
System Design Classroom 299 implied HN points 16 May 24
  1. Getting timeouts right is important. If you wait too long, your system slows down, but if you timeout too fast, you might miss a successful call.
  2. Circuit breakers help manage failures. They quickly stop requests to a failing service, allowing your system to recover faster.
  3. Bulkheads keep parts of your system separate. If one part fails, the others keep working, preventing a complete shutdown of the system.
The Tech Buffet 59 implied HN points 06 Nov 23
  1. You can index data in different ways to improve how retrieval works. This means you don't always have to use the same data for both indexing and retrieving.
  2. One method is to break chunks of data into smaller parts. This helps ensure that the information retrieved is more relevant to what the user is looking for.
  3. Another approach is to index data by the questions they answer or their summaries. This makes it easier to find the right content, even if a user isn't very clear in their queries.
Sonal’s Newsletter 19 implied HN points 29 Jul 23
  1. Performance tuning Snowpark on Snowflake can significantly reduce processing time, from half a day to half an hour.
  2. Utilizing the query profiler by Snowflake and making targeted optimizations can have a high impact on performance.
  3. Optimizations like converting UDTFs to UDFs, caching Dataframes, and using batch size annotations can further optimize Snowpark workflows.
Thái | Hacker | Kỹ sư tin tặc 0 implied HN points 26 Jul 07
  1. Having 'hàng khủng' (powerful resources) isn't always good. Inappropriate use can lead to system issues like overloading.
  2. Hugepages in Linux provide benefits by reducing page table lookup time and ensuring non-swappable memory for important services.
  3. Understanding the tools and resources you are using thoroughly is crucial to avoid unintended consequences.
DataSketch’s Substack 0 implied HN points 14 Oct 24
  1. Properly configuring resources in Spark is really important. Make sure you adjust settings like memory and cores to fit your cluster's total resources.
  2. Good data partitioning helps Spark job performance a lot. For example, repartitioning your data based on a relevant column can lead to faster processing times.
  3. Using broadcast joins can save time and reduce workload. When joining smaller tables, broadcasting can make the process much quicker.
Get a weekly roundup of the best Substack posts, by hacker news affinity: