This week in Python, the popular topics included the differences between Polars and pandas, the Pex tool for generating Python executable files, and a proposal for a Django project template. Articles covered running local LLMs, scaling Django, Python's garbage collector, rate limiting Python async API requests, and Python dependency management. Interesting projects included the RAG chunking library 'chonkie', AlphaFold 3 inference pipeline, and the TransformerEngine library for accelerating Transformer models on NVIDIA GPUs. Have a great week and happy coding!
The Polars vs pandas difference nobody is talking about
A closer look at non-elementary group-by aggregations.
Pex: A tool for generating .pex (Python EXecutable) files, lock files and venvs
Proposal for a Django project template
The author's take on what could be a project template for Django advanced usage, with modern tooling (for Python and UI dependencies, as well as configuration/environment management), but not too opinionated.
microsoft / autogen
A programming framework for agentic AI 🤖
Everything I've learned so far about running local LLMs
A post about running large language models (LLMs) locally on a computer. It discusses what LLMs are and how to set them up to run on your own machine. The article also covers some of the limitations of LLMs, but highlights their potential for tasks like proofreading and creative writing.
The Practical Guide to Scaling Django
Most Django scaling guides focus on theoretical maximums. But real scaling isn’t about handling hypothetical millions of users - it’s about systematically eliminating bottlenecks as you grow. Here’s how to do it right, based on patterns that work in production.
CPython's Garbage Collector and Its Impact on Application Performance
Tutorial: How to rate limit Python async API requests
With an example that performs 100 simultaneous requests to the Etherscan API
Python, C++ inspired language that transpiles to C and can be embedded within C
Python dependency management is a dumpster fire
This article is all about fire safety techniques and tools. It's about how you should think about dependency management, which tools you should consider for different scenarios, and what trade offs you'll have to make. Finally, it exposes the complexity and lingering problems in the ecosystem.
NanoDjango - single-file Django apps | uv integration
NanoDjango is a cool package that lets you build small scripts using all the power of Django, and also supports django-ninja for APIs. We'll dive into NanoDjango in this video, and will use uv and inline script metadata for dependency management.
Thoughts on Django’s Core
Django's longevity is attributed to its stable core, time-based releases, and API stability policy. While there's enthusiasm for expanding Django's features, the author argues that the core should remain focused and prioritize stability. Instead, the community should embrace third-party packages as a way to innovate and extend Django's capabilities without compromising its core.
Understanding Multimodal LLMs
An introduction to the main techniques and latest models.
Flash Attention derived and coded from first principles with Triton (Python)
This video provides an in-depth, step-by-step explanation of Flash Attention, covering its derivation, implementation, and underlying concepts. The presenter explains CUDA, Triton, and autograd from scratch, then derives and codes both the forward and backward passes of Flash Attention.
ML in Go with a Python Sidecar
Python Hub Weekly Digest for 2024-11-17
chonkie
The no-nonsense RAG chunking library that's lightweight, lightning-fast, and ready to CHONK your texts.
alphafold3
AlphaFold 3 inference pipeline.
pipe-operator
Elixir's pipe operator in Python.
Muon
Muon optimizer for neural networks: >30% extra sample efficiency, <3% wallclock overhead.
Protenix
A trainable PyTorch reproduction of AlphaFold 3.
Cosmos-Tokenizer
A suite of image and video neural tokenizers.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
BeamerQt
PyQt-based application to create Beamer-LaTeX Presentations.
Project by Ruslan Keba. Since 2012. Powered by Python. Made in 🇺🇦Ukraine.