Home / Companies / Modal / Hacker News

Modal on HN

47 posts with 1+ points since 2022

Filters
Since:
Posts by Month (47 total)
Hacker News Posts
Title Points Comments Date
DoppelBot: Replace Your CEO with an LLM 232 -- 2025-02-04
The Missing Nvidia GPU Glossary 230 -- 2025-01-12
'I paid for the whole GPU, I am going to use the … 154 -- 2025-05-07
Lambda on hard mode: serverless HTTP in Rust 131 -- 2024-03-16
Static IPs for Serverless Containers 125 -- 2024-12-02
Three types of LLM workloads and how to serve them 75 -- 2026-01-21
Linear Programming for Fun and Profit 62 -- 2025-05-09
GPU Programming Glossary 13 -- 2024-12-12
Modal is now generally available 11 -- 2023-10-10
Catching crypto miners using syscall signatures 9 -- 2024-06-07
Checkpoint/restore for sub-second container startup 9 -- 2025-01-29
GPU Memory Snapshots: fast container cold boots 9 -- 2025-07-31
DoppelBot: Replace Your CEO with an LLM 8 -- 2023-05-15
Embedding (RAG) all of Wikipedia in less than 15 minutes 7 -- 2024-01-24
The future of AI needs more flexible GPU capacity 6 -- 2024-10-25
How to beat proprietary embedding models with open-source 6 -- 2024-04-29
Generating diffusion QR codes that work 5 -- 2025-07-02
We reverse-engineered Flash Attention 4 5 -- 2025-09-26
Beat GPT-4o at Python by searching with 100 dumb LLaMAs 4 -- 2024-08-06
Modal is GA and raised a 16M Series A 4 -- 2023-10-10
The LLM Engine Advisor 4 -- 2025-06-03
Dollars per Token Considered Harmful 4 -- 2025-07-16
Transcribe speech 100x faster and 100x cheaper with open models 4 -- 2025-07-28
Modal Notebooks, a real-time collaborative notebook with cloud GPUs 4 -- 2025-09-09
Modal Notebooks: How we built a cloud GPU notebook that boots in … 4 -- 2025-09-17
A beginner's guide to LLM fine-tuning 3 -- 2023-11-08
Modal – Run code in the cloud without managing your own infrastructure 3 -- 2023-01-04
Inside vLLM: Anatomy of a High-Throughput LLM Inference System 3 -- 2025-09-13
Modal's $87M Series B 3 -- 2025-09-29
One second voice-to-voice latency with just open models 3 -- 2025-11-09
Agents need good developer experience too 3 -- 2025-11-20
Host overhead is killing your inference efficiency 3 -- 2025-11-19
Keeping 20k GPUs Healthy 3 -- 2026-01-09
Modal now charging for reserved containers(minimum of 0.125 cores per container) 2 -- 2024-07-23
Using CUDA on Modal 2 -- 2024-06-24
Run GPU Jobs from Airflow 2 -- 2024-06-21
How Ramp automated receipt processing with fine-tuned LLMs 2 -- 2024-04-02
Modal Launches Sandboxes 2 -- 2025-01-21
Modal SDKs for JavaScript and Go 2 -- 2025-04-30
Modal's Serverless KV Store Now Scales to Infinity 2 -- 2025-05-20
The GPU Glossary: Performance 2 -- 2025-09-04
Finetune Any Llama in Minutes on Modal 1 -- 2023-12-01
Modal – an end-to-end stack for cloud compute 1 -- 2022-12-23
Tidbyt Is Joining Modal 1 -- 2024-12-02
Using the Lamborghini of inference engines for serverless Llama 3 1 -- 2025-04-21
Introducing: B200s and H200s on Modal 1 -- 2025-06-04
The LLM Engine Almanac 1 -- 2025-06-09