Ran TinyLLM on an M1 Mac:Quantized model = 4GB Great for offline QA tasks Too slow for real chat UXStill… seeing it run without the cloud is magical.