모델 압축과 Predictive Caching을 통한 오프라인 AI 튜터 구현 및 학습 세션 275% 증대
Building an AI Tutor That Works Without Internet: Lessons from Rural Ethiopia
Building an AI Tutor That Works Without Internet: Lessons from Rural Ethiopia
Model Compression Techniques for Edge Deployment
Why Building AI for Non-English Speakers is Harder Than You Think (And How I Did It Anyway)
Half-Quadratic Quantization of large machine learning models
Gemma 3n fully available in the open-source ecosystem!
Introducing AutoRound: Intel’s Advanced Quantization for LLMs and VLMs
SmolVLM2: Bringing Video Understanding to Every Device
SmolVLM Grows Smaller – Introducing the 256M & 500M Models!
Welcome to the Falcon 3 Family of Open Models!
SmolVLM - small yet mighty Vision Language Model
Welcome aMUSEd: Efficient Text-to-Image Generation
Introducing Würstchen: Fast Diffusion for Image Generation
Making LLMs lighter with AutoGPTQ and transformers
Open-sourcing Knowledge Distillation Code and Weights of SD-Small and SD-Tiny
Stable Diffusion XL on Mac with Advanced Core ML Quantization
Faster Stable Diffusion with Core ML on iPhone, iPad, and Mac
Smaller is better: Q8-Chat, an efficient generative AI experience on Xeon