WebGPU 기반 ONNX Runtime 도입을 통한 로컬 AI 파이프라인의 브라우저 단일화
Subtitles From a YouTube Link Without Leaving the Browser
Subtitles From a YouTube Link Without Leaving the Browser
Real-Time Speech, Audio, and Facial Analysis in Production AI Systems
Nobody Tells You This About Slow Transformer Models — I Fixed Mine in 3 Steps
Building a Browser-Based AI Background Removal Tool
I Started Building a Roguelike RPG — Powered by On-Device AI #2
Document Structure Extraction with Kreuzberg
Building a Local Voice AI Stack: Whisper + Ollama + Kokoro TTS on Apple Silicon
Supercharge Your Web Apps: Hardware Acceleration with WebGPU and WebAssembly
Transformers.js v4 Preview: Now Available on NPM!
From cloud to developers: Hugging Face and Microsoft Deepen Collaboration
Accelerating SD Turbo and SDXL Turbo Inference with ONNX Runtime and Olive
Accelerating over 130,000 Hugging Face models with ONNX Runtime
Optimum+ONNX Runtime - Easier, Faster training for your Hugging Face models
Accelerated Inference with Optimum and Transformers Pipelines
Scaling-up BERT Inference on CPU (Part 1)
How we sped up transformer inference 100x for 🤗 API customers