Skip to main content

Filed under

Edge Computing

  1. Edge Computing·14 min read·

    Running LLaMA 3.1 on a Raspberry Pi: Memory-Efficient Edge AI with PIPELOAD

    Run LLaMA 3.1 on Raspberry Pi with PIPELOAD pipeline inference—achieve 90% memory reduction and deploy 7B models on 8GB edge devices at 2.5 tokens/sec.

  2. Ai·45 min read·

    AI Learning in Resource-Constrained Environments

    Train AI models on resource-constrained hardware with quantization, pruning, and distillation—run GPT-3 capabilities 100x faster through compression.

  3. Ai·15 min read·

    AI Meets Edge Computing: Transforming Real-Time Intelligence

    Deploy AI edge computing with YOLOv8 and TensorFlow Lite—achieve 15ms latency for real-time inference on Raspberry Pi with local processing for privacy.