RedPill
  • INTRODUCTION
    • Introduction
  • GET STARTED
    • How To Use RedPill API
      • Credits
    • RedPill Auto Router
    • Check List Models
    • Supported Models
    • API Reference
      • Create Chat Completion
      • Create Embeddings
    • Links
  • REDPILL ROUTER NETWORK
    • How to Become an API Node
      • FAQs
    • RedPill API Credits Exchange
    • Earn Effortlessly as an API Node
    • Liquid
      • Launching RedPill Liquidity Pool Rewards
        • LP Rewards
        • LP Season1
      • LP FAQs
  • Confidential AI Inference
    • Introduction
    • Get Started
    • Host LLM in TEE
    • Implementation
    • LLM in TEE Benchmark
Powered by GitBook
On this page
  • Introduction
  • Overview
  • References
  1. Confidential AI Inference

Introduction

PreviousLP FAQsNextGet Started

Last updated 3 months ago

Introduction

This chapter provides detailed technical information on the Confidential AI Inference, designed to ensure confidentiality, integrity, and verifiability of AI inference tasks. We use the the TEE technologies provided by NVIDIA GPU TEE and Intel TDX to secure AI workloads, allowing developers to easily deploy their LLMs in a secure environment.

Overview

Confidential inference addresses critical concerns such as data privacy, secure execution, and computation verifiability, making it indispensable for sensitive applications. As illustrated in the diagram below, people currently cannot fully trust the responses returned by LLMs from services like OpenAI or Meta, due to the lack of cryptographic verification. By running the LLM inside a TEE, we can add verification primitives alongside the returned response, known as a Remote Attestation (RA) Report. This allows users to verify the AI generation results locally without relying on any third parties.

Check the Get Started for more information.

Check the Host LLM in TEE for how to host your own private LLM in TEE.

Check the Implementation for the technical details of the Confidential AI Inference.

References

HCC-Whitepaper
Intel SGX DCAP Orientation
Phala's dcap-qvl
Automata’s Solidity Implementation
Phala Nvidia H200 TEE Benchmark Paper
Phala DeRoT Post on FlashBots forum
Phala Key Management Protocol Post on Flashbots forum