Settings

Theme

Mind the Trust Gap: Fast, Private Local-to-Cloud LLM Chat

hazyresearch.stanford.edu

11 points by wolecki 9 months ago · 2 comments

Reader

woleckiOP 9 months ago

A fast Trusted Execution Environment protocol utilizing the H100 confidential mode. Prompts are decrypted and processed in the GPU enclave. Key innovation is that it can be really fast especially on ≥10B parameter models, with latency overhead less than 1%. Like in CPU confidential cloud computing, that opens up a channel for communication with cloud GenAI models that even the provider cannot intercept. Wonder whether something like that could boost the trust for all the AI neoclouds out there.

danbiderman 9 months ago

author here!

Keyboard Shortcuts

j
Next item
k
Previous item
o / Enter
Open selected item
?
Show this help
Esc
Close modal / clear selection