Skip to content

Commit 2f9433e

Browse files
committed
Update
1 parent 80eeb44 commit 2f9433e

File tree

4 files changed

+22
-1
lines changed

4 files changed

+22
-1
lines changed
170 KB
Loading
Loading

confidential-ai/confidential-AI-API.md

Lines changed: 16 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,12 +24,14 @@ https://inference-api.phala.network/
2424
{
2525
"signing_address": "...",
2626
"nvidia_payload": "...",
27-
"dcap_payload": "..."
27+
"intel_quote": "..."
2828
}
2929
```
3030

3131
### Verify the Attestation
3232

33+
- Verify GPU Attestation
34+
3335
You can copy the value of `nvidia_payload` as the whole payload as followed to verify:
3436

3537
```
@@ -39,6 +41,19 @@ curl -X POST https://nras.attestation.nvidia.com/v3/attest/gpu \
3941
-d '__COPY_FROM_ABOVE__'
4042
```
4143

44+
- Verify TDX Quote
45+
46+
Theoretically, you can verify the Intel TDX quote with the value of `intel_quote` at anywhere that provide TDX quote verification service. The screenshot below is an example of how to verify the Intel TDX quote with the [Automata's on-chain attestation smart contract](https://explorer.ata.network/address/0xE26E11B257856B0bEBc4C759aaBDdea72B64351F/contract/65536_2/readContract#F6). For Automata example, just need to convert the returned base64 encoded quote to hex format (take Node for example).
47+
48+
```sh
49+
console.log('Quote bytes:', '0x' + Buffer.from(intel_quote, 'base64').toString('hex'));
50+
51+
// Use on-chain smart contract function `verifyAndAttestOnChain` https://explorer.ata.network/address/0xE26E11B257856B0bEBc4C759aaBDdea72B64351F/contract/65536_2/readContract#F6
52+
// to verify with the printed quote bytes above.
53+
```
54+
55+
<figure><img src="../.gitbook/assets/automata-attestation.png" alt=""><figcaption></figcaption></figure>
56+
4257
## Chat API
4358

4459
OpenAI-compatible API. See: https://platform.openai.com/docs/api-reference/chat

confidential-ai/host-LLM-in-TEE.md

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,11 @@
11
# Host LLM in TEE
22

3+
Phala Network leverages NVIDIA's TEE GPU technology to provide confidential LLM inference services. By utilizing NVIDIA Blackwell's Confidential Computing capabilities, we ensure that AI model execution and data processing remain fully protected within secure enclaves. This solution enables organizations to run their LLM workloads with guaranteed privacy and security, preventing unauthorized access to both the model and user data during inference operations.
4+
5+
<figure><img src="../.gitbook/assets/hopper-arch-confidential-computing.jpeg" alt=""><figcaption></figcaption></figure>
6+
7+
*source: https://www.nvidia.com/en-us/data-center/solutions/confidential-computing/*
8+
39
The following diagram illustrates the architecture of the Phala Confidential AI Inference (private LLM Node) service. To host your own private LLM in TEE, you only need to wrap your LLM inference code into a docker image, then deploy your own container to the TEE network. To make the LLM fully verifiable, you need make sure the docker image is reproducible.
410

511
<figure><img src="../.gitbook/assets/host-llm-in-TEE.png" alt=""><figcaption></figcaption></figure>

0 commit comments

Comments
 (0)