Methodology for calculating the normalized, amortized emissions from fine-tuning AI models
Component | Disclosed data |
---|---|
Base model | Llama 2 |
GPU | Nvidia A100 80GB |
Server | HPE Apollo 6500 Gen10 Plus |
Number of GPUs | 4 |
Number of servers | 1 |
Server location | AWS US West (Oregon) |
Total reserved time | 12 hours |
Average CPU utilization | 12% |
Average GPU utilization | 47% |
Missing data point | Mechanism to replace |
---|---|
GPU model | Use the most common GPU for the training year (for instance, 2022 is Nvidia A100) |
Server model | Use the most common server or instance type for the training year |
Cluster size | Assume 1 server for fine-tuning |
Location | Use the US as a relatively high-carbon country |
Datacenter PUE | Use location average |
Datacenter WUE | Use location average |
Total fine-tuning time | Predict from number of tokens and model |
Start time | Use the published model date minus the total reserved time |
GPU and CPU utilization | Predict from model |
Component | Disclosed data |
---|---|
Base model | Llama 2 |
Managed service | AWS Bedrock |
Region | US West (Oregon) |
Start time | July 6, 2024 17:01 |
Tokens | 48,123 |