Spaces:
Running
Running
Commit
Β·
469bf5d
1
Parent(s):
ae24574
docs: Update documentation screen with accurate pricing
Browse files- Correct HuggingFace Jobs billing model: per-second (not per-hour)
- Remove incorrect "minimum 1 hour charge" statement
- Update HF Jobs GPU pricing to match official HF Spaces pricing
- T4-small: $0.40/hr, A10G-large: $1.50/hr, A100-large: $2.50/hr
- Update Modal GPU pricing with verified rates
- T4: $0.000164/sec, A100-80GB: $0.000694/sec, H200: $0.001261/sec
- Add B200 GPU option for Modal ($0.001736/sec)
- Add per-second rate column to both pricing tables
- Update platform comparison to reflect both use per-second billing
- Align documentation screen with JOB_SUBMISSION.md
References:
- https://huggingface.co/docs/hub/en/spaces-gpus
- https://modal.com/pricing
- screens/documentation.py +29 -26
screens/documentation.py
CHANGED
|
@@ -1886,8 +1886,8 @@ TraceMind-AI provides seamless integration with two cloud compute platforms, all
|
|
| 1886 |
|
| 1887 |
| Platform | Best For | Pricing Model | GPU Options | Free Tier |
|
| 1888 |
|----------|----------|---------------|-------------|-----------|
|
| 1889 |
-
| **HuggingFace Jobs** | Managed infrastructure, dataset integration | Per-
|
| 1890 |
-
| **Modal** | Serverless compute, pay-per-second | Per-second | T4, L4, A10, A100-80GB, H200 | β
Free credits available |
|
| 1891 |
|
| 1892 |
### Key Differences
|
| 1893 |
|
|
@@ -1895,13 +1895,14 @@ TraceMind-AI provides seamless integration with two cloud compute platforms, all
|
|
| 1895 |
- β
Native HuggingFace ecosystem integration
|
| 1896 |
- β
Managed infrastructure with guaranteed availability
|
| 1897 |
- β
Built-in dataset storage and versioning
|
|
|
|
| 1898 |
- β οΈ Requires Pro account ($9/month)
|
| 1899 |
-
- β οΈ
|
| 1900 |
|
| 1901 |
**Modal**:
|
| 1902 |
- β
Serverless architecture (no minimum charges)
|
| 1903 |
-
- β
Pay-per-second billing
|
| 1904 |
-
- β
Latest GPUs (H200 available)
|
| 1905 |
- β
Free tier with credits
|
| 1906 |
- β οΈ Requires separate account setup
|
| 1907 |
- β οΈ Container cold start time (~2-3 minutes first run)
|
|
@@ -1933,18 +1934,19 @@ Steps to create token:
|
|
| 1933 |
|
| 1934 |
### Hardware Options
|
| 1935 |
|
| 1936 |
-
| Hardware | vCPUs | GPU | Memory | Best For | Price/hr |
|
| 1937 |
-
|
| 1938 |
-
| `cpu-basic` | 2 | - | 16 GB | API models (OpenAI, Anthropic) |
|
| 1939 |
-
| `cpu-upgrade` | 8 | - | 32 GB | API models (high volume) |
|
| 1940 |
-
| `t4-small` | 4 | T4 (16GB) |
|
| 1941 |
-
| `t4-medium` | 8 | T4 (16GB) |
|
| 1942 |
-
| `
|
| 1943 |
-
| `a10g-
|
| 1944 |
-
| `
|
| 1945 |
-
| `
|
|
|
|
| 1946 |
|
| 1947 |
-
|
| 1948 |
|
| 1949 |
### Auto-Selection Logic
|
| 1950 |
|
|
@@ -2021,19 +2023,20 @@ Steps to get credentials:
|
|
| 2021 |
|
| 2022 |
| Hardware | GPU | Memory | Best For | Price/sec | Equivalent $/hr |
|
| 2023 |
|----------|-----|--------|----------|-----------|-----------------|
|
| 2024 |
-
| `CPU` | - | 16 GB | API models |
|
| 2025 |
-
| `T4` | T4 (16GB) | 16 GB | Small models (4B-8B) |
|
| 2026 |
-
| `L4` | L4 (24GB) | 24 GB | Small-medium models |
|
| 2027 |
-
| `A10G` | A10G (24GB) | 32 GB | Medium models (7B-13B) |
|
| 2028 |
-
| `L40S` | L40S (48GB) | 48 GB | Large models (optimized) |
|
| 2029 |
-
| `A100` | A100 (40GB) | 64 GB | Large models |
|
| 2030 |
-
| `A100-80GB` | A100 (80GB) | 128 GB | Very large models (70B+) |
|
| 2031 |
-
| `H100` | H100 (80GB) | 192 GB | Latest generation inference |
|
| 2032 |
-
| `H200` | H200 (141GB) | 256 GB | Cutting-edge, highest memory |
|
|
|
|
| 2033 |
|
| 2034 |
Full pricing: https://modal.com/pricing
|
| 2035 |
|
| 2036 |
-
**π‘
|
| 2037 |
|
| 2038 |
### Auto-Selection Logic
|
| 2039 |
|
|
|
|
| 1886 |
|
| 1887 |
| Platform | Best For | Pricing Model | GPU Options | Free Tier |
|
| 1888 |
|----------|----------|---------------|-------------|-----------|
|
| 1889 |
+
| **HuggingFace Jobs** | Managed infrastructure, dataset integration | Per-second | T4, L4, A10, A100, V5e | β ($9/mo Pro required) |
|
| 1890 |
+
| **Modal** | Serverless compute, pay-per-second | Per-second | T4, L4, A10, A100-80GB, H100, H200, B200 | β
Free credits available |
|
| 1891 |
|
| 1892 |
### Key Differences
|
| 1893 |
|
|
|
|
| 1895 |
- β
Native HuggingFace ecosystem integration
|
| 1896 |
- β
Managed infrastructure with guaranteed availability
|
| 1897 |
- β
Built-in dataset storage and versioning
|
| 1898 |
+
- β
Per-second billing (no minimums)
|
| 1899 |
- β οΈ Requires Pro account ($9/month)
|
| 1900 |
+
- β οΈ Limited GPU options (no H100/H200)
|
| 1901 |
|
| 1902 |
**Modal**:
|
| 1903 |
- β
Serverless architecture (no minimum charges)
|
| 1904 |
+
- β
Pay-per-second billing
|
| 1905 |
+
- β
Latest GPUs (H100, H200, B200 available)
|
| 1906 |
- β
Free tier with credits
|
| 1907 |
- β οΈ Requires separate account setup
|
| 1908 |
- β οΈ Container cold start time (~2-3 minutes first run)
|
|
|
|
| 1934 |
|
| 1935 |
### Hardware Options
|
| 1936 |
|
| 1937 |
+
| Hardware | vCPUs | GPU | Memory | Best For | Price/hr | Per-second |
|
| 1938 |
+
|----------|-------|-----|--------|----------|----------|------------|
|
| 1939 |
+
| `cpu-basic` | 2 | - | 16 GB | API models (OpenAI, Anthropic) | Free with Pro | - |
|
| 1940 |
+
| `cpu-upgrade` | 8 | - | 32 GB | API models (high volume) | Free with Pro | - |
|
| 1941 |
+
| `t4-small` | 4 | T4 (16GB) | 15 GB | Small models (4B-8B) | $0.40 | $0.000111 |
|
| 1942 |
+
| `t4-medium` | 8 | T4 (16GB) | 30 GB | Small models (batched) | $0.60 | $0.000167 |
|
| 1943 |
+
| `l4x1` | 8 | L4 (24GB) | 30 GB | Small-medium models | $0.80 | $0.000222 |
|
| 1944 |
+
| `a10g-small` | 4 | A10G (24GB) | 32 GB | Medium models (7B-13B) | $1.00 | $0.000278 |
|
| 1945 |
+
| `a10g-large` | 12 | A10G (24GB) | 92 GB | Medium models (high memory) | $1.50 | $0.000417 |
|
| 1946 |
+
| `a100-large` | 12 | A100 (80GB) | 142 GB | Large models (70B+) | $2.50 | $0.000694 |
|
| 1947 |
+
| `v5e-1x1` | 4 | TPU v5e | 16 GB | TPU-optimized workloads | TBD | TBD |
|
| 1948 |
|
| 1949 |
+
*Pricing from [HF Spaces GPU docs](https://huggingface.co/docs/hub/en/spaces-gpus). Billed **per-second** with no minimums.*
|
| 1950 |
|
| 1951 |
### Auto-Selection Logic
|
| 1952 |
|
|
|
|
| 2023 |
|
| 2024 |
| Hardware | GPU | Memory | Best For | Price/sec | Equivalent $/hr |
|
| 2025 |
|----------|-----|--------|----------|-----------|-----------------|
|
| 2026 |
+
| `CPU` | - | 16 GB | API models | $0.0000131/core | ~$0.05 |
|
| 2027 |
+
| `T4` | T4 (16GB) | 16 GB | Small models (4B-8B) | $0.000164 | ~$0.59 |
|
| 2028 |
+
| `L4` | L4 (24GB) | 24 GB | Small-medium models | $0.000222 | ~$0.80 |
|
| 2029 |
+
| `A10G` | A10G (24GB) | 32 GB | Medium models (7B-13B) | $0.000306 | ~$1.10 |
|
| 2030 |
+
| `L40S` | L40S (48GB) | 48 GB | Large models (optimized) | $0.000542 | ~$1.95 |
|
| 2031 |
+
| `A100` | A100 (40GB) | 64 GB | Large models | $0.000583 | ~$2.10 |
|
| 2032 |
+
| `A100-80GB` | A100 (80GB) | 128 GB | Very large models (70B+) | $0.000694 | ~$2.50 |
|
| 2033 |
+
| `H100` | H100 (80GB) | 192 GB | Latest generation inference | $0.001097 | ~$3.95 |
|
| 2034 |
+
| `H200` | H200 (141GB) | 256 GB | Cutting-edge, highest memory | $0.001261 | ~$4.54 |
|
| 2035 |
+
| `B200` | B200 (192GB) | 384 GB | Next-gen, massive memory | $0.001736 | ~$6.25 |
|
| 2036 |
|
| 2037 |
Full pricing: https://modal.com/pricing
|
| 2038 |
|
| 2039 |
+
**π‘ Both platforms use per-second billing!** Choose Modal for H100/H200/B200 GPUs or if you don't have HF Pro.
|
| 2040 |
|
| 2041 |
### Auto-Selection Logic
|
| 2042 |
|