SpaceTiD Calculating
Home
Data Center Transformation
The Road Ahead
Technical Support
About Us
Sign In
Logout
CN
Technical Services
EcoBox Computing Power Forecast
Enter the baseline forecast inputs
Project Name
Application
Dialog Assistant
Knowledge Q&A
AI Customer Service
Code Assistant
Other
Industry
Finance
Manufacturing
Healthcare
Government
Education
Retail
Internet
Energy
Other
Concurrent Users
users
Model Size
7B
14B
32B
70B
100B+
Uncertain
Request Length
Short text (about 100 Tokens)
Medium text (about 300 Tokens)
Long text (about 1000 Tokens)
Very long text (about 4000 Tokens)
Run Forecast
Project A
Finance / Dialog Assistant
This estimate uses a 14B model with medium-length requests.
0
cards
24GB / 80 TFLOPS
32GB / 120 TFLOPS
48GB / 160 TFLOPS
64GB / 220 TFLOPS
80GB / 300 TFLOPS
128GB / 450 TFLOPS
188GB / 600 TFLOPS
Recommended GPU Count
By compute: 0 cards; by memory: 0 cards
0
Peak Compute Demand
Estimated from token throughput and active model parameters.
0
Total VRAM Demand
Includes parameter residency, activation memory, and KV cache.
Technical Assumptions
Token Generation Rate
token/s
Model Precision
Select Model Precision
FP8
FP16
FP32
BF16
INT8
INT4
GPU Utilization
Redundancy Factor
GPU TFLOPS
Text Ratio
Image Ratio
Image FLOPs Multiplier
Trend Analysis
Concurrent User Growth Trend
Required GPUs and VRAM for Different Model Sizes
Request Length vs Compute Demand
Save This Forecast
Saved Forecasts