arxiv:2312.12624
Debasish Dhal
DebasishDhal99
·
AI & ML interests
None yet
Recent Activity
reacted
to
ImranzamanML's
post
with 👍
6 days ago
Here is how we can calculate the size of any LLM model:
Each parameter in LLM models is typically stored as a floating-point number. The size of each parameter in bytes depends on the precision.
32-bit precision: Each parameter takes 4 bytes.
16-bit precision: Each parameter takes 2 bytes
To calculate the total memory usage of the model:
Memory usage (in bytes) = No. of Parameters × Size of Each Parameter
For example:
32-bit Precision (FP32)
In 32-bit floating-point precision, each parameter takes 4 bytes.
Memory usage in bytes = 1 billion parameters × 4 bytes
1,000,000,000 × 4 = 4,000,000,000 bytes
In gigabytes: ≈ 3.73 GB
16-bit Precision (FP16)
In 16-bit floating-point precision, each parameter takes 2 bytes.
Memory usage in bytes = 1 billion parameters × 2 bytes
1,000,000,000 × 2 = 2,000,000,000 bytes
In gigabytes: ≈ 1.86 GB
It depends on whether you use 32-bit or 16-bit precision, a model with 1 billion parameters would use approximately 3.73 GB or 1.86 GB of memory, respectively.
upvoted
a
paper
8 days ago
A Survey on Knowledge Distillation of Large Language Models
new activity
11 days ago
DebasishDhal99/cloud-vs-temperature-data:[bot] Conversion to Parquet
Organizations
Papers
1
spaces
3
models
None public yet
datasets
8
DebasishDhal99/cloud-vs-temperature-data
Viewer
•
Updated
•
984k
•
31
DebasishDhal99/punjabi-instruction-dataset
Viewer
•
Updated
•
245k
•
68
DebasishDhal99/hindi-eval-dataset
Viewer
•
Updated
•
3k
•
38
DebasishDhal99/qwen_eval_dataset
Viewer
•
Updated
•
3k
•
6
DebasishDhal99/german-polish-paired-placenames
Viewer
•
Updated
•
9.79k
•
45
DebasishDhal99/German_Names_Central_And_Eastern_Europe
Viewer
•
Updated
•
25k
•
84
DebasishDhal99/german-czech-paired-placenames
Viewer
•
Updated
•
5.72k
•
50
DebasishDhal99/exonyms-for-lithuanian-places
Viewer
•
Updated
•
102
•
7