User:Caleb Cooper/sandbox: Difference between revisions
Appearance
Content deleted Content added
Caleb Cooper (talk | contribs) No edit summary |
Caleb Cooper (talk | contribs) No edit summary |
||
| Line 10: | Line 10: | ||
<pre class="inline"> |
:<pre class="inline"> |
||
this |
this |
||
that |
that |
||
Revision as of 10:29, 15 July 2020
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
this that the other
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
this that the other
this that the other