User:Caleb Cooper/sandbox: Difference between revisions
Content deleted Content added
Caleb Cooper (talk | contribs) No edit summary |
Caleb Cooper (talk | contribs) No edit summary |
||
Line 1:
<pre>
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
</pre>
<pre class="scroll">
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
</pre>
<pre class="inline">
this
that
the other
</pre>
<pre class="blue-white">
We have that issue on Unity because Data Reduction (Compression/Deduplication) works in 8 KB blocks. when we have volume with large numbers of writes smaller than 8 KB have to backfilled from the volume (a read to the backend), to get enough data to compress, which causes extra load on the CPU. The average write size masks the number of small writes to a certain extent, because any large writes can
| |||