Data compression is the reduction of the number of bits that have to be saved or transmitted and the process is rather important in the internet hosting field because info filed on HDDs is typically compressed in order to take less space. You'll find various algorithms for compressing info and they provide different efficiency based upon the content. Many of them remove just the redundant bits, so no data will be lost, while others delete unnecessary bits, which leads to worse quality once your data is uncompressed. The process consumes a lot of processing time, so a web hosting server should be powerful enough to be able to compress and uncompress data in real time. One example how binary code could be compressed is by "remembering" that there're five sequential 1s, for example, as an alternative to storing all five 1s.
Data Compression in Shared Web Hosting
The cloud web hosting platform where your shared web hosting
account will be made employs the impressive ZFS file system. The LZ4 compression method which the aforementioned uses is superior in many aspects, and not only does it compress data better than any compression method which a variety of other file systems use, but it is also much quicker. The gains will be significant particularly on compressible content such as website files. While it could sound irrational, uncompressing data with LZ4 is faster than reading uncompressed info from a hard disk, so the performance of each website hosted on our servers shall be improved. The better and quicker compression rates also allow us to produce a large number of daily backups of the entire content in every single hosting account, so if you delete anything by mistake, the last back-up copy that we have won't be more than a few hours old. This is possible as the backups take much less space and their generation is fast enough, to not change the performance of the servers.
Data Compression in Semi-dedicated Hosting
The semi-dedicated hosting
plans that we offer are created on a powerful cloud hosting platform which runs on the ZFS file system. ZFS employs a compression algorithm known as LZ4 that is superior to any other algorithm you can find in terms of speed and compression ratio when it comes to processing web content. This is valid particularly when data is uncompressed since LZ4 does that a lot faster than it would be to read uncompressed data from a hard disk and because of this, Internet sites running on a platform where LZ4 is enabled will function at a higher speed. We are able to take advantage of this feature despite of the fact that it needs quite a considerable amount of CPU processing time as our platform uses a lot of powerful servers working together and we do not make accounts on a single machine like many companies do. There's one more reward of using LZ4 - considering the fact that it compresses data really well and does that extremely fast, we can also make multiple daily backups of all accounts without affecting the performance of the servers and keep them for a whole month. By doing this, you can always restore any content that you erase by mistake.