A lot of their speed comes from special tricks like bypassing the kernel and going direct to hardware (DPDK).
Intel's DPDK doesn't support Mac OS X.
You should probably develop in a VM anyways, or at least with a platform + hardware that supports these features so you can better match what production deployments will be like.
I always cringe a little, for backend systems software, when people develop in an environment completely divorced from their production environment. It just creates headaches.
Problem with Snebu: it only does file-level deduplication I think which means storing something like virtual machine snapshots or images will not deduplicate well.
You need block-level deduplication to work well with things like VM images.
That is correct -- at this time it is only doing file level dedup, due to overhead and complexity of block level (or better yet, variable block level) deduplication. The file level dedup is done by computing the sha1 checksum of the file, and using that as the name to store the file contents under.
I've got a solution in the works specifically for KVM images, that hopefully I'll be able to finish up the next time I get a week off of work. The way I'm planning on handling that (at least for LibVirt based VMs) is to use libguestfs to create an equivalent to the `find` command (with all the -printf options that it supports), and to generate tar file output of selected files from the VM. (Snebu is designed around only requiring find/tar on the client side -- so anything that can produce this output will work). Although I don't like that libguestfs actually fires up a VM in order to work with the image files -- I may work on an approach to read the qcow2 file images directly. Again, looking forward to a vacation this spring so I can pound out that module.
My experience with attic (seems to be a trend with all of your dedup reviewed systems) is that it also takes a very long time to restore large amounts of data _however_ you have the option of restoring individual files via a FUSE file system which is _immensely_ useful.
An example was that a restore of ~200 GiB of VM snapshots took over a day from a NAS to the server in question.
Usually, the backups take about an hour to write out, so reading data from attic does take significantly longer.
This is probably because dedup is non-trivial to restore from (it can involve lots of random reads/disk seeks).
This tool certainly looks interesting and does support several sources for the backup.
I was describing a meta tool that would act like an experienced backup admin and setup the whole process / environment. E.g. this tool would configure network shares or a webserver with authentication to access the data, that deltaic restored.
However, such a tool does not exist, so I will study deltaic when there is some more time to waste with backup research. :)