Skip to main content

Giving Privacy a Pass

As I've written recently written I am in the process of moving my backups to TimeMachine, SuperDuper and BackBlaze. One of the things that I was thinking about was the amount of time that it was going to take for the initial upload to BackBlaze. When I was looking at other solutions the initial upload was going to be on the order of months. BackBlaze, on the other hand, is said to auto-throttle.

I'm assuming that auto-throttle means that they watch the system load and then throttle the amount of bandwidth that it takes on the single system. Since I'm backing up 2 systems with individual accounts (one master account) to BackBlaze both machines seem to be humming alone nicely.

Back to my thought... in what order are the files backed up during the initial upload and thereafter. I'm certain that there have been studies on the average user's average file size and the number of edits over it's lifetime. And this information would be key.

During the initial upload I'd probably sort the files by size and not date. Doing all of the smallest files first. There are two reasons. (1) because it shows that the backup is making progress and the user is less likely to abandon the upload if they make quick progress. (2) in the event of a crash during the initial backup you might have a better chance of recovering more of the system in terms of individual files. The same cannot be said for the incremental backups; the largest files might get starved.

Anyway, BackBlaze does not appear to be taking months to achieve the online backup I was hoping for. Let's hope I never have to perform a restore... but there is something to be said for purchasing the occasional snapshot.

But now for privacy. Let's say for the sake of argument that I have elected not to encrypt my data that is being backed up. Now immediately before the backup, if BackBlaze generated some sort of signature of each target file and compared it to the entire dataset on it's servers it could reduce the backup time and duplicate storage costs by consolidating duplicates. This would work well for movies, videos, music but not individual unique titles. However, at this exact moment BackBlaze is backing up my iTunes library. I know I have about 8,000+ files and BackBlaze is reporting that it has 8,000 files to backup. I do not care if anyone knows what's in my music library. It's all paid for. Seems to me that there is no reason to upload them (same for iTunes match).

Comments

Popular posts from this blog

Entry level cost for CoreOS+Tectonic

CoreOS and Tectonic start their pricing at 10 servers. Managed CoreOS starts at $1000 per month for those first 10 servers and Tectonic is $5000 for the same 10 servers. Annualized that is $85K or at least one employee depending on your market. As a single employee company I'd rather hire the employee. Specially since I only have 3 servers.

The pricing is biased toward the largest servers with the largest capacities; my dual core 32GB i5 IntelNuc can never be mistaken for a 96-CPU dual or quad core DELL

If CoreOS does not figure out a different barrier of entry they are going to follow the Borland path to obscurity.

UPDATE 2017-10-30: With gratitude the CoreOS team has provided updated information on their pricing, however, I stand by my conclusion that the effective cost is lower when you deploy monster machines. The cost per node of my 1 CPU Intel NUC is the same as a 96 CPU server when you get beyond 10 nodes. I'll also reiterate that while my pricing notes are not currently…

eGalax touch on default Ubuntu 14.04.2 LTS

I have not had success with the touch drivers as yet.  The touch works and evtest also seems to report events, however, I have noticed that the button click is not working and no matter what I do xinput refuses to configure the buttons correctly.  When I downgraded to ubuntu 10.04 LTS everything sort of worked... there must have been something in the kermel as 10.04 was in the 2.6 kernel and 4.04 is in the 3.x branch.

One thing ... all of the documentation pointed to the wrong website or one in Taiwanese. I was finally able to locate the drivers again: http://www.eeti.com.tw/drivers_Linux.html (it would have been nice if they provided the install instructions in text rather than PDF)
Please open the document "EETI_eGTouch_Programming_Guide" under the Guide directory, and follow the Guidline to install driver.
download the appropriate versionunzip the fileread the programming manual And from that I'm distilling to the following: execute the setup.sh answer all of the questio…

Prometheus vs Bosun

In conclusion... while Bosun(B) is still not the ideal monitoring system neither is Prometheus(P).

TL;DR;

I am running Bosun in a Docker container hosted on CoreOS. Fleet service/unit files keep it running. However in once case I have experienced at least one severe crash as a result of a disk full condition. That it is implemented as part golang, java and python is an annoyance. The MIT license is about the only good thing.

I am trying to integrate Prometheus into my pipeline but losing steam fast. The Prometheus design seems to desire that you integrate your own cache inside your application and then allow the server to scrape the data, however, if the interval between scrapes is shorter than the longest transient session of your application then you need a gateway. A place to shuttle your data that will be a little more persistent.

(1) storing the data in my application might get me started more quickly
(2) getting the server to pull the data might be more secure
(3) using a push g…