Skip to main content

The ultimate disaster recovery

I'm watching a presentation on Camlistore from 2016 and it's one I have seen before. I've also been watching "Fear of the dead" which asks me to think about a few different scenarios. But while Camlistore is meant to be storage for life or storage of life ... given my life experience and code-life experience how much of this matters to me any more?

In 30 years as a professional/journeyman programmer I have written millions of lines of code. Most of it belongs to my employers or customers but a good deal belongs to me. There have been countless experiments that I've used to learn, experience and experiment. But even as I think about all that code there is no way I would want to DR any of it in the short term let alone in the long term. On the other hand having copies of photos and videos would be handy and nice.

However....

In the case of a zombie attack and the end of days I do not think any of that is going to matter. Also, "we" know so much more now about programming and devices that if we had to reboot all of society I do not imagine that we would be looking at linux, windows or osx anymore.

I suppose it would have been nice to see my answer to Microsoft's DBCC interview question but at the time there was no way to capture that info; and the same could be said for the Amazon javascript text editor or card shuffling an dealing programs.

I guess what I'm saying is that beyond the normal lifespan of a project and part way into it's half-life just let it go. [a] you'll spend more time reverse engineering it instead of rewriting a modern version. [b] frankly by the time you're reached the half-life there has been enough attrition both personnel and cognitive that starting over is better for everyone.

Which brings me to my final point. The absolute core of the "work" needs to be implemented in a very descriptive and easy to reconstruct DSL. And with relative deference to Donald Knuth and Literate Programming the challenge is that literate programming's execution is difficult.

PS: one thing to take from Literate Programming is code around the documentation instead of documentation around the code. I think the modern version of this a variation on markdown.

Popular posts from this blog

Prometheus vs Bosun

In conclusion... while Bosun(B) is still not the ideal monitoring system neither is Prometheus(P).

TL;DR;

I am running Bosun in a Docker container hosted on CoreOS. Fleet service/unit files keep it running. However in once case I have experienced at least one severe crash as a result of a disk full condition. That it is implemented as part golang, java and python is an annoyance. The MIT license is about the only good thing.

I am trying to integrate Prometheus into my pipeline but losing steam fast. The Prometheus design seems to desire that you integrate your own cache inside your application and then allow the server to scrape the data, however, if the interval between scrapes is shorter than the longest transient session of your application then you need a gateway. A place to shuttle your data that will be a little more persistent.

(1) storing the data in my application might get me started more quickly
(2) getting the server to pull the data might be more secure
(3) using a push g…

Entry level cost for CoreOS+Tectonic

CoreOS and Tectonic start their pricing at 10 servers. Managed CoreOS starts at $1000 per month for those first 10 servers and Tectonic is $5000 for the same 10 servers. Annualized that is $85K or at least one employee depending on your market. As a single employee company I'd rather hire the employee. Specially since I only have 3 servers.

The pricing is biased toward the largest servers with the largest capacities; my dual core 32GB i5 IntelNuc can never be mistaken for a 96-CPU dual or quad core DELL

If CoreOS does not figure out a different barrier of entry they are going to follow the Borland path to obscurity.

Weave vs Flannel

While Weave and Flannel have some features in common weave includes DNS for service discovery and a wrapper process for capturing that info. In order to get some parity you'd need to add a DNS service like SkyDNS and then write your own script to weave the two together.
In Weave your fleet file might have some of this:
[Service] . . . ExecStartPre=/opt/bin/weave run --net=host --name bob ncx/bob ExecStart=/usr/bin/docker attach bob
In sky + flannel it might look like:
[Service] . . . ExecStartPre=docker run -d --net=host --name bob ncx/bob ExecStartPre=etcdctl set /skydns/local/ncx/bob '{"host":"`docker inspect --format '{{ .NetworkSettings.IPAddress }}' bob`","port":8080}' ExecStart=/usr/bin/docker attach bob
I'd like it to look like this:
[Service] . . . ExecStartPre=skyrun --net=host --name bob ncx/bob ExecStart=/usr/bin/docker attach bob
That's the intent anyway. I'm not sure the exact commands will work and that's partly why we…