Skip to main content

Travel Routers Part 1

The first two of the travel routers I purchased have arrived and while I hate meaningless unboxing videos I'll spare you the details other than to say that the RavPower All-In-One FileHub is minimally packaged both inside and out; just the router and a USB cable leaving me to provide the wallcharger. The ZyXEL Wireless Router included the device, USB cable, universal power br. Both had very little documentation although the RavPower was easier to comprehend and if you want to get beyond QuickStart use-case you will have to download the ZyXEL manual.



RavPower
Battery - 6000mAH (quick press the power button to see power level) While the documentation calls it an "external battery" it seems to also power the router.
Ports - SDCard slot, mini USB for charging the battery, USB port for charging other devices and flash drives, RJ45 ethernet port.
Modes - Access Point, bridge, file server
GUI - web admin, smartphone app
Default SSID - FileHubPlus-XXXX
Default password - 11111111
Default IP Address - 10.10.10.254
Default uid/pwd - admin/(no password)
Manual: link

ZyXEL
Battery - none
Ports - USB power, RJ45 ethernet port
Special - WPS button, Clone MAC button, mode switch
Default SSID - ZyXELXXXX
Default password - 00000000
Default IP Address - (router mode)192.168.1.1 (other modes 192.168.1.2)
Default uid/pwd - admin/1234
Note - There is a Windows driver for autoplay.
Killer feature - From the quick scan of the user manual I discovered that there is a QOS feature. This could be helpful with my Apple Photo Sync problem crashing my modem and routers.

Sandisk Media Drive
- worth a mention but that's it. It was meant as a file server more that a router.

Sandisk Connect
- worth a mention but that's it. It was meant as a file server more that a router.

In part 2 I'll put the ZyXEL and RavPower to the test.

Popular posts from this blog

Prometheus vs Bosun

In conclusion... while Bosun(B) is still not the ideal monitoring system neither is Prometheus(P).

TL;DR;

I am running Bosun in a Docker container hosted on CoreOS. Fleet service/unit files keep it running. However in once case I have experienced at least one severe crash as a result of a disk full condition. That it is implemented as part golang, java and python is an annoyance. The MIT license is about the only good thing.

I am trying to integrate Prometheus into my pipeline but losing steam fast. The Prometheus design seems to desire that you integrate your own cache inside your application and then allow the server to scrape the data, however, if the interval between scrapes is shorter than the longest transient session of your application then you need a gateway. A place to shuttle your data that will be a little more persistent.

(1) storing the data in my application might get me started more quickly
(2) getting the server to pull the data might be more secure
(3) using a push g…

Entry level cost for CoreOS+Tectonic

CoreOS and Tectonic start their pricing at 10 servers. Managed CoreOS starts at $1000 per month for those first 10 servers and Tectonic is $5000 for the same 10 servers. Annualized that is $85K or at least one employee depending on your market. As a single employee company I'd rather hire the employee. Specially since I only have 3 servers.

The pricing is biased toward the largest servers with the largest capacities; my dual core 32GB i5 IntelNuc can never be mistaken for a 96-CPU dual or quad core DELL

If CoreOS does not figure out a different barrier of entry they are going to follow the Borland path to obscurity.

Weave vs Flannel

While Weave and Flannel have some features in common weave includes DNS for service discovery and a wrapper process for capturing that info. In order to get some parity you'd need to add a DNS service like SkyDNS and then write your own script to weave the two together.
In Weave your fleet file might have some of this:
[Service] . . . ExecStartPre=/opt/bin/weave run --net=host --name bob ncx/bob ExecStart=/usr/bin/docker attach bob
In sky + flannel it might look like:
[Service] . . . ExecStartPre=docker run -d --net=host --name bob ncx/bob ExecStartPre=etcdctl set /skydns/local/ncx/bob '{"host":"`docker inspect --format '{{ .NetworkSettings.IPAddress }}' bob`","port":8080}' ExecStart=/usr/bin/docker attach bob
I'd like it to look like this:
[Service] . . . ExecStartPre=skyrun --net=host --name bob ncx/bob ExecStart=/usr/bin/docker attach bob
That's the intent anyway. I'm not sure the exact commands will work and that's partly why we…