Skip to main content

PART 1: Gitlab:latest and docker

I've been exercising with Digital Ocean and my need to get gitlab running in such a way that I can publish or register container instances. Sure I could use the public FREE version but I really want to integrate with my own CI/CD and so on.

So there is a lot going on... launch gitlab as a container:

docker run --detach \
    --hostname git.example.com \
    --publish 443:443 --publish 80:80 --publish 2222:22 \
    --name gitlab \
    --restart always \
    --volume /srv/gitlab/config:/etc/gitlab \
    --volume /srv/gitlab/logs:/var/log/gitlab \
    --volume /srv/gitlab/data:/var/opt/gitlab \
    gitlab/gitlab-ce:latest

This is going to launch the container for you, however, one important point is that there is no SSL here and it's quite complicated so to start this is the wrong way to start the container... instead this is preferred:

docker run --detach \
    --hostname git.example.com \
    --publish 10443:443 --publish 10080:80 --publish 10022:22 \
    --name gitlab \
    --restart always \
    --volume /srv/gitlab/config:/etc/gitlab \
    --volume /srv/gitlab/logs:/var/log/gitlab \
    --volume /srv/gitlab/data:/var/opt/gitlab \
    gitlab/gitlab-ce:latest

* one nasty side effect of storing all of the snippets in gitlab is trying to locate them when you need them to restart the service.

Mounting a host volume feels better than mounting a container. There are certain advantages including container HA and some functions that do not exist yet but there are serious problems when cleaning stale containers... you could delete all your data.  I recently survived a system crash only because the files were mounted on the host.

It's not great but it's better. What this means is that git is going to be accessible:

http://1.2.3.4:10080

This is not terrible be it also means you're using the host's IP address and not the ephemeral IP. Think in terms of Kubernetes the public IP is not stitched in.

At this point we have a running gitlab. The next step is setting up the smtp email relay. My preference is to deploy using mailgun so you'll need an account there. It's free until you get to scale.

Next configure gitlab to use mailgun by editing the file on the host volume:

vi /srv/gitlab/config/gitlab.rb

And you'll want to get these fields from the mailgun config(docs):

gitlab_rails['smtp_enable'] = true
gitlab_rails['smtp_address'] = "smtp.mailgun.org"
gitlab_rails['smtp_port'] = 587
gitlab_rails['smtp_authentication'] = "plain"
gitlab_rails['smtp_enable_starttls_auto'] = true
gitlab_rails['smtp_user_name'] = "postmaster@mg.example.com"
gitlab_rails['smtp_password'] = "password here"
gitlab_rails['smtp_domain'] = "mg.example.com"

Now you'll need to reload the config by first shelling into the running container

docker exec -it gitlab /bin/sh

And then tell gitlab to reconfigure

gitlab-ctl reconfigure

UPDATE: I forgot to add the gitlab runner

docker run -d --name gitlab-runner --restart always \
  -v /srv/gitlab-runner/config:/etc/gitlab-runner \
  -v /var/run/docker.sock:/var/run/docker.sock \
  gitlab/gitlab-runner:latest

The actual registration info is here. Essentially with a running runner... do this:

docker exec -it gitlab-runner gitlab-runner register

and answer the questions with info in gitlab

NEXT: In Part 2 I'll document the certbot configuration and Part 3 will be haproxy.

Comments

Popular posts from this blog

Entry level cost for CoreOS+Tectonic

CoreOS and Tectonic start their pricing at 10 servers. Managed CoreOS starts at $1000 per month for those first 10 servers and Tectonic is $5000 for the same 10 servers. Annualized that is $85K or at least one employee depending on your market. As a single employee company I'd rather hire the employee. Specially since I only have 3 servers.

The pricing is biased toward the largest servers with the largest capacities; my dual core 32GB i5 IntelNuc can never be mistaken for a 96-CPU dual or quad core DELL

If CoreOS does not figure out a different barrier of entry they are going to follow the Borland path to obscurity.

UPDATE 2017-10-30: With gratitude the CoreOS team has provided updated information on their pricing, however, I stand by my conclusion that the effective cost is lower when you deploy monster machines. The cost per node of my 1 CPU Intel NUC is the same as a 96 CPU server when you get beyond 10 nodes. I'll also reiterate that while my pricing notes are not currently…

eGalax touch on default Ubuntu 14.04.2 LTS

I have not had success with the touch drivers as yet.  The touch works and evtest also seems to report events, however, I have noticed that the button click is not working and no matter what I do xinput refuses to configure the buttons correctly.  When I downgraded to ubuntu 10.04 LTS everything sort of worked... there must have been something in the kermel as 10.04 was in the 2.6 kernel and 4.04 is in the 3.x branch.

One thing ... all of the documentation pointed to the wrong website or one in Taiwanese. I was finally able to locate the drivers again: http://www.eeti.com.tw/drivers_Linux.html (it would have been nice if they provided the install instructions in text rather than PDF)
Please open the document "EETI_eGTouch_Programming_Guide" under the Guide directory, and follow the Guidline to install driver.
download the appropriate versionunzip the fileread the programming manual And from that I'm distilling to the following: execute the setup.sh answer all of the questio…

Prometheus vs Bosun

In conclusion... while Bosun(B) is still not the ideal monitoring system neither is Prometheus(P).

TL;DR;

I am running Bosun in a Docker container hosted on CoreOS. Fleet service/unit files keep it running. However in once case I have experienced at least one severe crash as a result of a disk full condition. That it is implemented as part golang, java and python is an annoyance. The MIT license is about the only good thing.

I am trying to integrate Prometheus into my pipeline but losing steam fast. The Prometheus design seems to desire that you integrate your own cache inside your application and then allow the server to scrape the data, however, if the interval between scrapes is shorter than the longest transient session of your application then you need a gateway. A place to shuttle your data that will be a little more persistent.

(1) storing the data in my application might get me started more quickly
(2) getting the server to pull the data might be more secure
(3) using a push g…