mirror of
https://github.com/IT4Change/Ocelot-Social.git
synced 2025-12-13 07:45:56 +00:00
This keeps our configuration DRY and helps us to save keystrokes.
Human-Connection Nitro | Deployment Configuration
Currently the deployment is not primetime ready as you still have to do some manual work. That we need to change, the following list gives some glimpse of the missing steps.
Todo`s
- check labels and selectors if they all are correct
- configure NGINX from yaml
- configure Let's Encrypt cert-manager from yaml
- configure ingress form yaml
- configure persistent & shared storage between nodes
- reproduce setup locally
The dummy directory has some lb configurations that did not work properly on Digital Ocean but could be used as a starting point for getting it right
Install Minikube, kubectl
There are many Kubernetes distributions, but if you're just getting started, Minikube is a tool that you can use to get your feet wet.
Create a namespace locally
kubectl create -f namespace-staging.yml
Change config maps according to your needs
cd config/
cp db-migration-worker-configmap.template.yaml db-migration-worker-configmap.yaml
# edit all variables according to the setup of the remote legacy server
cd ..
Apply the config map to staging namespace
kubectl apply -f config/
Setup secrets and deploy themn
cp secrets.yaml.template secrets.yaml
# change all vars as needed and deploy it afterwards
kubectl apply -f secrets.yaml
Create volumes and deployments
kubectl apply -f volumes/
kubectl apply -f deployments/
This can take a while. Sit back and relax and have a look into your minikube dashboard:
minikube dashboard
Wait until all pods turn green and they don't show a warning Waiting: ContainerCreating anymore.
Expose the services
kubectl apply -f services/
Access the service
minikube service nitro-backend --namespace=staging
minikube service nitro-web --namespace=staging
Provisioning db-migration-worker
Copy your private ssh key and the .known-hosts file of your remote legacy server.
# check the corresponding db-migration-worker pod
kubectl --namespace=staging get pods
# change <POD_ID> below
kubectl cp path/to/your/ssh/keys/.ssh staging/nitro-db-migration-worker-<POD_ID>:/root/
Run the migration:
# change <POD_IDs> below
kubectl --namespace=staging exec -it nitro-db-migration-worker-<POD_ID> ./import.sh
kubectl --namespace=staging exec -it nitro-neo4j-<POD_ID> ./import/import.sh
Description
Languages
TypeScript
35.7%
Vue
24.9%
JavaScript
19.2%
HTML
17.8%
SCSS
0.9%
Other
1.4%