diff --git a/README.md b/README.md index 2062865dd..9d657f98b 100644 --- a/README.md +++ b/README.md @@ -26,7 +26,7 @@ In other words, we are interested in a network of networks and in keeping the da Post feed -Check out more screenshots [here](/wiki/en:Screenshots). +Check out more screenshots [here](https://github.com/Ocelot-Social-Community/Ocelot-Social/wiki/en:Screenshots). ## Features @@ -40,14 +40,14 @@ Ocelot.social networks feature: * filters * and more … -Check out the [full feature list](/wiki/en:FAQ#what-are-the-features). +Check out the [full feature list](https://github.com/Ocelot-Social-Community/Ocelot-Social/wiki/en:FAQ#what-are-the-features). ## User Guide and Frequently Asked Questions -In the [wiki](/wiki) you can find more information. +In the [wiki](https://github.com/Ocelot-Social-Community/Ocelot-Social/wiki) you can find more information. -* [User Guide](/wiki/en:User-Guide) -* [Frequently Asked Questions](/wiki/en:FAQ) +* [User Guide](https://github.com/Ocelot-Social-Community/Ocelot-Social/wiki/en:User-Guide) +* [Frequently Asked Questions](https://github.com/Ocelot-Social-Community/Ocelot-Social/wiki/en:FAQ) ## Demo diff --git a/cypress/cypress.config.js b/cypress/cypress.config.js index b1d80575e..db8308557 100644 --- a/cypress/cypress.config.js +++ b/cypress/cypress.config.js @@ -31,7 +31,8 @@ const { parsed } = dotenv.config({ path: '../backend/.env' }) module.exports = defineConfig({ e2e: { projectId: "qa7fe2", - defaultCommandTimeout: 10000, + defaultCommandTimeout: 60000, + pageLoadTimeout:180000, chromeWebSecurity: false, baseUrl: "http://localhost:3000", specPattern: "cypress/e2e/**/*.feature", diff --git a/deployment/TODO-next-update.md b/deployment/TODO-next-update.md index 4470efa14..591ce899f 100644 --- a/deployment/TODO-next-update.md +++ b/deployment/TODO-next-update.md @@ -4,7 +4,7 @@ When you introduce a new version and branding and deploy it on your network, you ## Version >= 3.3.0 with 'ocelotDockerVersionTag' 3.3.0-XXX -- We have the new option to configure DKIM for sent e-mails in environment variables (`.env` or `values.yaml`), see [here](deployment.md): +- We have the new option to configure DKIM for sent e-mails in environment variables (`.env` or `values.yaml`), see [deployment-values.md](deployment-values.md): - `SMTP_DKIM_DOMAINNAME=` - `SMTP_DKIM_KEYSELECTOR=` - `SMTP_DKIM_PRIVATKEY=` diff --git a/deployment/docker-compose.yml b/deployment/docker-compose.yml index 973dc6ac5..5236a7025 100644 --- a/deployment/docker-compose.yml +++ b/deployment/docker-compose.yml @@ -142,19 +142,35 @@ services: - 3001:80 neo4j: + # Neo4j v3.5.14-community + # image: wollehuss/neo4j-community-branded:latest + # Neo4j 4.4-community image: ocelotsocialnetwork/neo4j-community:latest container_name: neo4j-branded networks: - test-network + ports: + - 7687:7687 + # only for development + # - 7474:7474 + - 7474:7474 volumes: - neo4j_data:/data environment: + # settings reference: https://neo4j.com/docs/operations-manual/4.4/docker/ref-settings/ + # TODO: This sounds scary for a production environment - NEO4J_AUTH=none - NEO4J_dbms_security_procedures_unrestricted=algo.*,apoc.* - - NEO4J_ACCEPT_LICENSE_AGREEMENT=yes - ports: - - 7687:7687 - - 7474:7474 + - NEO4J_dbms_allow__format__migration=true + - NEO4J_dbms_allow__upgrade=true + # TODO: clarify if that is the only thing needed to unlock the Enterprise version + # - NEO4J_ACCEPT_LICENSE_AGREEMENT=yes + # Uncomment following line for Neo4j Enterprise version instead of Community version + # TODO: clarify if that is the only thing needed to unlock the Enterprise version + # - NEO4J_ACCEPT_LICENSE_AGREEMENT=yes + # TODO: Remove the playground from production + # bring the database in offline mode to export or load dumps + # command: ["tail", "-f", "/dev/null"] mailserver: image: djfarrelly/maildev diff --git a/deployment/scripts/cluster.backup.sh b/deployment/scripts/cluster.backup.sh index 103309419..9503061cf 100755 --- a/deployment/scripts/cluster.backup.sh +++ b/deployment/scripts/cluster.backup.sh @@ -18,16 +18,8 @@ BACKUP_FOLDER=${BACKUP_FOLDER:-${SCRIPT_DIR}/../configurations/${CONFIGURATION}/ # create backup fodler mkdir -p ${BACKUP_FOLDER} -# maintenance mode on -${SCRIPT_DIR}/cluster.maintenance.sh on - -# shutdown database -kubectl --kubeconfig=${KUBECONFIG} get deployment ocelot-neo4j -o json \ - | jq '.spec.template.spec.containers[] += {"command": ["tail", "-f", "/dev/null"]}' \ - | kubectl --kubeconfig=${KUBECONFIG} apply -f - - -# wait for the container to restart -sleep 60 +# cluster maintenance mode on && Neo4j maintenance mode on +${SCRIPT_DIR}/cluster.neo4j.sh maintenance on # database backup kubectl --kubeconfig=${KUBECONFIG} -n default exec -it \ @@ -40,13 +32,5 @@ kubectl --kubeconfig=${KUBECONFIG} cp \ kubectl --kubeconfig=${KUBECONFIG} cp \ default/$(kubectl --kubeconfig=${KUBECONFIG} -n default get pods | grep ocelot-backend |awk '{ print $1 }'):/app/public/uploads $BACKUP_FOLDER/public-uploads -# restart database -kubectl --kubeconfig=${KUBECONFIG} get deployment ocelot-neo4j -o json \ - | jq 'del(.spec.template.spec.containers[].command)' \ - | kubectl --kubeconfig=${KUBECONFIG} apply -f - - -# wait for the container to restart -sleep 60 - -# maintenance mode off -${SCRIPT_DIR}/cluster.maintenance.sh off \ No newline at end of file +# Neo4j maintenance mode off && cluster maintenance mode off +${SCRIPT_DIR}/cluster.neo4j.sh maintenance off \ No newline at end of file diff --git a/deployment/scripts/cluster.install.sh b/deployment/scripts/cluster.install.sh new file mode 100755 index 000000000..9dd7ed15a --- /dev/null +++ b/deployment/scripts/cluster.install.sh @@ -0,0 +1,58 @@ +#!/bin/bash + +# !!! never tested !!! + +# base setup +SCRIPT_PATH=$(realpath $0) +SCRIPT_DIR=$(dirname $SCRIPT_PATH) + +# check CONFIGURATION +if [ -z ${CONFIGURATION} ]; then + echo "You must provide a `CONFIGURATION` via environment variable" + exit 1 +fi +echo "Using CONFIGURATION=${CONFIGURATION}" + +# configuration +KUBECONFIG=${KUBECONFIG:-${SCRIPT_DIR}/../configurations/${CONFIGURATION}/kubeconfig.yaml} +VALUES=${SCRIPT_DIR}/../configurations/${CONFIGURATION}/kubernetes/values.yaml +DOCKERHUB_OCELOT_TAG=${DOCKERHUB_OCELOT_TAG:-"latest"} + + +## install Ingress-Nginx +helm repo add ingress-nginx https://kubernetes.github.io/ingress-nginx +helm repo update +helm install \ + ingress-nginx ingress-nginx/ingress-nginx \ + --kubeconfig=${KUBECONFIG} \ + -f ${SCRIPT_DIR}/../src/kubernetes/nginx.values.yaml + +## install Cert-Manager +helm repo add jetstack https://charts.jetstack.io +helm repo update +helm install \ + cert-manager jetstack/cert-manager \ + --kubeconfig=${KUBECONFIG} \ + --namespace cert-manager \ + --create-namespace \ + --version v1.13.1 \ + --set installCRDs=true + +## install Ocelot with helm +helm install \ + ocelot \ + --kubeconfig=${KUBECONFIG} \ + --values ${VALUES} \ + --set appVersion="${DOCKERHUB_OCELOT_TAG}" \ + ${SCRIPT_DIR}/../src/kubernetes/ \ + --timeout 10m + +## set Neo4j database indexes, constrains, and initial admin account plus run migrate up +kubectl --kubeconfig=${KUBECONFIG} \ + -n default \ + exec -it \ + $(kubectl --kubeconfig=${KUBECONFIG} -n default get pods | grep ocelot-backend | awk '{ print $1 }') -- \ + /bin/sh -c "yarn prod:migrate init && yarn prod:migrate up" + # /bin/sh -c "node --experimental-repl-await build/src/db/clean.js && node --experimental-repl-await build/src/db/seed.js" + +echo "!!! You must install a firewall or similar !!! (for DigitalOcean see: deployment/src/kubernetes/README.md)" diff --git a/deployment/scripts/cluster.neo4j.sh b/deployment/scripts/cluster.neo4j.sh new file mode 100755 index 000000000..41ebe4227 --- /dev/null +++ b/deployment/scripts/cluster.neo4j.sh @@ -0,0 +1,53 @@ +#!/bin/bash + +# base setup +SCRIPT_PATH=$(realpath $0) +SCRIPT_DIR=$(dirname $SCRIPT_PATH) + +# check CONFIGURATION +if [ -z ${CONFIGURATION} ]; then + echo "You must provide a `CONFIGURATION` via environment variable" + exit 1 +fi + +# configuration +KUBECONFIG=${KUBECONFIG:-${SCRIPT_DIR}/../configurations/${CONFIGURATION}/kubeconfig.yaml} + +case $1 in + maintenance) + case $2 in + on) + # maintenance mode on + ${SCRIPT_DIR}/cluster.maintenance.sh on + + # set Neo4j in offline mode (maintenance) + kubectl --kubeconfig=${KUBECONFIG} get deployment ocelot-neo4j -o json \ + | jq '.spec.template.spec.containers[] += {"command": ["tail", "-f", "/dev/null"]}' \ + | kubectl --kubeconfig=${KUBECONFIG} apply -f - + + # wait for the container to restart + sleep 60 + ;; + off) + # set Neo4j in online mode + kubectl --kubeconfig=${KUBECONFIG} get deployment ocelot-neo4j -o json \ + | jq 'del(.spec.template.spec.containers[].command)' \ + | kubectl --kubeconfig=${KUBECONFIG} apply -f - + + # wait for the container to restart + sleep 60 + + # maintenance mode off + ${SCRIPT_DIR}/cluster.maintenance.sh off + ;; + *) + echo -e "Run this script with first argument either 'off' or 'on'" + exit + ;; + esac + ;; + *) + echo -e "Run this script with first argument 'maintenance'" + exit + ;; +esac diff --git a/deployment/src/kubernetes/Backup.md b/deployment/src/kubernetes/Backup.md index 227b5765f..34011a512 100644 --- a/deployment/src/kubernetes/Backup.md +++ b/deployment/src/kubernetes/Backup.md @@ -137,7 +137,7 @@ $ kubectl -n default exec -it $(kubectl -n default get pods | grep ocelot-neo4j # bash: enter bash of Neo4j $ kubectl -n default exec -it $(kubectl -n default get pods | grep ocelot-neo4j | awk '{ print $1 }') -- bash # generate Dump -neo4j% neo4j-admin dump --to=/var/lib/neo4j/$(date +%F)-neo4j-dump +neo4j% neo4j-admin dump --database=graph.db --to=/var/lib/neo4j/$(date +%F)-neo4j-dump # exit bash neo4j% exit diff --git a/deployment/src/kubernetes/DigitalOcean.md b/deployment/src/kubernetes/DigitalOcean.md index 2c919d5f2..71c49d783 100644 --- a/deployment/src/kubernetes/DigitalOcean.md +++ b/deployment/src/kubernetes/DigitalOcean.md @@ -75,7 +75,7 @@ The IPs of the DigitalOcean machines are not necessarily stable, so the cluster' ## Deploy -Yeah, you're done here. Back to [Deployment with Helm for Kubernetes](/deployment/kubernetes/README.md). +Yeah, you're done here. Back to [Deployment with Helm for Kubernetes](/deployment/src/kubernetes/README.md). ## Backups On DigitalOcean diff --git a/deployment/src/kubernetes/README.md b/deployment/src/kubernetes/README.md index 126c37e9c..17773d736 100644 --- a/deployment/src/kubernetes/README.md +++ b/deployment/src/kubernetes/README.md @@ -9,7 +9,7 @@ Please contact us if you are interested in options not listed below. Managed Kubernetes: -- [DigitalOcean](/deployment/kubernetes/DigitalOcean.md) +- [DigitalOcean](/deployment/src/kubernetes/DigitalOcean.md) ## Configuration @@ -46,29 +46,20 @@ Please have a look here: - [Installing with Helm](https://cert-manager.io/docs/installation/helm/#installing-with-helm) -Our Helm installation is optimized for cert-manager version `v1.9.1` and `kubectl` version `"v1.24.2`. +Our Helm installation is optimized for cert-manager version `v1.13.1` and `kubectl` version `"v1.28.2`. Please search here for cert-manager versions that are compatible with your `kubectl` version on the cluster and on the client: [cert-manager Supported Releases](https://cert-manager.io/docs/installation/supported-releases/#supported-releases). ***ATTENTION:*** *When uninstalling cert-manager, be sure to use the same method as for installation! Otherwise, we could end up in a broken state, see [Uninstall](https://cert-manager.io/docs/installation/kubectl/#uninstalling).* - @@ -101,10 +100,15 @@ $ helm repo update #### 2. Install ingress-nginx ```bash +# in configuration/ + # kubeconfig.yaml set globaly -$ helm install ingress-nginx ingress-nginx/ingress-nginx -f nginx.values.yaml +helm install ingress-nginx ingress-nginx/ingress-nginx -f ../../src/kubernetes/nginx.values.yaml + # or kubeconfig.yaml in your repo, then adjust -$ helm --kubeconfig=/../kubeconfig.yaml install ingress-nginx ingress-nginx/ingress-nginx -f nginx.values.yaml +helm install \ + ingress-nginx ingress-nginx/ingress-nginx -f ../../src/kubernetes/nginx.values.yaml \ + --kubeconfig ./kubeconfig.yaml ``` ### DigitalOcean Firewall @@ -159,6 +163,8 @@ $ doctl compute firewall get --context ### DNS +***ATTENTION:** This seems not to work at all so we leave it away at the moment* + ***TODO:** I thought this is necessary if we use the DigitalOcean DNS management service? See [Manage DNS With DigitalOcean](/deployment/kubernetes/DigitalOcean.md#manage-dns-with-digitalocean)* This chart is only necessary (recommended is more precise) if you run DigitalOcean without load balancer. @@ -174,6 +180,8 @@ $ helm repo update #### 2. Install DNS ```bash +# !!! untested for now for new deployment structure !!! + # kubeconfig.yaml set globaly $ helm install dns bitnami/external-dns -f dns.values.yaml # or kubeconfig.yaml in your repo, then adjust @@ -191,10 +199,22 @@ All commands for ocelot need to be executed in the kubernetes folder. Therefore Only run once for the first time of installation: ```bash +# in configuration/ + # kubeconfig.yaml set globaly -$ helm install ocelot ./ +helm install ocelot \ + --values ./kubernetes/values.yaml \ + --set appVersion="latest" \ + ../../src/kubernetes/ \ + --timeout 10m + # or kubeconfig.yaml in your repo, then adjust -$ helm --kubeconfig=/../kubeconfig.yaml install ocelot ./ +helm install ocelot \ + --kubeconfig ./kubeconfig.yaml \ + --values ./kubernetes/values.yaml \ + --set appVersion="latest" \ + ../../src/kubernetes/ \ + --timeout 10m ``` #### Upgrade & Update @@ -202,10 +222,24 @@ $ helm --kubeconfig=/../kubeconfig.yaml install ocelot ./ Run for all upgrades and updates: ```bash +# !!! untested for now for new deployment structure !!! + +# in configuration/ + # kubeconfig.yaml set globaly -$ helm upgrade ocelot ./ +helm upgrade ocelot \ + --values ./kubernetes/values.yaml \ + --set appVersion="latest" \ + ../../src/kubernetes/ \ + --timeout 10m + # or kubeconfig.yaml in your repo, then adjust -$ helm --kubeconfig=/../kubeconfig.yaml upgrade ocelot ./ +helm upgrade ocelot \ + --kubeconfig ./kubeconfig.yaml \ + --values ./kubernetes/values.yaml \ + --set appVersion="latest" \ + ../../src/kubernetes/ \ + --timeout 10m ``` #### Rollback @@ -213,10 +247,17 @@ $ helm --kubeconfig=/../kubeconfig.yaml upgrade ocelot ./ Run for a rollback, in case something went wrong: ```bash +# !!! untested for now for new deployment structure !!! + +# in configuration/ + # kubeconfig.yaml set globaly -$ helm rollback ocelot +helm rollback ocelot --timeout 10m + # or kubeconfig.yaml in your repo, then adjust -$ helm --kubeconfig=/../kubeconfig.yaml rollback ocelot +helm rollback ocelot \ + --kubeconfig ./kubeconfig.yaml \ + --timeout 10m ``` #### Uninstall @@ -224,10 +265,17 @@ $ helm --kubeconfig=/../kubeconfig.yaml rollback ocelot Be aware that if you uninstall ocelot the formerly bound volumes become unbound. Those volumes contain all data from uploads and database. You have to manually free their reference in order to bind them again when reinstalling. Once unbound from their former container references they should automatically be rebound (considering the sizes did not change) ```bash +# !!! untested for now for new deployment structure !!! + +# in configuration/ + # kubeconfig.yaml set globaly -$ helm uninstall ocelot +helm uninstall ocelot --timeout 10m + # or kubeconfig.yaml in your repo, then adjust -$ helm --kubeconfig=/../kubeconfig.yaml uninstall ocelot +helm uninstall ocelot \ + --kubeconfig ./kubeconfig.yaml \ + --timeout 10m ``` ## Backups @@ -292,8 +340,11 @@ $ kubectl -n default rollout status deployment/ocelot-neo4j --timeout=240s # !!! be aware of the correct kube context !!! $ kubectl config get-contexts -# reset and seed Neo4j database via backend for staging +# for staging: reset and seed Neo4j database via backend $ kubectl -n default exec -it $(kubectl -n default get pods | grep ocelot-backend | awk '{ print $1 }') -- /bin/sh -c "node --experimental-repl-await build/src/db/clean.js && node --experimental-repl-await build/src/db/seed.js" +# or alternatively +# for production: set Neo4j database indexes, constrains, and initial admin account plus run migrate up via backend +$ kubectl -n default exec -it $(kubectl -n default get pods | grep ocelot-backend | awk '{ print $1 }') -- /bin/sh -c "yarn prod:migrate init && yarn prod:migrate up" ``` diff --git a/docker-compose.yml b/docker-compose.yml index 41e4dc261..78cee69b3 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -108,6 +108,7 @@ services: volumes: - neo4j_data:/data environment: + # settings reference: https://neo4j.com/docs/operations-manual/4.4/docker/ref-settings/ # TODO: This sounds scary for a production environment - NEO4J_AUTH=none - NEO4J_dbms_security_procedures_unrestricted=algo.*,apoc.* diff --git a/neo4j/README.md b/neo4j/README.md index 1e0a1d477..dcf5714ea 100644 --- a/neo4j/README.md +++ b/neo4j/README.md @@ -53,15 +53,45 @@ Start Neo4J and confirm the database is running at [http://localhost:7474](http: ## Operations on Neo4j -### Import Neo4j Dump Locally in Docker +### Docker or Docker Compose + +- we need to set `command: ["tail", "-f", "/dev/null"]` in the Neo4j block of `docker-compose.yml` on top level so the Neo4j database is in maintenance mode + +### Create Neo4j Dump + +To create a dump in Neo4j running in a Docker container: + +- set the database to maintenance mode, see above +- entering the following commands: + +```bash +# connect to the Docker containers Neo4j terminal +$ docker exec -it neo4j bash +# generate Dump +neo4j% neo4j-admin dump --database=graph.db --to=/var/lib/neo4j/$(date +%F)-neo4j-dump +# exit bash +neo4j% exit +# copy the dump out of the running Docker container +$ docker cp :/var/lib/neo4j/neo4j-dump /$(date +%F)-neo4j-dump +``` + +### Import Neo4j Dump To import a dump into Neo4j running in a Docker container: -- we need to set `command: ["tail", "-f", "/dev/null"]` in the Neo4j block of `docker-compose.yml` on top level so the Neo4j database is in maintenance mode -- copy the dump into the running Docker container: `docker cp /path/to/dump :/existing-directory-in-docker/` -- connect to the Docker containers Neo4j terminal: `docker exec -it neo4j bash` -- to load the dump into the database we need the following command in this terminal: `neo4j-admin load --expand-commands --database=graph.db --from /backups/neo4j-dump --force` -- leave the terminal by entering: `exit` +- set the database to maintenance mode, see above +- entering the following commands: + +```bash +# copy the dump into the running Docker container +$ docker cp /neo4j-dump :/var/lib/neo4j/$(date +%F)-neo4j-dump +# connect to the Docker containers Neo4j terminal +$ docker exec -it neo4j bash +# to load the dump into the database we need the following command in this terminal +neo4j% neo4j-admin load --expand-commands --database=graph.db --from /var/lib/neo4j/$(date +%F)-neo4j-dump --force +# leave the terminal by entering +neo4j% exit +``` ## Commands diff --git a/webapp/components/DonationInfo/DonationInfo.spec.js b/webapp/components/DonationInfo/DonationInfo.spec.js index 76aef9444..ac537be5b 100644 --- a/webapp/components/DonationInfo/DonationInfo.spec.js +++ b/webapp/components/DonationInfo/DonationInfo.spec.js @@ -46,7 +46,7 @@ describe('DonationInfo.vue', () => { // it looks to me that toLocaleString for some reason is not working as expected it.skip('creates a label from the given amounts and a translation string', () => { - expect(mocks.$t).nthCalledWith(1, 'donations.amount-of-total', { + expect(mocks.$t).toHaveBeenNthCalledWith(1, 'donations.amount-of-total', { amount: '10.000', total: '50.000', }) @@ -55,7 +55,7 @@ describe('DonationInfo.vue', () => { describe('given english locale', () => { it('creates a label from the given amounts and a translation string', () => { - expect(mocks.$t).toBeCalledWith( + expect(mocks.$t).toHaveBeenCalledWith( 'donations.amount-of-total', expect.objectContaining({ amount: '10,000', diff --git a/webapp/components/FilterMenu/FilterMenuComponent.vue b/webapp/components/FilterMenu/FilterMenuComponent.vue index e362e758f..7897a285d 100644 --- a/webapp/components/FilterMenu/FilterMenuComponent.vue +++ b/webapp/components/FilterMenu/FilterMenuComponent.vue @@ -3,7 +3,7 @@

{{ $t('filter-menu.filter-by') }}

-
+
{ - this.$emit('showFilterMenu') - this.$toast.success(this.$t('filter-menu.save.success')) - }) - .catch(() => { - this.$toast.error(this.$t('filter-menu.save.error')) - }) + if (this.categoriesActive) { + this.$apollo + .mutate({ + mutation: SaveCategories(), + variables: { activeCategories: this.filteredCategoryIds }, + }) + .then(() => { + this.$emit('showFilterMenu') + this.$toast.success(this.$t('filter-menu.save.success')) + }) + .catch(() => { + this.$toast.error(this.$t('filter-menu.save.error')) + }) + } }, }, } diff --git a/webapp/components/Group/GroupMember.spec.js b/webapp/components/Group/GroupMember.spec.js index ef8b96568..8efc754ad 100644 --- a/webapp/components/Group/GroupMember.spec.js +++ b/webapp/components/Group/GroupMember.spec.js @@ -94,20 +94,20 @@ describe('GroupMember', () => { describe('with server error', () => { it('toasts an error message', () => { - expect(toastErrorMock).toBeCalledWith('Oh no!') + expect(toastErrorMock).toHaveBeenCalledWith('Oh no!') }) }) describe('with server success', () => { it('calls the API', () => { - expect(apolloMock).toBeCalledWith({ + expect(apolloMock).toHaveBeenCalledWith({ mutation: changeGroupMemberRoleMutation(), variables: { groupId: 'group-id', userId: 'user', roleInGroup: 'admin' }, }) }) it('toasts a success message', () => { - expect(toastSuccessMock).toBeCalledWith('group.changeMemberRole') + expect(toastSuccessMock).toHaveBeenCalledWith('group.changeMemberRole') }) }) }) @@ -150,7 +150,7 @@ describe('GroupMember', () => { }) it('toasts an error message', () => { - expect(toastErrorMock).toBeCalledWith('Oh no!!') + expect(toastErrorMock).toHaveBeenCalledWith('Oh no!!') }) it('closes the modal', () => { @@ -165,7 +165,7 @@ describe('GroupMember', () => { }) it('calls the API', () => { - expect(apolloMock).toBeCalledWith({ + expect(apolloMock).toHaveBeenCalledWith({ mutation: removeUserFromGroupMutation(), variables: { groupId: 'group-id', userId: 'user' }, }) @@ -176,7 +176,7 @@ describe('GroupMember', () => { }) it('toasts a success message', () => { - expect(toastSuccessMock).toBeCalledWith('group.memberRemoved') + expect(toastSuccessMock).toHaveBeenCalledWith('group.memberRemoved') }) it('closes the modal', () => { diff --git a/webapp/components/LoginForm/LoginForm.spec.js b/webapp/components/LoginForm/LoginForm.spec.js index e0972b453..2004656a5 100644 --- a/webapp/components/LoginForm/LoginForm.spec.js +++ b/webapp/components/LoginForm/LoginForm.spec.js @@ -75,8 +75,8 @@ describe('LoginForm', () => { describe('no categories saved', () => { it('resets the categories', async () => { await fillIn(Wrapper()) - expect(storeMocks.mutations['posts/RESET_CATEGORIES']).toBeCalled() - expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).not.toBeCalled() + expect(storeMocks.mutations['posts/RESET_CATEGORIES']).toHaveBeenCalled() + expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).not.toHaveBeenCalled() }) }) @@ -84,11 +84,11 @@ describe('LoginForm', () => { it('sets the categories', async () => { authUserMock.mockReturnValue({ activeCategories: ['cat1', 'cat9', 'cat12'] }) await fillIn(Wrapper()) - expect(storeMocks.mutations['posts/RESET_CATEGORIES']).toBeCalled() - expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toBeCalledTimes(3) - expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toBeCalledWith({}, 'cat1') - expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toBeCalledWith({}, 'cat9') - expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toBeCalledWith({}, 'cat12') + expect(storeMocks.mutations['posts/RESET_CATEGORIES']).toHaveBeenCalled() + expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toHaveBeenCalledTimes(3) + expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toHaveBeenCalledWith({}, 'cat1') + expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toHaveBeenCalledWith({}, 'cat9') + expect(storeMocks.mutations['posts/TOGGLE_CATEGORY']).toHaveBeenCalledWith({}, 'cat12') }) }) }) diff --git a/webapp/components/PostTeaser/PostTeaser.spec.js b/webapp/components/PostTeaser/PostTeaser.spec.js index 0d48729d3..315f922cd 100644 --- a/webapp/components/PostTeaser/PostTeaser.spec.js +++ b/webapp/components/PostTeaser/PostTeaser.spec.js @@ -73,7 +73,7 @@ describe('PostTeaser', () => { it('has no validation errors', () => { const spy = jest.spyOn(global.console, 'error') Wrapper() - expect(spy).not.toBeCalled() + expect(spy).not.toHaveBeenCalled() spy.mockReset() }) diff --git a/webapp/pages/index.spec.js b/webapp/pages/index.spec.js index c16560a84..86a852a83 100644 --- a/webapp/pages/index.spec.js +++ b/webapp/pages/index.spec.js @@ -114,11 +114,11 @@ describe('PostIndex', () => { }) it('resets the category filter', () => { - expect(mutations['posts/RESET_CATEGORIES']).toBeCalled() + expect(mutations['posts/RESET_CATEGORIES']).toHaveBeenCalled() }) it('sets the category', () => { - expect(mutations['posts/TOGGLE_CATEGORY']).toBeCalledWith({}, 'cat3') + expect(mutations['posts/TOGGLE_CATEGORY']).toHaveBeenCalledWith({}, 'cat3') }) }) }) diff --git a/webapp/pages/index.vue b/webapp/pages/index.vue index 7b5c4a903..6067119e1 100644 --- a/webapp/pages/index.vue +++ b/webapp/pages/index.vue @@ -19,7 +19,7 @@
-
+
{ asyncData = true tosVersion = '0.0.4' wrapper = await Wrapper() - expect(redirect).toBeCalledWith('/') + expect(redirect).toHaveBeenCalledWith('/') }) }) }) diff --git a/webapp/pages/logout.spec.js b/webapp/pages/logout.spec.js index 460d321aa..c0edd1c4b 100644 --- a/webapp/pages/logout.spec.js +++ b/webapp/pages/logout.spec.js @@ -36,8 +36,8 @@ describe('logout.vue', () => { }) it('logs out and redirects to login', () => { - expect(mocks.$store.dispatch).toBeCalledWith('auth/logout') - expect(mocks.$router.replace).toBeCalledWith('/login') + expect(mocks.$store.dispatch).toHaveBeenCalledWith('auth/logout') + expect(mocks.$router.replace).toHaveBeenCalledWith('/login') }) }) }) diff --git a/webapp/pages/map.spec.js b/webapp/pages/map.spec.js index a9e21c876..a0ad95605 100644 --- a/webapp/pages/map.spec.js +++ b/webapp/pages/map.spec.js @@ -114,19 +114,19 @@ describe('map', () => { }) it('initializes on style load', () => { - expect(mapOnMock).toBeCalledWith('style.load', expect.any(Function)) + expect(mapOnMock).toHaveBeenCalledWith('style.load', expect.any(Function)) }) it('initializes on mouseenter', () => { - expect(mapOnMock).toBeCalledWith('mouseenter', 'markers', expect.any(Function)) + expect(mapOnMock).toHaveBeenCalledWith('mouseenter', 'markers', expect.any(Function)) }) it('initializes on mouseleave', () => { - expect(mapOnMock).toBeCalledWith('mouseleave', 'markers', expect.any(Function)) + expect(mapOnMock).toHaveBeenCalledWith('mouseleave', 'markers', expect.any(Function)) }) it('calls add map control', () => { - expect(mapAddControlMock).toBeCalled() + expect(mapAddControlMock).toHaveBeenCalled() }) describe('trigger style load event', () => { @@ -137,7 +137,7 @@ describe('map', () => { }) it('calls loadMarkersIconsAndAddMarkers', () => { - expect(spy).toBeCalled() + expect(spy).toHaveBeenCalled() }) }) diff --git a/webapp/pages/password-reset.spec.js b/webapp/pages/password-reset.spec.js index 9337d1d59..328f04606 100644 --- a/webapp/pages/password-reset.spec.js +++ b/webapp/pages/password-reset.spec.js @@ -68,7 +68,7 @@ describe('password-reset.vue', () => { asyncData = true isLoggedIn = true wrapper = await Wrapper() - expect(redirect).toBeCalledWith('/') + expect(redirect).toHaveBeenCalledWith('/') }) }) }) diff --git a/webapp/pages/post/edit/_id.spec.js b/webapp/pages/post/edit/_id.spec.js index 2773483a4..205a7c4fe 100644 --- a/webapp/pages/post/edit/_id.spec.js +++ b/webapp/pages/post/edit/_id.spec.js @@ -76,7 +76,10 @@ describe('post/_id.vue', () => { authorId = 'some-author' userId = 'some-user' wrapper = await Wrapper() - expect(error).toBeCalledWith({ message: 'error-pages.cannot-edit-post', statusCode: 403 }) + expect(error).toHaveBeenCalledWith({ + message: 'error-pages.cannot-edit-post', + statusCode: 403, + }) }) it('renders with asyncData of same user', async () => { diff --git a/webapp/pages/registration.spec.js b/webapp/pages/registration.spec.js index a276c2b8b..b34ba3ba7 100644 --- a/webapp/pages/registration.spec.js +++ b/webapp/pages/registration.spec.js @@ -327,7 +327,7 @@ describe('Registration', () => { asyncData = true isLoggedIn = true wrapper = await Wrapper() - expect(redirect).toBeCalledWith('/') + expect(redirect).toHaveBeenCalledWith('/') }) // copied from webapp/components/Registration/Signup.spec.js as testing template diff --git a/webapp/pages/terms-and-conditions-confirm.spec.js b/webapp/pages/terms-and-conditions-confirm.spec.js index c8faeb6f1..268eb83ed 100644 --- a/webapp/pages/terms-and-conditions-confirm.spec.js +++ b/webapp/pages/terms-and-conditions-confirm.spec.js @@ -71,7 +71,7 @@ describe('terms-and-conditions-confirm.vue', () => { asyncData = true tosAgree = true wrapper = await Wrapper() - expect(redirect).toBeCalledWith('/') + expect(redirect).toHaveBeenCalledWith('/') }) }) })