diff --git a/docker/defaults/main.yml b/docker/defaults/main.yml index fb9aa1b..3eed9ec 100644 --- a/docker/defaults/main.yml +++ b/docker/defaults/main.yml @@ -13,3 +13,7 @@ docker_packages: docker_run_as_docker_user: True docker_user_home: /home/docker +docker_defaults_file: /etc/default/docker +docker_enable_tcp_socket: False +docker_tcp_socket_port: 2375 +docker_tcp_socket_host: 127.0.0.1 diff --git a/docker/tasks/pkgs.yml b/docker/tasks/pkgs.yml index a37461c..e56d3d5 100644 --- a/docker/tasks/pkgs.yml +++ b/docker/tasks/pkgs.yml @@ -14,6 +14,10 @@ apt: pkg={{ item }} state={{ docker_pkg_status }} update_cache=yes cache_valid_time=3600 with_items: '{{ docker_packages }}' + - name: Install the Docker default options + template: src=docker-default.j2 dest={{ docker_defaults_file }} owner=root group=root mode=0644 + notify: Restart docker + - name: Ensure that the service is started and enabled service: name=docker state=started enabled=yes diff --git a/docker/templates/docker-default.j2 b/docker/templates/docker-default.j2 new file mode 100644 index 0000000..dec0522 --- /dev/null +++ b/docker/templates/docker-default.j2 @@ -0,0 +1,15 @@ +# Customize location of Docker binary (especially for development testing). +#DOCKERD="/usr/local/bin/dockerd" + + +CUSTOM_DOCKER_SOCKET="-H tcp://{{ docker_tcp_socket_host }}:{{ docker_tcp_socket_port }} -H unix:///var/run/docker.sock" +#CUSTOM_DOCKER_DNS_SERVERS="--dns 8.8.8.8 --dns 8.8.4.4" + +# Use DOCKER_OPTS to modify the daemon startup options. +DOCKER_OPTS="{% if docker_enable_tcp_socket %}$CUSTOM_DOCKER_SOCKET {% endif %}" + +# If you need Docker to use an HTTP proxy, it can also be specified here. +#export http_proxy="http://127.0.0.1:3128/" + +# This is also a handy place to tweak where Docker's temporary files go. +#export DOCKER_TMPDIR="/mnt/bigdrive/docker-tmp" diff --git a/jenkins/common/defaults/main.yml b/jenkins/common/defaults/main.yml new file mode 100644 index 0000000..dc125c3 --- /dev/null +++ b/jenkins/common/defaults/main.yml @@ -0,0 +1,12 @@ +--- +# There are some duplicates here +jenkins_dest: "/var/lib/jenkins" +jenkins_username: jenkins +jenkins_group: jenkins +jenkins_shell: /bin/bash + +jenkins_maven_config: True +jenkins_maven_settings_dirs: + - .m2 + +jenkins_maven_settings_url: http://localhost/settings.xml diff --git a/jenkins/common/tasks/main.yml b/jenkins/common/tasks/main.yml new file mode 100644 index 0000000..b57d2d2 --- /dev/null +++ b/jenkins/common/tasks/main.yml @@ -0,0 +1,20 @@ +--- +- block: + - name: Create the maven setting directory + file: dest={{ jenkins_dest }}/{{ item }} state=directory + with_items: '{{ jenkins_maven_settings_dirs }}' + + - name: Fetch the maven settings template file + become: False + become_user: root + run_once: True + get_url: url={{ jenkins_maven_settings_url }} dest=/tmp/settings.xml.j2 force=yes + delegate_to: localhost + + - name: Install the maven settings + template: src=/tmp/settings.xml.j2 dest={{ jenkins_dest }}/.m2/settings.xml + + become: True + become_user: '{{ jenkins_username }}' + when: jenkins_maven_config + tags: [ 'jenkins', 'jenkins_common', 'jenkins_master', 'jenkins_slave' ] diff --git a/jenkins/master/defaults/main.yml b/jenkins/master/defaults/main.yml index 9d8a240..f3562bc 100644 --- a/jenkins/master/defaults/main.yml +++ b/jenkins/master/defaults/main.yml @@ -74,3 +74,5 @@ jenkins_plugins: enabled: True build-pipeline-plugin: enabled: True + build-timeout-plugin: + enabled: True diff --git a/jenkins/master/tasks/main.yml b/jenkins/master/tasks/main.yml index 289fdc2..63365db 100644 --- a/jenkins/master/tasks/main.yml +++ b/jenkins/master/tasks/main.yml @@ -32,14 +32,13 @@ service: name=jenkins state=started enabled=yes when: jenkins_install - tags: jenkins + tags: [ 'jenkins', 'jenkins_master' ] - block: # Handle plugins # If Jenkins is installed or updated, wait for pulling the Jenkins CLI, assuming 10s should be sufficiant - name: 120 seconds delay while starting Jenkins wait_for: port={{ jenkins_http_port }} delay={{ jenkins_restart_delay }} - when: jenkins_install | changed # Create Jenkins CLI destination directory - name: "Create Jenkins CLI destination directory" @@ -55,14 +54,14 @@ - name: Install plugins without a specific version jenkins_plugin: name="{{ item.key }}" params='{{ jenkins_access_params }}' register: my_jenkins_plugin_unversioned - when: 'version' not in item.value - with_dict: jenkins_plugins + when: '"version" not in item.value' + with_dict: '{{ jenkins_plugins }}' - name: Install plugins with a specific version jenkins_plugin: name="{{ item.key }}" version="{{ item.value['version'] }}" params='{{ jenkins_access_params }}' register: my_jenkins_plugin_versioned - when: 'version' in item.value - with_dict: jenkins_plugins + when: '"version" in item.value' + with_dict: '{{ jenkins_plugins }}' - name: Initiate the jenkins_restart_required fact set_fact: @@ -72,16 +71,17 @@ set_fact: jenkins_restart_required: yes when: item.changed - with_items: my_jenkins_plugin_versioned.results + with_items: '{{ my_jenkins_plugin_versioned.results }}' - name: Check if restart is required by any of the unversioned plugins set_fact: jenkins_restart_required: yes when: item.changed - with_items: my_jenkins_plugin_unversioned.results + with_items: '{{ my_jenkins_plugin_unversioned.results }}' - name: Restart Jenkins if required - service: name=jenkins state= restarted + become_user: root + service: name=jenkins state=restarted when: jenkins_restart_required - name: Wait for Jenkins to start up @@ -105,18 +105,18 @@ - name: Plugin pinning jenkins_plugin: name="{{ item.key }}" state="{{ 'pinned' if item.value['pinned'] else 'unpinned'}}" params='{{ jenkins_access_params }}' - when: 'pinned' in item.value - with_dict: jenkins_plugins + when: '"pinned" in item.value' + with_dict: '{{ jenkins_plugins }}' - name: Plugin enabling jenkins_plugin: name="{{ item.key }}" state="{{ 'enabled' if item.value['enabled'] else 'disabled'}}" params='{{ jenkins_access_params }}' - when: 'enabled' in item.value - with_dict: jenkins_plugins + when: '"enabled" in item.value' + with_dict: '{{ jenkins_plugins }}' become: True become_user: '{{ jenkins_username }}' when: jenkins_install - tags: [ 'jenkins', 'jenkins_plugins' ] + tags: [ 'jenkins', 'jenkins_plugins', 'jenkins_master' ] - block: - name: Ensure that jenkins is stoppend and disabled @@ -140,4 +140,4 @@ apt_key: url='{{ jenkins_repo_key }}' state=absent when: not jenkins_install - tags: jenkins + tags: [ 'jenkins', 'jenkins_master' ] diff --git a/jenkins/slave/defaults/main.yml b/jenkins/slave/defaults/main.yml index e69de29..4d5bc9a 100644 --- a/jenkins/slave/defaults/main.yml +++ b/jenkins/slave/defaults/main.yml @@ -0,0 +1,10 @@ +--- +jenkins_slave: False +# There are some duplicates here +jenkins_dest: "/var/lib/jenkins" +jenkins_username: jenkins +jenkins_group: jenkins +jenkins_shell: /bin/bash +jenkins_tmp_retain_days: 5 +# TODO: fetch the public key from the master +#jenkins_master_pubkey: diff --git a/jenkins/slave/tasks/main.yml b/jenkins/slave/tasks/main.yml index e69de29..916f5b6 100644 --- a/jenkins/slave/tasks/main.yml +++ b/jenkins/slave/tasks/main.yml @@ -0,0 +1,17 @@ +--- +- block: + - name: add the user that will run the jenkins jobs + user: name={{ jenkins_username }} home={{ jenkins_dest }} shell={{ jenkins_shell }} generate_ssh_key=yes + + - name: ensure the jenkins master has ssh access on each slave, as jenkins user + authorized_key: user={{ jenkins_username }} key="{{ item }}" state=present + with_items: + - '{{ jenkins_master_pubkey }}' + + - name: Daily cron job to cleanup the /tmp junk + template: src={{ item }}.j2 dest=/etc/cron.daily/{{ item }} owner=root group=root mode=0755 + with_items: + - tmp-cleaner + + when: jenkins_slave + tags: [ 'jenkins', 'jenkins_slave' ] diff --git a/jenkins/slave/templates/tmp-cleaner.j2 b/jenkins/slave/templates/tmp-cleaner.j2 new file mode 100644 index 0000000..54e8c23 --- /dev/null +++ b/jenkins/slave/templates/tmp-cleaner.j2 @@ -0,0 +1,8 @@ +#!/bin/bash + +RETAIN_DAYS={{ jenkins_tmp_retain_days }} +LOG_FILE=/var/log/tmp-cleaner.log +find /tmp/ -ctime +${RETAIN_DAYS} -exec rm -fr {} \; >>$LOG_FILE 2>&1 + +exit 0 + diff --git a/shinyproxy/defaults/main.yml b/shinyproxy/defaults/main.yml new file mode 100644 index 0000000..7fd569e --- /dev/null +++ b/shinyproxy/defaults/main.yml @@ -0,0 +1,20 @@ +--- +shinyproxy_install: False +shinyproxy_version: 0.8.6 +shinyproxy_file_name: 'shinyproxy-{{ shinyproxy_version }}.jar' +shinyproxy_url: 'https://www.shinyproxy.io/downloads/{{ shinyproxy_file_name }}' +shinyproxy_app_name: shinyproxy.jar +shinyproxy_user: shinyproxy +shinyproxy_install_dir: /opt/shinyproxy +shinyproxy_http_port: 8080 + +shinyproxy_app_title: 'Open Analytics Shiny Proxy' +shinyproxy_logo_url: 'http://www.openanalytics.eu/sites/www.openanalytics.eu/themes/oa/logo.png' +shinyproxy_authentication: ldap +shinyproxy_admin_group: '' +shinyproxy_ldap_server: 'ldap://ldap.forumsys.com:389/dc=example,dc=com' +shinyproxy_ldap_admin: cn=read-only-admin,dc=example,dc=com +shinyproxy_ldap_admin_pwd: password +shinyproxy_user_dn_pattern: 'uid={0}' +shinyproxy_group_search_base: '' +shinyproxy_group_search_filter: '(uniqueMember={0})' diff --git a/shinyproxy/handlers/main.yml b/shinyproxy/handlers/main.yml new file mode 100644 index 0000000..5e3b523 --- /dev/null +++ b/shinyproxy/handlers/main.yml @@ -0,0 +1,3 @@ +--- +- name: Restart shinyproxy + service: name=shinyproxy state=restarted diff --git a/shinyproxy/tasks/main.yml b/shinyproxy/tasks/main.yml new file mode 100644 index 0000000..b66a1f1 --- /dev/null +++ b/shinyproxy/tasks/main.yml @@ -0,0 +1,30 @@ +--- +- block: + - name: Create the shinyproxy user + user: name={{ shinyproxy_user }} home={{ shinyproxy_install_dir }} createhome=yes system=yes shell=/usr/sbin/nologin + + - name: Download the shinyproxy jar + become: True + become_user: '{{ shinyproxy_user }}' + get_url: url={{ shinyproxy_url }} dest={{ shinyproxy_install_dir }} + + - name: Set up a symlink to an unversioned app name + become: True + become_user: '{{ shinyproxy_user }}' + file: src={{ shinyproxy_install_dir }}/{{ shinyproxy_file_name }} dest={{ shinyproxy_install_dir }}/{{ shinyproxy_app_name }} state=link + + - name: Install the upstart init file + template: src=upstart-shinyproxy.conf.j2 dest=/etc/init/shinyproxy.conf owner=root group=root mode=0644 + when: is_trusty + + - name: Install the shinyproxy configuration file + template: src=shinyproxy-conf.yml.j2 dest={{ shinyproxy_install_dir }}/application.yml owner=root group={{ shinyproxy_user }} mode=0640 + notify: Restart shinyproxy + tags: [ 'shinyproxy', 'shinyproxy_conf' ] + + - name: Ensure that the shinyproxy service is enabled and running + service: name=shinyproxy state=started enabled=yes + + when: shinyproxy_install + tags: shinyproxy + diff --git a/shinyproxy/templates/shinyproxy-conf.yml.j2 b/shinyproxy/templates/shinyproxy-conf.yml.j2 new file mode 100644 index 0000000..5bc848c --- /dev/null +++ b/shinyproxy/templates/shinyproxy-conf.yml.j2 @@ -0,0 +1,40 @@ +shiny: + proxy: + title: {{ shinyproxy_app_title }} + logo-url: {{ shinyproxy_logo_url }} + landing-page: / + heartbeat-rate: 10000 + heartbeat-timeout: 60000 + port: 8080 + authentication: {{ shinyproxy_authentication }} + admin-groups: {{ shinyproxy_admin_group }} + # LDAP configuration + ldap: + url: {{ shinyproxy_ldap_server }} + user-dn-pattern: {{ shinyproxy_user_dn_pattern }} + group-search-base: {{ shinyproxy_group_search_base }} + group-search-filter: {{ shinyproxy_group_search_filter }} + manager-dn: {{ shinyproxy_ldap_admin }} + manager-password: {{ shinyproxy_ldap_admin_pwd }} +# Docker configuration + docker: + cert-path: /home/none + url: http://localhost:2375 + host: 127.0.0.1 + port-range-start: 20000 + apps: + - name: 01_hello + display-name: Hello Application + description: Application which demonstrates the basics of a Shiny app + docker-cmd: ["R", "-e shinyproxy::run_01_hello()"] + docker-image: openanalytics/shinyproxy-demo + groups: scientists, mathematicians + - name: 06_tabsets + docker-cmd: ["R", "-e shinyproxy::run_06_tabsets()"] + docker-image: openanalytics/shinyproxy-demo + groups: scientists + +logging: + file: + shinyproxy.log + diff --git a/shinyproxy/templates/upstart-shinyproxy.conf.j2 b/shinyproxy/templates/upstart-shinyproxy.conf.j2 new file mode 100644 index 0000000..93c4b4b --- /dev/null +++ b/shinyproxy/templates/upstart-shinyproxy.conf.j2 @@ -0,0 +1,17 @@ +# shinyproxy - springboot based shiny executor + +description "shinyproxy service" + +start on runlevel [2345] +stop on runlevel [!2345] +respawn +respawn limit 10 5 + +setuid {{ shinyproxy_user }} +setgid {{ shinyproxy_user }} + +script + cd {{ shinyproxy_install_dir }} + exec java -jar ./{{ shinyproxy_app_name }} +end script + diff --git a/smartgears/accounting-service/defaults/main.yml b/smartgears/accounting-service/defaults/main.yml new file mode 100644 index 0000000..c75b366 --- /dev/null +++ b/smartgears/accounting-service/defaults/main.yml @@ -0,0 +1,10 @@ +--- + +accounting_service_gcube_repository: 'gcube-snapshots' +accounting_service_install: False +accounting_service_ver: 1.0.0-20170323.102652-1 +accounting_service_snap: 1.0.0-SNAPSHOT +accounting_service_name: accounting-service +accounting_service_filename: '{{ accounting_service_name }}-{{ accounting_service_ver }}.war' +accounting_service_url: 'http://maven.research-infrastructures.eu/nexus/service/local/repositories/{{ accounting_service_gcube_repository }}/content/org/gcube/data/publishing/{{ accounting_service_name }}/{{ accounting_service_snap}}/{{ accounting_service_filename }}' +accounting_service_war_file: '{{ accounting_service_name }}.war' diff --git a/smartgears/accounting-service/tasks/main.yml b/smartgears/accounting-service/tasks/main.yml new file mode 100644 index 0000000..e76436b --- /dev/null +++ b/smartgears/accounting-service/tasks/main.yml @@ -0,0 +1,17 @@ +--- +- block: + + - name: Remove the installed accounting-service before upgrading + file: dest={{ item }} state=absent + with_items: + - '{{ smartgears_instance_path }}/webapps/accounting-service' + - '{{ smartgears_instance_path }}/webapps/accounting-service.war' + when: smartgears_upgrade + + - name: Get the accounting-service war file + get_url: url={{ accounting_service_url }} dest={{ smartgears_instance_path }}/webapps/{{ accounting_service_war_file }} + + become: True + become_user: '{{ smartgears_user }}' + when: accounting_service_install + tags: [ 'smartgears', 'accounting_service', 'tomcat' ] diff --git a/smartgears/dataminer_app/tasks/install-gcube-key.yml b/smartgears/dataminer_app/tasks/install-gcube-key.yml index d55e529..b83a35b 100644 --- a/smartgears/dataminer_app/tasks/install-gcube-key.yml +++ b/smartgears/dataminer_app/tasks/install-gcube-key.yml @@ -7,6 +7,7 @@ - '{{ gcube_prod_key_2 }}' - '{{ gcube_prod_key_3 }}' - '{{ gcube_prod_key_4 }}' + - '{{ gcube_prod_key_5 }}' notify: Restart smartgears when: install_gcube_prod_key @@ -17,6 +18,7 @@ - '{{ gcube_prod_key_2 }}' - '{{ gcube_prod_key_3 }}' - '{{ gcube_prod_key_4 }}' + - '{{ gcube_prod_key_5 }}' notify: Restart smartgears when: not install_gcube_prod_key diff --git a/smartgears/home_library/defaults/main.yml b/smartgears/home_library/defaults/main.yml new file mode 100644 index 0000000..5e25055 --- /dev/null +++ b/smartgears/home_library/defaults/main.yml @@ -0,0 +1,8 @@ +--- +home_library_gcube_repository: 'gcube-staging' +home_library_install: False +home_library_ver: 1.7.0-4.3.0-144852 +home_library_name: home-library-webapp +home_library_filename: '{{ home_library_name }}-{{ home_library_ver }}.war' +home_library_url: 'http://maven.research-infrastructures.eu/nexus/content/repositories/{{ home_library_gcube_repository }}/org/gcube/data/access/{{ home_library_name }}/{{ home_library_ver}}/{{ home_library_filename }}' +home_library_war_file: '{{ home_library_name }}.war' diff --git a/smartgears/home_library/tasks/main.yml b/smartgears/home_library/tasks/main.yml new file mode 100644 index 0000000..53ecb8b --- /dev/null +++ b/smartgears/home_library/tasks/main.yml @@ -0,0 +1,22 @@ +--- +- block: + + - name: Remove the installed HOME LIBRARY connector before upgrading + file: dest={{ item }} state=absent + with_items: + - '{{ smartgears_instance_path }}/webapps/home-library-webapp' + - '{{ smartgears_instance_path }}/webapps/home-library-webapp.war' + when: smartgears_upgrade + + - name: Get the HOME LIBRARY connector war file + get_url: url={{ home_library_url }} dest={{ smartgears_instance_path }}/webapps/{{ home_library_war_file }} + +# - name: Unpack the HOME LIBRARY connector war file +# shell: mkdir {{ smartgears_instance_path }}/webapps/home-library-webapp ; cd {{ smartgears_instance_path }}/webapps/home-library-webapp ; jar xf {{ smartgears_instance_path }}/webapps/{{ home_library_war_file }} +# args: +# creates: '{{ smartgears_instance_path }}/webapps/home-library-webapp/WEB-INF/web.xml' + + become: True + become_user: '{{ smartgears_user }}' + when: home_library_install + tags: [ 'smartgears', 'home_library', 'tomcat' ] diff --git a/smartgears/r_connector/defaults/main.yml b/smartgears/r_connector/defaults/main.yml index af47ee3..2582836 100644 --- a/smartgears/r_connector/defaults/main.yml +++ b/smartgears/r_connector/defaults/main.yml @@ -1,7 +1,7 @@ --- gcube_repository: 'gcube-staging' r_connector_install: False -r_connector_ver: 2.1.2-4.3.0-144071 +r_connector_ver: 2.1.3-4.4.0-146364 r_connector_name: r-connector r_connector_group_id: org.gcube.data.analysis r_connector_extension: war diff --git a/smartgears/r_connector/templates/rusersadd.j2 b/smartgears/r_connector/templates/rusersadd.j2 index 009b716..1853a4b 100755 --- a/smartgears/r_connector/templates/rusersadd.j2 +++ b/smartgears/r_connector/templates/rusersadd.j2 @@ -10,11 +10,18 @@ ALLOW_LOCAL_USERS=1 RPROFILE_FILE='{{ r_connector_rprofile_path }}/{{ r_connector_rprofile_filename }}' if [ -d $HDIR ] ; then - logger "$LOG_PREFIX user $HDIR directory exists" - sudo /usr/bin/touch ${HDIR}/{{ r_connector_userconfig }} - sudo /bin/chown ${USER}:{{ smartgears_user }} $HDIR/{{ r_connector_userconfig }} - sudo /usr/bin/setfacl -m u:${USER}:rw,g:{{ smartgears_user }}:rw $HDIR/{{ r_connector_userconfig }} - exit 0 + if id -u $USER >/dev/null 2>&1 + then + logger "$LOG_PREFIX user $HDIR directory exists. Touching the userconfig.csv file to ensure that it exists with the correct permissions" + sudo /usr/bin/touch ${HDIR}/{{ r_connector_userconfig }} + sudo /bin/chown ${USER}:{{ smartgears_user }} $HDIR/{{ r_connector_userconfig }} + sudo /usr/bin/setfacl -m u:${USER}:rw,g:{{ smartgears_user }}:rw $HDIR/{{ r_connector_userconfig }} + sudo /bin/ln -s $RPROFILE_FILE $HDIR/{{ r_connector_rprofile_filename }} + exit 0 + else + logger "$LOG_PREFIX user $HDIR directory exists but the user not. Aborting." + exit 1 + fi else if id -u $USER >/dev/null 2>&1 then diff --git a/smartgears/smartgears/defaults/main.yml b/smartgears/smartgears/defaults/main.yml index a50d484..2e1e2b0 100644 --- a/smartgears/smartgears/defaults/main.yml +++ b/smartgears/smartgears/defaults/main.yml @@ -16,19 +16,32 @@ smartgears_user: '{{ d4science_user }}' smartgears_user_home: '{{ d4science_user_home }}' smartgears_instance_path: '{{ smartgears_user_home }}/tomcat' smartgears_install_path: '{{ smartgears_user_home }}/SmartGears' -smartgears_distribution_version: 2.1.0-4.3.0-142337 +#smartgears_distribution_version: 2.1.0-4.3.0-142337 +smartgears_distribution_version: 2.1.2-4.4.0-146408 smartgears_file: 'smartgears-distribution-{{ smartgears_distribution_version }}.tar.gz' smartgears_url: 'http://maven.research-infrastructures.eu/nexus/content/repositories/{{ gcube_repository }}/org/gcube/distribution/smartgears-distribution/{{ smartgears_distribution_version }}/{{ smartgears_file }}' smartgears_mode: online smartgears_application_mode: online # Production infra smartgears_infrastructure_name: "d4science.research-infrastructures.eu" +# Production VOs +smartgears_production_vo: + - '/{{ smartgears_infrastructure_name }}' + - '/{{ smartgears_infrastructure_name }}/FARM' + - '/{{ smartgears_infrastructure_name }}/SoBigData' + - '/{{ smartgears_infrastructure_name }}/SmartArea' + - '/{{ smartgears_infrastructure_name }}/gCubeApps' + - '/{{ smartgears_infrastructure_name }}/D4Research' +# Set to 'true' or 'false'. Pay attention to the case +smartgears_authorized_on_all_scopes: 'false' +smartgears_scopes: + - '/{{ smartgears_infrastructure_name }}' smartgears_hostname: '{{ ansible_fqdn }}' smartgears_country: it smartgears_location: pisa smartgears_latitude: 41.9000 smartgears_longitude: 12.5000 -smartgears_publication_frequency: 180 +smartgears_publication_frequency: 300 smartgears_http_port: 9000 smartgears_service_name: 'tomcat-instance-{{ smartgears_http_port }}' @@ -36,8 +49,6 @@ smartgears_loglevel: WARN smartgears_tomcat_contexts: [ 'whn-manager' ] -smartgears_scopes: - - '/{{ smartgears_infrastructure_name }}' # The iptables rules use this http_port: '{{ smartgears_http_port }}' diff --git a/smartgears/smartgears/tasks/install-gcube-keys.yml b/smartgears/smartgears/tasks/install-gcube-keys.yml index 9c45615..add8743 100644 --- a/smartgears/smartgears/tasks/install-gcube-keys.yml +++ b/smartgears/smartgears/tasks/install-gcube-keys.yml @@ -7,6 +7,7 @@ - '{{ gcube_prod_key_2 }}' - '{{ gcube_prod_key_3 }}' - '{{ gcube_prod_key_4 }}' + - '{{ gcube_prod_key_5 }}' notify: Restart smartgears when: install_gcube_prod_key @@ -17,6 +18,7 @@ - '{{ gcube_prod_key_2 }}' - '{{ gcube_prod_key_3 }}' - '{{ gcube_prod_key_4 }}' + - '{{ gcube_prod_key_5 }}' notify: Restart smartgears when: not install_gcube_prod_key diff --git a/smartgears/smartgears/tasks/smartgears-app.yml b/smartgears/smartgears/tasks/smartgears-app.yml index 3af684e..da2c239 100644 --- a/smartgears/smartgears/tasks/smartgears-app.yml +++ b/smartgears/smartgears/tasks/smartgears-app.yml @@ -58,6 +58,26 @@ copy: src=TokenGenerator.class dest=/usr/local/lib/TokenGenerator.class owner=root group=root mode=0644 tags: [ 'smartgears', 'smartgears_conf', 'tomcat' ] +- name: Install the script that manages the smartgears loglevel + template: src=change-logback-loglevel.sh.j2 dest=/usr/local/bin/change-logback-loglevel owner=root group=root mode=0755 + with_items: '{{ tomcat_m_instances }}' + when: not item.skip_smartgears + tags: [ 'smartgears', 'smartgears_loglevel', 'tomcat' ] + +- name: Change the smartgears log level + become: True + become_user: '{{ smartgears_user }}' + shell: /usr/local/bin/change-logback-loglevel + tags: [ 'smartgears', 'tomcat', 'smartgears_loglevel' ] + +- name: Remove some wrong symbolic links created by the install/upgrade script + file: dest={{ item }} state=absent + with_items: + - '{{ smartgears_install_path }}/lib/lib' + - '{{ smartgears_install_path }}/apps/webapps' + when: smartgears_upgrade + tags: [ 'smartgears', 'smartgears_conf', 'tomcat' ] + - name: Install the script that fetches the scope tokens template: src=get-scopes.j2 dest=/usr/local/bin/get-scopes owner=root group={{ smartgears_user }} mode=0750 with_items: '{{ tomcat_m_instances }}' @@ -81,38 +101,20 @@ with_items: '{{ tomcat_m_instances }}' tags: [ 'smartgears', 'smartgears_conf', 'tomcat' ] -- name: Get the scope tokens from the authorization service and assemble the container.xml file +- name: Get the scope tokens from the authorization service and assemble the container.xml file when whe have an authorization token or we are upgrading become: True become_user: '{{ smartgears_user }}' - shell: /usr/local/bin/get-scopes {{ gcube_admin_token | default(omit) }} - when: gcube_admin_token is defined + shell: /usr/local/bin/get-scopes {{ gcube_admin_token | default('') }} + when: gcube_admin_token is defined or smartgears_upgrade + notify: Restart smartgears tags: [ 'smartgears', 'smartgears_conf', 'tomcat' ] - name: Remove the smartgears application state if requested - #file: dest={{ smartgears_install_path }}/state state=absent - file: dest=/home/gcube/SmartGears/state state=absent - when: smartgears_remove_state - notify: Restart smartgears - tags: [ 'smartgears', 'smartgears_conf', 'tomcat' ] - -- name: Install the script that manages the smartgears loglevel - template: src=change-logback-loglevel.sh.j2 dest=/usr/local/bin/change-logback-loglevel owner=root group=root mode=0755 - with_items: '{{ tomcat_m_instances }}' - when: not item.skip_smartgears - tags: [ 'smartgears', 'smartgears_loglevel', 'tomcat' ] - -- name: Change the smartgears log level become: True become_user: '{{ smartgears_user }}' - shell: /usr/local/bin/change-logback-loglevel - tags: [ 'smartgears', 'tomcat', 'smartgears_loglevel' ] - -- name: Remove some wrong symbolic links created by the install/upgrade script - file: dest={{ item }} state=absent - with_items: - - '{{ smartgears_install_path }}/lib/lib' - - '{{ smartgears_install_path }}/apps/webapps' - when: smartgears_upgrade + shell: . {{ smartgears_user_home }}/.bashrc ; cd {{ smartgears_user_home }}/SmartGears/scripts ; ./clean-container-state -s old_saved_scopes.xml + when: smartgears_remove_state and not smartgears_upgrade + notify: Restart smartgears tags: [ 'smartgears', 'smartgears_conf', 'tomcat' ] - name: Force a smartgears restart diff --git a/smartgears/smartgears/templates/container.xml.j2 b/smartgears/smartgears/templates/container.xml.j2 index 9f20eac..961537b 100644 --- a/smartgears/smartgears/templates/container.xml.j2 +++ b/smartgears/smartgears/templates/container.xml.j2 @@ -24,6 +24,9 @@ {{ smartgears_longitude }} + +{{ smartgears_authorized_on_all_scopes }} + {{ smartgears_publication_frequency }} diff --git a/smartgears/smartgears/templates/get-scopes.j2 b/smartgears/smartgears/templates/get-scopes.j2 index 5f835fb..70a74d1 100644 --- a/smartgears/smartgears/templates/get-scopes.j2 +++ b/smartgears/smartgears/templates/get-scopes.j2 @@ -6,11 +6,37 @@ CONTAINER_XML_TAIL={{ smartgears_user_home }}/.containerxml/3-container.xml LOCAL_LIB=/usr/local/lib LOCAL_ETC=/usr/local/etc LOG_PREFIX="get-scopes: " +GHN_ENV_FILE=/etc/default/tomcat-instance-{{ item.http_port }}.local +SMARTGEARS_VO_AUTH={{ smartgears_authorized_on_all_scopes }} +SMARTGEARS_UPGRADE={{ smartgears_upgrade }} +SMARTGEARS_SAVED_STATE_F=saved_scopes_list.xml +SMARTGEARS_SAVED_STATE_PATH={{ smartgears_user_home }}/SmartGears/$SMARTGEARS_SAVED_STATE_F +SMARTGEARS_SCRIPTS_DIR={{ smartgears_user_home }}/SmartGears/scripts + CONTAINER_XML_FILE={{ smartgears_install_path }}/container.xml +# 0: True, 1: False +USE_SAVED_STATE=1 + RETVAL= +# Scenario: +# - First installation, no upgrade. +# - The node must run on all VOs +# In these cases we use our scopes list + +if [ "$SMARTGEARS_VO_AUTH" == 'false' ] ; then + if [ "$SMARTGEARS_UPGRADE" == 'True' ] ; then + USE_SAVED_STATE=0 + logger "$LOG_PREFIX setting the correct variables so that we are going to use the local scopes" + else + logger "$LOG_PREFIX We are going to use our scopes list. A valid token is mandatory" + fi +else + logger "$LOG_PREFIX We are going to use our scopes list. A valid token is mandatory" +fi + SCOPES_LIST="" if [ -f $LOCAL_ETC/scopes.list ] ; then . $LOCAL_ETC/scopes.list @@ -18,14 +44,16 @@ else logger "$LOG_PREFIX There is no token list, aborting" exit 1 fi - -if [ $# -ne 1 ] ; then - echo "The token must be passed as the sole argument" - logger "$LOG_PREFIX The token must be passed as the sole argument" + +if [ $# -ne 1 -a $USE_SAVED_STATE -ne 0 ] ; then + echo "The token must be passed as the sole argument when we are not using the local state" + logger "$LOG_PREFIX The token must be passed as the sole argument when we are not using the local state" exit 1 +elif [ $# -eq 1 ] ; then + logger "$LOG_PREFIX We have an authorization token" + TOKEN=$1 fi -TOKEN=$1 {%if setup_nginx %} {%if https_port is defined %} @@ -43,21 +71,33 @@ HTTP_PORT={{ http_port }} HTTP_PORT={{ item.http_port }} {% endif %} -for jar in $( ls -1 /home/gcube/tomcat/lib/ ) ; do - export CLASSPATH="/home/gcube/SmartGears/lib/${jar}:$CLASSPATH" -done +if [ $USE_SAVED_STATE -ne 0 ] ; then + logger "$LOG_PREFIX First installation or moving avay to a configuration that needs to be present on all the VREs. Using our scopes list and not the state memorized one" + for jar in $( ls -1 /home/gcube/tomcat/lib/ ) ; do + export CLASSPATH="/home/gcube/SmartGears/lib/${jar}:$CLASSPATH" + done -cd $LOCAL_LIB + cd $LOCAL_LIB -java TokenGenerator {{ smartgears_hostname }} $TOKEN $HTTP_PORT $SCOPES_FILE $SCOPES_LIST -RETVAL=$? -if [ $RETVAL -eq 0 ] ; then - logger "$LOG_PREFIX We got the scope tokens" + java TokenGenerator {{ smartgears_hostname }} $TOKEN $HTTP_PORT $SCOPES_FILE $SCOPES_LIST >/dev/null 2>&1 + RETVAL=$? + if [ $RETVAL -eq 0 ] ; then + logger "$LOG_PREFIX We got the scope tokens" + else + logger "$LOG_PREFIX Unable to obtain the scope tokens, aborting" + exit 1 + fi else - logger "$LOG_PREFIX Unable to obtain the scope tokens, aborting" - exit 1 + logger "$LOG_PREFIX We are going to use the scopes memorized into the state" + SCOPES_FILE=$SMARTGEARS_SAVED_STATE_PATH fi +# We always remove the current state +cd $SMARTGEARS_SCRIPTS_DIR +. $GHN_ENV_FILE +./clean-container-state -s $SMARTGEARS_SAVED_STATE_F + + # Now that we have the tokens, we can assemble the container.xml file chmod 640 $CONTAINER_XML_FILE CREATE_CONTAINER_XML_RES=0