spark-env.install 1.4 KB

1234567891011121314151617181920212223242526
  1. mkdir -p ${CONFIG_DIR}/rpmnew
  2. mkdir -p ${CONFIG_DIR}/externals/spark-hadoop
  3. printf "Installing %-44s ..." "spark-env.sh"
  4. if [ ! -e ${CONFIG_DIR}/spark-env.sh ]; then
  5. # Always install new files without comment
  6. cp -f ${INSTALL_DIR}/externals/spark-hadoop/conf/spark-env.sh ${CONFIG_DIR}/externals/spark-hadoop/spark-env.sh
  7. cp -f ${INSTALL_DIR}/externals/spark-hadoop/conf/spark-env.sh ${CONFIG_DIR}/rpmnew/spark-env.sh
  8. log_success_msg
  9. elif [ -e ${CONFIG_DIR}/rpmnew/spark-env.sh ] && ! `diff -q ${CONFIG_DIR}/rpmnew/spark-env.sh ${INSTALL_DIR}/externals/spark-hadoop/conf/spark-env.sh >/dev/null` ; then
  10. # There are changes in the default config since last installed
  11. if ! `diff -q ${CONFIG_DIR}/rpmnew/spark-env.sh ${CONFIG_DIR}/externals/spark-hadoop/spark-env.sh >/dev/null` ; then
  12. # User has made their own changes too, so don't overwrite
  13. log_failure_msg "Not overwriting modified configuration file spark-env.sh"
  14. else
  15. # User has NOT made their own changes - ok to update
  16. cp -f ${INSTALL_DIR}/externals/spark-hadoop/conf/spark-env.sh ${CONFIG_DIR}/externals/spark-hadoop/spark-env.sh
  17. cp -f ${INSTALL_DIR}/externals/spark-hadoop/conf/spark-env.sh ${CONFIG_DIR}/rpmnew/spark-env.sh
  18. log_success_msg "Updated configuration file spark-env.sh"
  19. fi
  20. else
  21. log_success_msg "No changes to configuration file spark-env.sh"
  22. fi