summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorChen, Chih-Chia <pigfoot@gmail.com>2021-11-26 17:07:02 +0800
committerChen, Chih-Chia <pigfoot@gmail.com>2021-11-26 17:07:02 +0800
commitfe9160c393bad0eb25f91e5e683e2a35f47ae8e8 (patch)
treea68f3630caca7826afb6dca591b2d955edff6f0b
parent[net-vpn/cloudflared] upgrade from upstream (diff)
downloadpigfoot-fe9160c393bad0eb25f91e5e683e2a35f47ae8e8.tar.gz
pigfoot-fe9160c393bad0eb25f91e5e683e2a35f47ae8e8.tar.bz2
pigfoot-fe9160c393bad0eb25f91e5e683e2a35f47ae8e8.zip
Remove out-of-dated packages
FIX: #824710 Signed-off-by: Chen, Chih-Chia <pigfoot@gmail.com>
-rw-r--r--sys-cluster/apache-spark-bin/Manifest1
-rw-r--r--sys-cluster/apache-spark-bin/apache-spark-bin-2.2.0.ebuild74
-rw-r--r--sys-cluster/apache-spark-bin/files/spark-master.service13
-rw-r--r--sys-cluster/apache-spark-bin/files/spark-worker.service14
-rw-r--r--sys-cluster/apache-spark-bin/files/spark.init33
5 files changed, 0 insertions, 135 deletions
diff --git a/sys-cluster/apache-spark-bin/Manifest b/sys-cluster/apache-spark-bin/Manifest
deleted file mode 100644
index b775885..0000000
--- a/sys-cluster/apache-spark-bin/Manifest
+++ /dev/null
@@ -1 +0,0 @@
-DIST spark-2.2.0-bin-hadoop2.7.tgz 203728858 SHA256 97fd2cc58e08975d9c4e4ffa8d7f8012c0ac2792bcd9945ce2a561cf937aebcc SHA512 7a186a2a007b2dfd880571f7214a7d329c972510a460a8bdbef9f7f2a891019343c020f74b496a61e5aa42bc9e9a79cc99defe5cb3bf8b6f49c07e01b259bc6b WHIRLPOOL 5d53db496d7f3f3224c089e19e70120fd1e00d25a3cdff333c14aebdb02fed82d596aa6446e0c289c2dcd397f33aea9f219e5fa86dc49b9e4b82ab71efe723f5
diff --git a/sys-cluster/apache-spark-bin/apache-spark-bin-2.2.0.ebuild b/sys-cluster/apache-spark-bin/apache-spark-bin-2.2.0.ebuild
deleted file mode 100644
index 990f6a9..0000000
--- a/sys-cluster/apache-spark-bin/apache-spark-bin-2.2.0.ebuild
+++ /dev/null
@@ -1,74 +0,0 @@
-# Copyright 1999-2017 Gentoo Foundation
-# Distributed under the terms of the GNU General Public License v2
-
-EAPI="6"
-inherit user systemd
-
-MY_PN="spark"
-
-DESCRIPTION="Software framework for fast cluster computing"
-HOMEPAGE="http://spark.apache.org/"
-SRC_URI="mirror://apache/${MY_PN}/${MY_PN}-${PV}/${MY_PN}-${PV}-bin-hadoop2.7.tgz"
-
-LICENSE="Apache-2.0"
-SLOT="0"
-KEYWORDS="~amd64 ~x86"
-IUSE="scala"
-
-DEPEND="scala? ( dev-lang/scala )"
-
-RDEPEND="virtual/jre
-scala? ( dev-lang/scala )"
-
-S="${WORKDIR}/${MY_PN}-${PV}-bin-hadoop2.7"
-INSTALL_DIR=/opt/${MY_PN}-${PV}
-
-pkg_setup(){
- enewgroup hadoop
- enewuser spark -1 /bin/bash /home/spark hadoop
-}
-
-src_install() {
- sandbox=`egrep -c "^[0-9].*#.* sandbox" /etc/hosts`
- workmem=1024m
- [ $sandbox -ne 0 ] && workmem=192m
-
- # create file spark-env.sh
- cat > conf/spark-env.sh <<-EOF
-SPARK_LOG_DIR=/var/log/spark
-SPARK_PID_DIR=/var/run/pids
-SPARK_LOCAL_DIRS=/var/lib/spark
-SPARK_WORKER_MEMORY=${workmem}
-SPARK_WORKER_DIR=/var/lib/spark
-EOF
-
- dodir "${INSTALL_DIR}"
- diropts -m770 -o spark -g hadoop
- dodir /var/log/spark
- dodir /var/lib/spark
- rm -f bin/*.cmd
- # dobin bin/*
- fperms g+w conf/*
- mv "${S}"/* "${D}${INSTALL_DIR}"
- fowners -Rf root:hadoop "${INSTALL_DIR}"
-
- # conf symlink
- dosym ${INSTALL_DIR}/conf /etc/spark
-
- cat > 99spark <<EOF
-SPARK_HOME="${INSTALL_DIR}"
-SPARK_CONF_DIR="/etc/spark"
-PATH="${INSTALL_DIR}/bin"
-EOF
- doenvd 99spark
-
- # init/systemd scripts
- newinitd "${FILESDIR}"/"${MY_PN}.init" "${MY_PN}-worker"
- systemd_newunit "${FILESDIR}/${MY_PN}-worker.service" "${MY_PN}-worker.service"
- if [ `egrep -c "^[0-9].*${HOSTNAME}.*#.* sparkmaster" /etc/hosts` -eq 1 ] ; then
- dosym /etc/init.d/"${MY_PN}-worker" /etc/init.d/"${MY_PN}-master"
- systemd_newunit "${FILESDIR}/${MY_PN}-master.service" "${MY_PN}-master.service"
- fi
-
- dosym "${INSTALL_DIR}" "/opt/${MY_PN}"
-}
diff --git a/sys-cluster/apache-spark-bin/files/spark-master.service b/sys-cluster/apache-spark-bin/files/spark-master.service
deleted file mode 100644
index 1dd573e..0000000
--- a/sys-cluster/apache-spark-bin/files/spark-master.service
+++ /dev/null
@@ -1,13 +0,0 @@
-[Unit]
-Description=Apache Spark master Daemon
-After=network.target systemd-resolved.service
-Requires=network.target
-
-[Service]
-Type=simple
-EnvironmentFile=/etc/env.d/99spark
-User=spark
-ExecStart={SPARK_HOME}/sbin/start-master.sh
-
-[Install]
-WantedBy=multi-user.target
diff --git a/sys-cluster/apache-spark-bin/files/spark-worker.service b/sys-cluster/apache-spark-bin/files/spark-worker.service
deleted file mode 100644
index 8a35379..0000000
--- a/sys-cluster/apache-spark-bin/files/spark-worker.service
+++ /dev/null
@@ -1,14 +0,0 @@
-[Unit]
-Description=Apache Spark worker Daemon
-After=network.target systemd-resolved.service
-Requires=network.target
-
-[Service]
-Type=simple
-EnvironmentFile=/etc/env.d/99spark
-User=spark
-ExecStartPre=/bin/bash -c "/bin/systemctl set-environment sparkmaster=$(egrep "^[0-9].*#.* sparkmaster" /etc/hosts | awk '{ print $3}')"
-ExecStart={SPARK_HOME}/sbin/start-slave.sh -- spark://${sparkmaster}:7077"
-
-[Install]
-WantedBy=multi-user.target
diff --git a/sys-cluster/apache-spark-bin/files/spark.init b/sys-cluster/apache-spark-bin/files/spark.init
deleted file mode 100644
index b55fcae..0000000
--- a/sys-cluster/apache-spark-bin/files/spark.init
+++ /dev/null
@@ -1,33 +0,0 @@
-#!/sbin/openrc-run
-#
-# Copyright 1999-2011 Gentoo Foundation
-# Distributed under the terms of the GNU General Public License v2
-# $Header: $
-
-DAEMON=${SVCNAME/spark-}
-SPARK_PID_DIR=/var/run/pids
-
-depend() {
- use dns net ssh
-}
-
-start() {
- [ -d $SPARK_PID_DIR ] || mkdir -m 777 $SPARK_PID_DIR
- if [ "$DAEMON" == "worker" ]; then
- sparkmaster=`egrep "^[0-9].*#.* sparkmaster" /etc/hosts | awk '{ print $3}' `
- CMD="$SPARK_HOME/sbin/start-slave.sh -- spark://$sparkmaster:7077"
- else
- CMD="$SPARK_HOME/sbin/start-master.sh"
- fi
-
- ebegin "Starting Apache Spark $DAEMON"
- start-stop-daemon --start --quiet --pidfile $SPARK_PID_DIR/spark-spark-org.apache.spark.deploy.${DAEMON}.${DAEMON^}-1.pid \
- -u spark -x ${CMD}
- eend $?
-}
-
-stop() {
- ebegin "Stopping Apache Spark $DAEMON"
- start-stop-daemon --stop --quiet --pidfile $SPARK_PID_DIR/spark-spark-org.apache.spark.deploy.${DAEMON}.${DAEMON^}-1.pid
- eend $?
-}