diff --git a/bin/cloud-local.sh b/bin/cloud-local.sh
index 4866fab..e7561e1 100755
--- a/bin/cloud-local.sh
+++ b/bin/cloud-local.sh
@@ -84,9 +84,11 @@ function download_packages {
local maven=${pkg_src_maven}
declare -a urls=("${apache_archive_url}/hadoop/common/hadoop-${pkg_hadoop_ver}/hadoop-${pkg_hadoop_ver}.tar.gz"
- "${apache_archive_url}/zookeeper/zookeeper-${pkg_zookeeper_ver}/zookeeper-${pkg_zookeeper_ver}.tar.gz"
- "${apache_archive_url}/spark/spark-${pkg_spark_ver}/spark-${pkg_spark_ver}-bin-${pkg_spark_hadoop_ver}.tgz")
+ "${apache_archive_url}/zookeeper/zookeeper-${pkg_zookeeper_ver}/zookeeper-${pkg_zookeeper_ver}.tar.gz")
+ if [[ "$spark_enabled" -eq 1 ]]; then
+ urls=("${urls[@]}" "${apache_archive_url}/spark/spark-${pkg_spark_ver}/spark-${pkg_spark_ver}-bin-${pkg_spark_hadoop_ver}.tgz")
+ fi
if [[ "$kafka_enabled" -eq 1 ]]; then
urls=("${urls[@]}" "${apache_archive_url}/kafka/${pkg_kafka_ver}/kafka_${pkg_kafka_scala_ver}-${pkg_kafka_ver}.tgz")
@@ -134,9 +136,10 @@ function unpackage {
[[ "$acc_enabled" -eq 1 ]] && (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/accumulo-${pkg_accumulo_ver}-bin.tar.gz") && echo "Unpacked accumulo"
[[ "$hbase_enabled" -eq 1 ]] && (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/hbase-${pkg_hbase_ver}-bin.tar.gz") && echo "Unpacked hbase"
[[ "$zeppelin_enabled" -eq 1 ]] && (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/zeppelin-${pkg_zeppelin_ver}-bin-all.tgz") && echo "Unpacked zeppelin"
- (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/hadoop-${pkg_hadoop_ver}.tar.gz") && echo "Unpacked hadoop"
[[ "$kafka_enabled" -eq 1 ]] && (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/kafka_${pkg_kafka_scala_ver}-${pkg_kafka_ver}.tgz") && echo "Unpacked kafka"
- (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/spark-${pkg_spark_ver}-bin-${pkg_spark_hadoop_ver}.tgz") && echo "Unpacked spark"
+ [[ "$spark_enabled" -eq 1 ]] \
+ && (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/spark-${pkg_spark_ver}-bin-${pkg_spark_hadoop_ver}.tgz") && echo "Unpacked spark"
+ (cd -P "${CLOUD_HOME}" && tar $targs "${CLOUD_HOME}/pkg/hadoop-${pkg_hadoop_ver}.tar.gz") && echo "Unpacked hadoop"
}
function configure {
@@ -144,18 +147,16 @@ function configure {
cp -r ${CLOUD_HOME}/templates/* ${CLOUD_HOME}/tmp/staging/
# accumulo config before substitutions
- [[ "$acc_enabled" -eq 1 ]] && cp $ACCUMULO_HOME/conf/examples/3GB/standalone/* $ACCUMULO_HOME/conf/
+ ##[[ "$acc_enabled" -eq 1 ]] && cp $ACCUMULO_HOME/conf/examples/3GB/standalone/* $ACCUMULO_HOME/conf/
## Substitute env vars
sed -i~orig "s#LOCAL_CLOUD_PREFIX#${CLOUD_HOME}#;s#CLOUD_LOCAL_HOSTNAME#${CL_HOSTNAME}#;s#CLOUD_LOCAL_BIND_ADDRESS#${CL_BIND_ADDRESS}#" ${CLOUD_HOME}/tmp/staging/*/*
if [[ "$acc_enabled" -eq 1 ]]; then
# accumulo config
- # make accumulo bind to all network interfaces (so you can see the monitor from other boxes)
- sed -i~orig "s/\# export ACCUMULO_MONITOR_BIND_ALL=\"true\"/export ACCUMULO_MONITOR_BIND_ALL=\"true\"/" "${ACCUMULO_HOME}/conf/accumulo-env.sh"
echo "${CL_HOSTNAME}" > ${ACCUMULO_HOME}/conf/gc
echo "${CL_HOSTNAME}" > ${ACCUMULO_HOME}/conf/masters
- echo "${CL_HOSTNAME}" > ${ACCUMULO_HOME}/conf/slaves
+ echo "${CL_HOSTNAME}" > ${ACCUMULO_HOME}/conf/tservers
echo "${CL_HOSTNAME}" > ${ACCUMULO_HOME}/conf/monitor
echo "${CL_HOSTNAME}" > ${ACCUMULO_HOME}/conf/tracers
fi
@@ -187,11 +188,27 @@ function configure {
[[ "$zeppelin_enabled" -eq 1 ]] && cp ${CLOUD_HOME}/tmp/staging/zeppelin/* ${ZEPPELIN_HOME}/conf/
# If Spark doesn't have log4j settings, use the Spark defaults
- test -f $SPARK_HOME/conf/log4j.properties || cp $SPARK_HOME/conf/log4j.properties.template $SPARK_HOME/conf/log4j.properties
+ test -f $SPARK_HOME/conf/log4j.properties || [[ "$spark_enabled" -eq 1 ]] && cp $SPARK_HOME/conf/log4j.properties.template $SPARK_HOME/conf/log4j.properties
# configure port offsets
configure_port_offset
+ # As of Accumulo 2 accumulo-site.xml is nolonger allowed. To avoid a lot of work rewriting the ports script we'll just use accumulo's converter.
+ if [ -f "$ACCUMULO_HOME/conf/accumulo-site.xml" ]; then
+ rm -f "$ACCUMULO_HOME/conf/accumulo.properties"
+ "$ACCUMULO_HOME/bin/accumulo" convert-config \
+ -x "$ACCUMULO_HOME/conf/accumulo-site.xml" \
+ -p "$ACCUMULO_HOME/conf/accumulo.properties"
+ rm -f "$ACCUMULO_HOME/conf/accumulo-site.xml"
+ fi
+
+ # Configure accumulo-client.properties
+ if [ -f "$ACCUMULO_HOME/conf/accumulo-client.properties" ]; then
+ sed -i "s/.*instance.name=.*$/instance.name=$cl_acc_inst_name/" "$ACCUMULO_HOME/conf/accumulo-client.properties"
+ sed -i "s/.*auth.principal=.*$/auth.principal=root/" "$ACCUMULO_HOME/conf/accumulo-client.properties"
+ sed -i "s/.*auth.token=.*$/auth.token=$cl_acc_inst_pass/" "$ACCUMULO_HOME/conf/accumulo-client.properties"
+
+ fi
rm -rf ${CLOUD_HOME}/tmp/staging
}
@@ -213,15 +230,15 @@ function start_first_time {
# format namenode
echo "Formatting namenode..."
- $HADOOP_HOME/bin/hadoop namenode -format
+ $HADOOP_HOME/bin/hdfs namenode -format
# start hadoop
echo "Starting hadoop..."
- $HADOOP_HOME/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR start namenode
- $HADOOP_HOME/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR start secondarynamenode
- $HADOOP_HOME/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR start datanode
- $HADOOP_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR start resourcemanager
- $HADOOP_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR start nodemanager
+ $HADOOP_HOME/bin/hdfs --config $HADOOP_CONF_DIR --daemon start namenode
+ $HADOOP_HOME/bin/hdfs --config $HADOOP_CONF_DIR --daemon start secondarynamenode
+ $HADOOP_HOME/bin/hdfs --config $HADOOP_CONF_DIR --daemon start datanode
+ $HADOOP_HOME/bin/yarn --config $HADOOP_CONF_DIR --daemon start resourcemanager
+ $HADOOP_HOME/bin/yarn --config $HADOOP_CONF_DIR --daemon start nodemanager
# Wait for HDFS to exit safemode:
echo "Waiting for HDFS to exit safemode..."
@@ -244,7 +261,7 @@ function start_first_time {
# starting accumulo
echo "Starting accumulo..."
- $ACCUMULO_HOME/bin/start-all.sh
+ $ACCUMULO_HOME/bin/accumulo-cluster start
fi
if [[ "$hbase_enabled" -eq 1 ]]; then
@@ -286,9 +303,9 @@ function start_cloud {
# start hadoop
echo "Starting hadoop..."
- hadoop-daemon.sh --config $HADOOP_CONF_DIR start namenode
- hadoop-daemon.sh --config $HADOOP_CONF_DIR start secondarynamenode
- hadoop-daemon.sh --config $HADOOP_CONF_DIR start datanode
+ hdfs --config $HADOOP_CONF_DIR --daemon start namenode
+ hdfs --config $HADOOP_CONF_DIR --daemon start secondarynamenode
+ hdfs --config $HADOOP_CONF_DIR --daemon start datanode
start_yarn
# Wait for HDFS to exit safemode:
@@ -298,7 +315,7 @@ function start_cloud {
if [[ "$acc_enabled" -eq 1 ]]; then
# starting accumulo
echo "starting accumulo..."
- $ACCUMULO_HOME/bin/start-all.sh
+ $ACCUMULO_HOME/bin/accumulo-cluster start
fi
if [[ "$hbase_enabled" -eq 1 ]]; then
@@ -321,8 +338,8 @@ function start_cloud {
}
function start_yarn {
- $HADOOP_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR start resourcemanager
- $HADOOP_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR start nodemanager
+ $HADOOP_HOME/bin/yarn --config $HADOOP_CONF_DIR --daemon start resourcemanager
+ $HADOOP_HOME/bin/yarn --config $HADOOP_CONF_DIR --daemon start nodemanager
}
function start_geoserver {
@@ -349,7 +366,7 @@ function stop_cloud {
if [[ "$acc_enabled" -eq 1 ]]; then
echo "Stopping accumulo..."
- $ACCUMULO_HOME/bin/stop-all.sh
+ $ACCUMULO_HOME/bin/accumulo-cluster stop
fi
if [[ "$hbase_enabled" -eq 1 ]]; then
@@ -359,9 +376,9 @@ function stop_cloud {
echo "Stopping yarn and dfs..."
stop_yarn
- $HADOOP_HOME/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR stop namenode
- $HADOOP_HOME/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR stop secondarynamenode
- $HADOOP_HOME/sbin/hadoop-daemon.sh --config $HADOOP_CONF_DIR stop datanode
+ $HADOOP_HOME/bin/hdfs --config $HADOOP_CONF_DIR --daemon stop namenode
+ $HADOOP_HOME/bin/hdfs --config $HADOOP_CONF_DIR --daemon stop secondarynamenode
+ $HADOOP_HOME/bin/hdfs --config $HADOOP_CONF_DIR --daemon stop datanode
echo "Stopping zookeeper..."
$ZOOKEEPER_HOME/bin/zkServer.sh stop
@@ -434,8 +451,8 @@ function verify_stop {
}
function stop_yarn {
- $HADOOP_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR stop resourcemanager
- $HADOOP_HOME/sbin/yarn-daemon.sh --config $HADOOP_CONF_DIR stop nodemanager
+ $HADOOP_HOME/bin/yarn --config $HADOOP_CONF_DIR --daemon stop resourcemanager
+ $HADOOP_HOME/bin/yarn --config $HADOOP_CONF_DIR --daemon stop nodemanager
}
function stop_geoserver {
diff --git a/bin/config.sh b/bin/config.sh
index 646a4ae..0fe00bf 100755
--- a/bin/config.sh
+++ b/bin/config.sh
@@ -30,7 +30,7 @@ function validate_config {
# allowed versions are
local pkg_error=""
# hadoop 2.[5-9].x
- if [[ -z "$pkg_hadoop_ver" || ! $pkg_hadoop_ver =~ 2[.][56789][.]. ]]; then
+ if [[ -z "$pkg_hadoop_ver" || ! $pkg_hadoop_ver =~ 2[.][56789][.].|3[.]2[.]1 ]]; then
pkg_error="${pkg_error}Invalid hadoop version: '${pkg_hadoop_ver}' ${NL}"
fi
# zk 3.4.[5-10]
@@ -38,7 +38,7 @@ function validate_config {
pkg_error="${pkg_error}Invalid zookeeper version: '${pkg_zookeeper_ver}' ${NL}"
fi
# acc 1.[6-9].x
- if [[ -z "$pkg_accumulo_ver" || ! $pkg_accumulo_ver =~ 1[.][6789][.]. ]]; then
+ if [[ -z "$pkg_accumulo_ver" || ! $pkg_accumulo_ver =~ 1[.][6789][.].|2[.]0[.]0 ]]; then
pkg_error="${pkg_error}Invalid accumulo version: '${pkg_accumulo_ver}' ${NL}"
fi
# kafka 0.9.x, 0.10.x, 0.11.x, 1.0.x
@@ -78,8 +78,7 @@ function set_env_vars {
fi
export HADOOP_HOME="$CLOUD_HOME/hadoop-${pkg_hadoop_ver}"
- export HADOOP_PREFIX="${HADOOP_HOME}"
- export HADOOP_CONF_DIR="${HADOOP_PREFIX}/etc/hadoop"
+ export HADOOP_CONF_DIR="${HADOOP_HOME}/etc/hadoop"
export HADOOP_COMMON_HOME="${HADOOP_HOME}"
export HADOOP_HDFS_HOME="${HADOOP_HOME}"
export HADOOP_YARN_HOME="${HADOOP_HOME}"
@@ -87,8 +86,6 @@ function set_env_vars {
export HADOOP_IDENT_STRING=$(echo ${CLOUD_HOME} | (md5sum 2>/dev/null || md5) | cut -c1-32)
export YARN_HOME="${HADOOP_HOME}"
- export YARN_PID_DIR="${HADOOP_PID_DIR}"
- export YARN_IDENT_STRING="${HADOOP_IDENT_STRING}"
export SPARK_HOME="$CLOUD_HOME/spark-${pkg_spark_ver}-bin-${pkg_spark_hadoop_ver}"
diff --git a/bin/ports.sh b/bin/ports.sh
index 7f7694c..9485de0 100755
--- a/bin/ports.sh
+++ b/bin/ports.sh
@@ -96,9 +96,11 @@ function configure_port_offset {
sed -i~orig "s/clientPort=.*/clientPort=$zkPort/" $ZOOKEEPER_HOME/conf/zoo.cfg
# kafka (server.properties)
- kafkaPort=$((9092+offset))
- sed -i~orig "s/\/\/$CL_HOSTNAME:[0-9].*/\/\/$CL_HOSTNAME:$kafkaPort/" $KAFKA_HOME/config/server.properties
- sed -i~orig "s/zookeeper.connect=$CL_HOSTNAME:[0-9].*/zookeeper.connect=$CL_HOSTNAME:$zkPort/" $KAFKA_HOME/config/server.properties
+ if [[ "kafka_enabled" -eq 1 ]]; then
+ kafkaPort=$((9092+offset))
+ sed -i~orig "s/\/\/$CL_HOSTNAME:[0-9].*/\/\/$CL_HOSTNAME:$kafkaPort/" $KAFKA_HOME/config/server.properties
+ sed -i~orig "s/zookeeper.connect=$CL_HOSTNAME:[0-9].*/zookeeper.connect=$CL_HOSTNAME:$zkPort/" $KAFKA_HOME/config/server.properties
+ fi
# Zeppelin
if [[ "$zeppelin_enabled" -eq 1 ]]; then
diff --git a/conf/cloud-local.conf b/conf/cloud-local.conf
index 8be2f4d..27f1cef 100644
--- a/conf/cloud-local.conf
+++ b/conf/cloud-local.conf
@@ -24,10 +24,10 @@ pkg_src_maven="https://repo1.maven.org/maven2"
# VERSION MANAGEMENT - Versions of popular software
################################################################################
-pkg_accumulo_ver="1.9.2"
+pkg_accumulo_ver="2.0.0"
pkg_hbase_ver="1.3.1"
# Note pkg_spark_hadoop_ver below if modifying
-pkg_hadoop_ver="2.8.4"
+pkg_hadoop_ver="3.2.1"
# Note, just the major+minor from Hadoop, not patch level
hadoop_base_ver=${pkg_hadoop_ver:0:3}
@@ -100,7 +100,7 @@ hbase_enabled=0
kafka_enabled=0
# Download spark distribution
-spark_enabled=1
+spark_enabled=0
# Enable/Disable installation of GeoMesa
geomesa_enabled=0
diff --git a/templates/accumulo/accumulo-env.sh b/templates/accumulo/accumulo-env.sh
new file mode 100644
index 0000000..e11fd49
--- /dev/null
+++ b/templates/accumulo/accumulo-env.sh
@@ -0,0 +1,124 @@
+#! /usr/bin/env bash
+
+# Licensed to the Apache Software Foundation (ASF) under one or more
+# contributor license agreements. See the NOTICE file distributed with
+# this work for additional information regarding copyright ownership.
+# The ASF licenses this file to You under the Apache License, Version 2.0
+# (the "License"); you may not use this file except in compliance with
+# the License. You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+## Before accumulo-env.sh is loaded, these environment variables are set and can be used in this file:
+
+# cmd - Command that is being called such as tserver, master, etc.
+# basedir - Root of Accumulo installation
+# bin - Directory containing Accumulo scripts
+# conf - Directory containing Accumulo configuration
+# lib - Directory containing Accumulo libraries
+
+############################
+# Variables that must be set
+############################
+
+## Accumulo logs directory. Referenced by logger config.
+export ACCUMULO_LOG_DIR="${ACCUMULO_LOG_DIR:-${basedir}/logs}"
+## Hadoop installation
+export HADOOP_HOME="${HADOOP_HOME:-/path/to/hadoop}"
+## Hadoop configuration
+export HADOOP_CONF_DIR="${HADOOP_CONF_DIR:-${HADOOP_HOME}/etc/hadoop}"
+## Zookeeper installation
+export ZOOKEEPER_HOME="${ZOOKEEPER_HOME:-/path/to/zookeeper}"
+
+##########################
+# Build CLASSPATH variable
+##########################
+
+## Verify that Hadoop & Zookeeper installation directories exist
+if [[ ! -d "$ZOOKEEPER_HOME" ]]; then
+ echo "ZOOKEEPER_HOME=$ZOOKEEPER_HOME is not set to a valid directory in accumulo-env.sh"
+ exit 1
+fi
+if [[ ! -d "$HADOOP_HOME" ]]; then
+ echo "HADOOP_HOME=$HADOOP_HOME is not set to a valid directory in accumulo-env.sh"
+ exit 1
+fi
+
+## Build using existing CLASSPATH, conf/ directory, dependencies in lib/, and external Hadoop & Zookeeper dependencies
+if [[ -n "$CLASSPATH" ]]; then
+ CLASSPATH="${CLASSPATH}:${conf}"
+else
+ CLASSPATH="${conf}"
+fi
+CLASSPATH="${CLASSPATH}:${lib}/*:${HADOOP_CONF_DIR}:${ZOOKEEPER_HOME}/*:${HADOOP_HOME}/share/hadoop/client/*"
+export CLASSPATH
+
+##################################################################
+# Build JAVA_OPTS variable. Defaults below work but can be edited.
+##################################################################
+
+## JVM options set for all processes. Extra options can be passed in by setting ACCUMULO_JAVA_OPTS to an array of options.
+JAVA_OPTS=("${ACCUMULO_JAVA_OPTS[@]}"
+ '-XX:+UseConcMarkSweepGC'
+ '-XX:CMSInitiatingOccupancyFraction=75'
+ '-XX:+CMSClassUnloadingEnabled'
+ '-XX:OnOutOfMemoryError=kill -9 %p'
+ '-XX:-OmitStackTraceInFastThrow'
+ '-Djava.net.preferIPv4Stack=true'
+ "-Daccumulo.native.lib.path=${lib}/native")
+
+## Make sure Accumulo native libraries are built since they are enabled by default
+"${bin}"/accumulo-util build-native &> /dev/null
+
+## JVM options set for individual applications
+case "$cmd" in
+ master) JAVA_OPTS=("${JAVA_OPTS[@]}" '-Xmx512m' '-Xms512m') ;;
+ monitor) JAVA_OPTS=("${JAVA_OPTS[@]}" '-Xmx256m' '-Xms256m') ;;
+ gc) JAVA_OPTS=("${JAVA_OPTS[@]}" '-Xmx256m' '-Xms256m') ;;
+ tserver) JAVA_OPTS=("${JAVA_OPTS[@]}" '-Xmx3G' '-Xms3G') ;;
+ *) JAVA_OPTS=("${JAVA_OPTS[@]}" '-Xmx256m' '-Xms64m') ;;
+esac
+
+## JVM options set for logging. Review logj4 properties files to see how they are used.
+JAVA_OPTS=("${JAVA_OPTS[@]}"
+ "-Daccumulo.log.dir=${ACCUMULO_LOG_DIR}"
+ "-Daccumulo.application=${cmd}${ACCUMULO_SERVICE_INSTANCE}_$(hostname)")
+
+case "$cmd" in
+ monitor)
+ JAVA_OPTS=("${JAVA_OPTS[@]}" "-Dlog4j.configuration=log4j-monitor.properties")
+ ;;
+ gc|master|tserver|tracer)
+ JAVA_OPTS=("${JAVA_OPTS[@]}" "-Dlog4j.configuration=log4j-service.properties")
+ ;;
+ *)
+ # let log4j use its default behavior (log4j.xml, log4j.properties)
+ true
+ ;;
+esac
+
+export JAVA_OPTS
+
+############################
+# Variables set to a default
+############################
+
+export MALLOC_ARENA_MAX=${MALLOC_ARENA_MAX:-1}
+## Add Hadoop native libraries to shared library paths given operating system
+case "$(uname)" in
+ Darwin) export DYLD_LIBRARY_PATH="${HADOOP_HOME}/lib/native:${DYLD_LIBRARY_PATH}" ;;
+ *) export LD_LIBRARY_PATH="${HADOOP_HOME}/lib/native:${LD_LIBRARY_PATH}" ;;
+esac
+
+###############################################
+# Variables that are optional. Uncomment to set
+###############################################
+
+## Specifies command that will be placed before calls to Java in accumulo script
+# export ACCUMULO_JAVA_PREFIX=""
diff --git a/templates/accumulo/accumulo-site.xml b/templates/accumulo/accumulo-site.xml
index f7779a2..2667de8 100644
--- a/templates/accumulo/accumulo-site.xml
+++ b/templates/accumulo/accumulo-site.xml
@@ -123,51 +123,4 @@
1G
-
- general.classpaths
-
-
-
- $ACCUMULO_HOME/lib/accumulo-server.jar,
- $ACCUMULO_HOME/lib/accumulo-core.jar,
- $ACCUMULO_HOME/lib/accumulo-start.jar,
- $ACCUMULO_HOME/lib/accumulo-fate.jar,
- $ACCUMULO_HOME/lib/accumulo-proxy.jar,
- $ACCUMULO_HOME/lib/[^.].*.jar,
-
- $ZOOKEEPER_HOME/zookeeper[^.].*.jar,
-
- $HADOOP_CONF_DIR,
-
- $HADOOP_PREFIX/share/hadoop/common/[^.].*.jar,
- $HADOOP_PREFIX/share/hadoop/common/lib/(?!slf4j)[^.].*.jar,
- $HADOOP_PREFIX/share/hadoop/hdfs/[^.].*.jar,
- $HADOOP_PREFIX/share/hadoop/mapreduce/[^.].*.jar,
- $HADOOP_PREFIX/share/hadoop/yarn/[^.].*.jar,
- $HADOOP_PREFIX/share/hadoop/yarn/lib/jersey.*.jar,
-
-
-
-
-
- Classpaths that accumulo checks for updates and class files.
-