applied-ai-018's picture
Add files using upload-large-folder tool
dd4da36 verified
#!/bin/bash
set -e
trap cleanup_on_error ERR SIGINT SIGTERM
cleanup_on_error() {
echo "Error: $0:$LINENO: command \`$BASH_COMMAND\` failed with exit code $?"
exit 1
}
help() {
echo "Usage: apptainer_run_quality_signals.sh [ -c | --config ] [ -d | --dump_id ]"
exit 2
}
while [[ $# -gt 0 ]]; do
key="$1"
case $key in
-c | --config)
CONFIG_FILE="$2"
shift 2
;;
-d | --dump_id)
DUMP_ID="$2"
shift 2
;;
-l | --listings)
LISTINGS="$2"
shift 2
;;
-h | --help)
help
;;
--)
shift
break
;;
*)
echo "Invalid option: -$1"
help
;;
esac
done
# make environment variables available to downstream scripts
set -a
# shellcheck source=configs/base.conf
. "$CONFIG_FILE"
set +a
if [ -z "${MAX_DOCS}" ]; then
MAX_DOCS=-1
fi
ARTIFACTS_ARCHIVE="${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}.tar.gz"
if [ ! -d "${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}" ]; then
# download artifacts from bucket
echo "Downloading artifacts from ${INPUT_BASE_URI%/}/artifacts-${ARTIFACTS_ID}.tar.gz"
s5cmd --profile "$S3_PROFILE" --endpoint-url "$S3_ENDPOINT_URL" \
cp "${S3_BUCKET%/}/artifacts/artifacts-${ARTIFACTS_ID}.tar.gz" "${ARTIFACTS_ARCHIVE}"
# extract artifacts
mkdir -p "${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}"
echo "Extracting artifacts to ${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}"
tar -xzf "${ARTIFACTS_ARCHIVE}" -C "${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}"
rm "${ARTIFACTS_ARCHIVE}"
else
echo "Artifacts already exist at ${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}; skipping download."
fi
# run pipeline
ARTIFACTS_DIR="${DATA_ROOT%/}/artifacts-${ARTIFACTS_ID}"
if [ -z "${LISTINGS}" ]; then
LISTINGS="${ARTIFACTS_DIR%/}/listings/listings-${DUMP_ID}.txt"
fi
apptainer cache clean -f
apptainer run \
--env AWS_ACCESS_KEY_ID="$AWS_ACCESS_KEY_ID" --env AWS_SECRET_ACCESS_KEY="$AWS_SECRET_ACCESS_KEY" \
"docker://docker.io/${DOCKER_REPO}:amd64" \
python3 /usr/app/src/pipeline.py \
--input "${LISTINGS}" \
--input_base_uri "${INPUT_BASE_URI}" \
--output_base_uri "${OUTPUT_BASE_URI}" \
--cc_snapshot_id "${DUMP_ID}" \
--artifacts_dir "${ARTIFACTS_DIR}" \
--dsir_buckets "${DSIR_FEATURE_DIM}" \
--max_docs "${MAX_DOCS}" \
--inputs_per_process "${INPUTS_PER_PROCESS}" \
--langs "${LANGUAGES[@]}" \
--endpoint_url "${S3_ENDPOINT_URL}" \
--minhash_ngram_size "${MINHASH_NGRAM_SIZE}" \
--minhash_num_permutations "${MINHASH_NUM_PERMUTATIONS}" \
--minhash_similarities "${MINHASH_SIMILARITIES[@]}" \
--filename_keep_patterns "${FILENAME_KEEP_PATTERNS[@]}"