Skip to content

Commit 1ea2805

Browse files
committed
Run gzip with -f
1 parent 5fac4ee commit 1ea2805

File tree

47 files changed

+47
-47
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

47 files changed

+47
-47
lines changed

aurora-mysql/README.md

Lines changed: 1 addition & 1 deletion

aurora-postgresql/README.md

Lines changed: 1 addition & 1 deletion

bigquery/README.md

Lines changed: 1 addition & 1 deletion

bytehouse/NOTES.md

Lines changed: 1 addition & 1 deletion

bytehouse/README.md

Lines changed: 1 addition & 1 deletion

chdb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,7 @@ pip install --break-system-packages chdb==2.2.0b1
88

99
# Load the data
1010
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
11-
gzip -d hits.csv.gz
11+
gzip -d -f hits.csv.gz
1212
./load.py
1313

1414
# Run the queries

citus/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ sudo apt-get install -y postgresql-client
77
sudo docker run -d --name citus -p 5432:5432 -e POSTGRES_PASSWORD=mypass citusdata/citus:11.0
88

99
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
10-
gzip -d hits.tsv.gz
10+
gzip -d -f hits.tsv.gz
1111

1212
echo "*:*:*:*:mypass" > .pgpass
1313
chmod 400 .pgpass

clickhouse/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@ clickhouse-client < create"$SUFFIX".sql
4141
if [ ! -f hits.tsv ]
4242
then
4343
wget --no-verbose --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
44-
gzip -d hits.tsv.gz
44+
gzip -d -f hits.tsv.gz
4545
fi
4646

4747
clickhouse-client --time --query "INSERT INTO hits FORMAT TSV" < hits.tsv

cloudberry/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -105,7 +105,7 @@ elif [[ $1 == 'test' ]]; then
105105
chmod +x /home/gpadmin/run.sh
106106
chown gpadmin:gpadmin /home/gpadmin/*
107107
if [[ $2 != 'no_dl' ]]; then sudo -iu gpadmin wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'; fi
108-
if [[ $2 != 'no_dl' ]]; then sudo -iu gpadmin gzip -d hits.tsv.gz; fi
108+
if [[ $2 != 'no_dl' ]]; then sudo -iu gpadmin gzip -d -f hits.tsv.gz; fi
109109
sudo -iu gpadmin chmod 777 ~ hits.tsv
110110
sudo -iu gpadmin psql -d postgres -f /home/gpadmin/create.sql
111111
sudo -iu gpadmin nohup gpfdist &

cratedb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -35,7 +35,7 @@ do
3535
done
3636

3737
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz' -O /tmp/hits.tsv.gz
38-
gzip -d /tmp/hits.tsv.gz
38+
gzip -d -f /tmp/hits.tsv.gz
3939
chmod 444 /tmp/hits.tsv
4040

4141
psql -U crate -h localhost --no-password -t < $CREATE_FILE

doris/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -89,7 +89,7 @@ mysql -h 127.0.0.1 -P9030 -uroot hits <"$ROOT"/create.sql
8989
# Download data
9090
if [[ ! -f hits.tsv.gz ]] && [[ ! -f hits.tsv ]]; then
9191
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
92-
gzip -d hits.tsv.gz
92+
gzip -d -f hits.tsv.gz
9393
fi
9494

9595
# Load data

druid/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ echo "druid.query.groupBy.maxMergingDictionarySize=5000000000" >> apache-druid-$
2727
# Load the data
2828

2929
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
30-
gzip -d hits.tsv.gz
30+
gzip -d -f hits.tsv.gz
3131

3232
./apache-druid-${VERSION}/bin/post-index-task --file ingest.json --url http://localhost:8081
3333

duckdb-memory/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ pip install --break-system-packages duckdb==1.1.3 psutil
99
# Load the data
1010

1111
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
12-
gzip -d hits.csv.gz
12+
gzip -d -f hits.csv.gz
1313

1414
# Run the queries
1515

elasticsearch/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ curl -k -X PUT "https://localhost:9200/hits?pretty" -u "elastic:${PASSWORD}" -H
3838

3939
# Download and unzip dataset
4040
wget https://datasets.clickhouse.com/hits_compatible/hits.json.gz
41-
gzip -d hits.json.gz
41+
gzip -d -f hits.json.gz
4242

4343
# Prepare Elasticsearch for large bulk insert. To do the large upload, you have to break up JSON file into smaller files to prevent 'curl' from OOM while doing it, and adjust ELasticsearch HTTP upload size minimum. This creates roughly 250M files (note it takes a while)
4444
split -l 10000000 hits.json hits_

greenplum/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ sudo chmod 777 /gpmaster /gpdata1 /gpdata2 /gpdata3 /gpdata4 /gpdata5 /gpdata6 /
6464
gpinitsystem -ac gpinitsystem_singlenode
6565
export MASTER_DATA_DIRECTORY=/gpmaster/gpsne-1/
6666
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
67-
gzip -d hits.tsv.gz
67+
gzip -d -f hits.tsv.gz
6868
chmod 777 ~ hits.tsv
6969
psql -d postgres -f create.sql
7070
nohup gpfdist &

heavyai/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@ sudo systemctl enable heavydb
2828
# Load the data
2929

3030
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
31-
gzip -d hits.csv.gz
31+
gzip -d -f hits.csv.gz
3232
chmod 777 ~ hits.csv
3333

3434
sudo bash -c "echo 'allowed-import-paths = [\"/home/ubuntu/\"]' > /var/lib/heavyai/heavy.conf_"

hyper/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ sudo apt-get install -y python3-pip
55
pip install --break-system-packages tableauhyperapi
66

77
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
8-
gzip -d hits.csv.gz
8+
gzip -d -f hits.csv.gz
99

1010
./load.py
1111

infobright/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -14,7 +14,7 @@ sudo docker run -it --rm --network host mysql:5 mysql --host 127.0.0.1 --port 50
1414
# Load the data
1515

1616
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
17-
gzip -d hits.tsv.gz
17+
gzip -d -f hits.tsv.gz
1818

1919
# ERROR 2 (HY000) at line 1: Wrong data or column definition. Row: 93557187, field: 100.
2020
head -n 90000000 hits.tsv > hits90m.tsv

locustdb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ sudo apt-get install -y g++ capnproto libclang-14-dev
1515
cargo build --features "enable_rocksdb" --features "enable_lz4" --release
1616

1717
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
18-
gzip -d hits.csv.gz
18+
gzip -d -f hits.csv.gz
1919

2020
target/release/repl --load hits.csv --db-path db
2121

mariadb-columnstore/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ mysql --password="${PASSWORD}" --host 127.0.0.1 test < create.sql
1717
# Load the data
1818

1919
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
20-
gzip -d hits.tsv.gz
20+
gzip -d -f hits.tsv.gz
2121

2222
time mysql --password="${PASSWORD}" --host 127.0.0.1 test -e "
2323
LOAD DATA LOCAL INFILE 'hits.tsv' INTO TABLE hits

mariadb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ sudo service mariadb restart
1515
# Load the data
1616

1717
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
18-
gzip -d hits.tsv.gz
18+
gzip -d -f hits.tsv.gz
1919

2020
sudo mariadb -e "CREATE DATABASE test"
2121
sudo mariadb test < create.sql

monetdb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -23,7 +23,7 @@ sudo apt-get install -y expect
2323
./query.expect "$(cat create.sql)"
2424

2525
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
26-
gzip -d hits.tsv.gz
26+
gzip -d -f hits.tsv.gz
2727
chmod 777 ~ hits.tsv
2828

2929
./query.expect "COPY INTO hits FROM '$(pwd)/hits.tsv' USING DELIMITERS '\t'"

mongodb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -56,7 +56,7 @@ time mongosh --quiet --eval 'db.hits.createIndex({"ClientIP": 1, "WatchID": 1, "
5656
#################################
5757
# Load data and import
5858
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
59-
gzip -d hits.tsv.gz
59+
gzip -d -f hits.tsv.gz
6060

6161
# Use mongo import to load data into mongo. By default numInsertionWorkers is 1 so change to half of VM where it would be run
6262
#time mongoimport --collection hits --type tsv hits.tsv --fieldFile=create.txt --columnsHaveTypes --numInsertionWorkers=8

mysql-myisam/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ sudo service mysql restart
1010
# Load the data
1111

1212
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
13-
gzip -d hits.tsv.gz
13+
gzip -d -f hits.tsv.gz
1414

1515
sudo mysql -e "CREATE DATABASE test"
1616
sudo mysql test < create.sql

mysql/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -10,7 +10,7 @@ sudo service mysql restart
1010
# Load the data
1111

1212
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
13-
gzip -d hits.tsv.gz
13+
gzip -d -f hits.tsv.gz
1414

1515
sudo mysql -e "CREATE DATABASE test"
1616
sudo mysql test < create.sql

oxla/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ sudo DEBIAN_FRONTEND=noninteractive apt-get install -y build-essential
1111
echo "Download dataset."
1212
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
1313
echo "Unpack dataset."
14-
gzip -d hits.csv.gz
14+
gzip -d -f hits.csv.gz
1515
mkdir data
1616
mv hits.csv data
1717

pg_duckdb-indexed/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ set -eux
77
#sudo apt-get install -y postgresql-client
88

99
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
10-
gzip -d hits.tsv.gz
10+
gzip -d -f hits.tsv.gz
1111

1212
memory=$(awk '/MemTotal/ {print $2}' /proc/meminfo)
1313
threads=$(nproc)

pg_duckdb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ set -eux
77
#sudo apt-get install -y postgresql-client
88

99
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
10-
gzip -d hits.tsv.gz
10+
gzip -d -f hits.tsv.gz
1111

1212
memory=$(awk '/MemTotal/ {print $2}' /proc/meminfo)
1313
threads=$(nproc)

pgpro_tam/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -45,7 +45,7 @@ fi
4545
psql -h 127.0.0.1 -U postgres -t < create/"$CREATE_FILE".sql
4646

4747
#get and unpack hits.tsv
48-
sudo docker exec pgpro_tam bash -c "cd /tmp && wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz' && gzip -d hits.tsv.gz"
48+
sudo docker exec pgpro_tam bash -c "cd /tmp && wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz' && gzip -d -f hits.tsv.gz"
4949

5050
#insert data to table
5151
if [ "$1" == "parquet_fd_parall" ] ; then

pinot/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ sleep 30
1818
# Load the data
1919

2020
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
21-
gzip -d hits.tsv.gz
21+
gzip -d -f hits.tsv.gz
2222

2323
# Pinot was unable to load data as a single file wihout any errors returned. We have to split the data
2424
split -d --additional-suffix .tsv --verbose -n l/100 hits.tsv parts

postgresql-indexed/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ EOF
5252
sudo systemctl restart postgresql@$PGVERSION-main
5353

5454
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
55-
gzip -d hits.tsv.gz
55+
gzip -d -f hits.tsv.gz
5656

5757
sudo -u postgres psql -t -c 'CREATE DATABASE test'
5858
sudo -u postgres psql test -t <create.sql

postgresql/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ EOF
5252
sudo systemctl restart postgresql@$PGVERSION-main
5353

5454
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
55-
gzip -d hits.tsv.gz
55+
gzip -d -f hits.tsv.gz
5656

5757
sudo -u postgres psql -t -c 'CREATE DATABASE test'
5858
sudo -u postgres psql test -t <create.sql

questdb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ questdb/bin/questdb.sh start
1818
# Import the data
1919

2020
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
21-
gzip -d hits.csv.gz
21+
gzip -d -f hits.csv.gz
2222

2323
curl -G --data-urlencode "query=$(cat create.sql)" 'http://localhost:9000/exec'
2424

selectdb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -90,7 +90,7 @@ mysql -h 127.0.0.1 -P9030 -uroot hits <"$ROOT"/create.sql
9090
# Download data
9191
if [[ ! -f hits.tsv.gz ]] && [[ ! -f hits.tsv ]]; then
9292
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
93-
gzip -d hits.tsv.gz
93+
gzip -d -f hits.tsv.gz
9494
fi
9595

9696
# Load data

siglens/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -15,7 +15,7 @@ cd ..
1515

1616
echo "Download and unzip dataset"
1717
wget --continue https://datasets.clickhouse.com/hits_compatible/hits.json.gz
18-
gzip -d hits.json.gz
18+
gzip -d -f hits.json.gz
1919

2020
# Add the _index line and fix the UserID from string to num and preprocesses the dataset for loading
2121
python3 fix_hits.py

singlestore/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ sudo docker exec -it memsql-ciab memsql -p"${ROOT_PASSWORD}"
2222
# Load the data
2323

2424
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
25-
gzip -d hits.tsv.gz
25+
gzip -d -f hits.tsv.gz
2626
sudo docker cp hits.tsv memsql-ciab:/
2727

2828
sudo docker exec -it memsql-ciab memsql -p"${ROOT_PASSWORD}" -e "CREATE DATABASE test"

sqlite/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -6,7 +6,7 @@ sudo apt-get install -y sqlite3
66
sqlite3 mydb < create.sql
77

88
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.csv.gz'
9-
gzip -d hits.csv.gz
9+
gzip -d -f hits.csv.gz
1010

1111
time sqlite3 mydb '.import --csv hits.csv hits'
1212
wc -c mydb

starrocks/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ sleep 30
3939
# Prepare Data
4040
cd ../
4141
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
42-
gzip -d hits.tsv.gz
42+
gzip -d -f hits.tsv.gz
4343

4444
# Create Table
4545
mysql -h 127.0.0.1 -P9030 -uroot -e "CREATE DATABASE hits"

tablespace/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ sudo apt-get update
77
sudo apt-get install -y postgresql-client
88

99
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
10-
gzip -d hits.tsv.gz
10+
gzip -d -f hits.tsv.gz
1111
chmod 777 ~ hits.tsv
1212

1313
psql "host=$HOSTNAME port=5432 dbname=csdb user=csuser password=$PASSWORD sslmode=require" < create.sql

tembo-olap/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ sudo apt-get update
77
sudo apt-get install -y postgresql-client
88

99
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
10-
gzip -d hits.tsv.gz
10+
gzip -d -f hits.tsv.gz
1111
chmod 777 ~ hits.tsv
1212

1313
psql postgresql://postgres:$PASSWORD@$HOSTNAME:5432 -t -c 'CREATE DATABASE test'

timescale-cloud/README.md

Lines changed: 1 addition & 1 deletion

timescaledb-no-columnstore/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@ sudo -u postgres psql -c "CREATE DATABASE nocolumnstore"
1717
sudo -u postgres psql nocolumnstore -c "CREATE EXTENSION timescaledb WITH VERSION '2.17.2';"
1818

1919
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
20-
gzip -d hits.tsv.gz
20+
gzip -d -f hits.tsv.gz
2121
sudo chmod og+rX ~
2222
chmod 777 hits.tsv
2323

timescaledb/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ sudo -u postgres psql test -c "CREATE EXTENSION timescaledb WITH VERSION '2.17.2
1818

1919
# Import the data
2020
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
21-
gzip -d hits.tsv.gz
21+
gzip -d -f hits.tsv.gz
2222
sudo chmod og+rX ~
2323
chmod 777 hits.tsv
2424

umbra/benchmark.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@ sudo apt-get install -y postgresql-client gzip
99

1010
rm -rf hits.tsv
1111
wget --continue 'https://datasets.clickhouse.com/hits_compatible/hits.tsv.gz'
12-
gzip -d hits.tsv.gz
12+
gzip -d -f hits.tsv.gz
1313
chmod 777 hits.tsv
1414

1515
rm -rf umbra-25-01-23.tar.xz umbra

0 commit comments

Comments
 (0)