Skip to content

Commit 3895781

Browse files
committed
Merge branch 'main' into add_new_machine_type
2 parents 35e7877 + 79fe6a4 commit 3895781

File tree

30 files changed

+266
-45
lines changed

30 files changed

+266
-45
lines changed

.github/workflows/regression.yml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ jobs:
4848

4949
- name: Install all solutions
5050
shell: bash
51-
run: source path.env && python3 ./_setup_utils/install_all_solutions.py ${{ matrix.solution }}
51+
run: source path.env && python3 _setup_utils/install_all_solutions.py ${{ matrix.solution }}
5252

5353
- name: Turn swap off
5454
shell: bash

R-arrow/VERSION

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
17.0.0.1

_benchplot/benchplot-dict.R

Lines changed: 8 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@ solution.dict = {list(
3939
"spark" = list(name=c(short="spark", long="spark"), color=c(strong="#8000FFFF", light="#CC66FF")),
4040
"dask" = list(name=c(short="dask", long="dask"), color=c(strong="slategrey", light="lightgrey")),
4141
"juliadf" = list(name=c(short="DF.jl", long="DataFrames.jl"), color=c(strong="deepskyblue", light="darkturquoise")),
42-
"juliads" = list(name=c(short="IMD.jl", long="InMemoryDatasets.jl"), color=c(strong="#b80000", light="#ff1f1f")),
42+
"juliads" = list(name=c(short="IMD.jl", long="InMemData.jl"), color=c(strong="#b80000", light="#ff1f1f")),
4343
"clickhouse" = list(name=c(short="clickhouse", long="ClickHouse"), color=c(strong="hotpink4", light="hotpink1")),
4444
"polars" = list(name=c(short="polars", long="Polars"), color=c(strong="deepskyblue4", light="deepskyblue3")),
4545
"R-arrow" = list(name=c(short="R-arrow", long="R-arrow"), color=c(strong="aquamarine3", light="aquamarine1")),
@@ -299,10 +299,12 @@ groupby.data.exceptions = {list(
299299
"G1_1e8_2e0_0_0") # q3
300300
)},
301301
"juliadf" = {list(
302-
"timeout" = "G1_1e8_2e0_0_0",
303-
"out of memory" = c("G1_1e9_1e2_0_0","G1_1e9_1e1_0_0","G1_1e9_2e0_0_0","G1_1e9_1e2_0_1","G1_1e9_1e2_5_0") # CSV.File
302+
# "timeout" = "G1_1e8_2e0_0_0",
303+
# "out of memory" = c("G1_1e9_1e2_0_0","G1_1e9_1e1_0_0","G1_1e9_2e0_0_0","G1_1e9_1e2_0_1","G1_1e9_1e2_5_0"), # CSV.File
304+
"CSV import Segfault: JuliaLang#55765" = c("G1_1e7_1e2_0_0","G1_1e7_1e1_0_0","G1_1e7_2e0_0_0","G1_1e7_1e2_0_1","G1_1e7_1e2_5_0","G1_1e8_1e2_0_0","G1_1e8_1e1_0_0","G1_1e8_2e0_0_0","G1_1e8_1e2_0_1","G1_1e8_1e2_5_0","G1_1e9_1e2_0_0","G1_1e9_1e1_0_0","G1_1e9_2e0_0_0","G1_1e9_1e2_0_1","G1_1e9_1e2_5_0")
304305
)},
305306
"juliads" = {list(
307+
"CSV import Segfault: JuliaLang#55765" = c("G1_1e7_1e2_0_0","G1_1e7_1e1_0_0","G1_1e7_2e0_0_0","G1_1e7_1e2_0_1","G1_1e7_1e2_5_0","G1_1e8_1e2_0_0","G1_1e8_1e1_0_0","G1_1e8_2e0_0_0","G1_1e8_1e2_0_1","G1_1e8_1e2_5_0","G1_1e9_1e2_0_0","G1_1e9_1e1_0_0","G1_1e9_2e0_0_0","G1_1e9_1e2_0_1","G1_1e9_1e2_5_0")
306308
)},
307309
"clickhouse" = {list(
308310
)},
@@ -485,9 +487,11 @@ join.data.exceptions = {list(
485487
"out of memory" = c("J1_1e9_NA_0_0") # q1 even when using on-disk, after 47m (480m timeout)
486488
)},
487489
"juliadf" = {list(
488-
"out of memory" = c("J1_1e9_NA_0_0","J1_1e9_NA_5_0","J1_1e9_NA_0_1") # CSV.File
490+
# "out of memory" = c("J1_1e9_NA_0_0","J1_1e9_NA_5_0","J1_1e9_NA_0_1") # CSV.File
491+
"CSV import Segfault: JuliaLang#55765" = c("J1_1e7_NA_0_0", "J1_1e7_NA_5_0", "J1_1e7_NA_0_1", "J1_1e8_NA_0_0", "J1_1e8_NA_5_0", "J1_1e8_NA_0_1", "J1_1e9_NA_0_0")
489492
)},
490493
"juliads" = {list(
494+
"CSV import Segfault: JuliaLang#55765" = c("J1_1e7_NA_0_0", "J1_1e7_NA_5_0", "J1_1e7_NA_0_1", "J1_1e8_NA_0_0", "J1_1e8_NA_5_0", "J1_1e8_NA_0_1", "J1_1e9_NA_0_0")
491495
)},
492496
"clickhouse" = {list(
493497
)},

_report/index.Rmd

Lines changed: 27 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -217,6 +217,33 @@ loop_benchplot(dt_join, report_name="join", syntax.dict=join.syntax.dict, except
217217

218218
<!--
219219
##### **advanced questions**
220+
221+
![](./join/J1_1e7_NA_0_0_advanced.png)
222+
-->
223+
224+
#### 5 GB
225+
226+
##### **basic questions**
227+
228+
![](./join/J1_1e8_NA_0_0_basic.png)
229+
230+
231+
<!--
232+
##### **advanced questions**
233+
234+
![](./join/J1_1e8_NA_0_0_advanced.png)
235+
-->
236+
237+
#### 50 GB {.active}
238+
239+
##### **basic questions**
240+
241+
![](./join/J1_1e9_NA_0_0_basic.png)
242+
243+
<!--
244+
##### **advanced questions**
245+
246+
![](./join/J1_1e9_NA_0_0_advanced.png)
220247
-->
221248

222249
---

_run/partitioned_run.sh

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,5 @@
11
# set machine type
2+
<<<<<<< HEAD
23
./_run/run_small_medium.sh
34

4-
./_run/run_large.sh
5+
./_run/run_large.sh

_run/run_large_groupby_join.sh

Lines changed: 31 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,31 @@
1+
# download and expand large data
2+
3+
# get groupby large (0.5GB and 5GB datasets)
4+
aws s3 cp s3://duckdb-blobs/data/db-benchmark-data/groupby_large.duckdb data/groupby_large.duckdb
5+
# get join small (0.5GB and 5GB datasets)
6+
aws s3 cp s3://duckdb-blobs/data/db-benchmark-data/join_large.duckdb data/join_large.duckdb
7+
8+
9+
# expand groupby-small datasets to csv
10+
duckdb data/groupby_large.duckdb -c "copy G1_1e9_1e2_0_0 to 'data/G1_1e9_1e2_0_0.csv' (FORMAT CSV)"
11+
duckdb data/groupby_large.duckdb -c "copy G1_1e9_1e1_0_0 to 'data/G1_1e9_1e1_0_0.csv' (FORMAT CSV)"
12+
duckdb data/groupby_large.duckdb -c "copy G1_1e9_2e0_0_0 to 'data/G1_1e9_2e0_0_0.csv' (FORMAT CSV)"
13+
duckdb data/groupby_large.duckdb -c "copy G1_1e9_1e2_0_1 to 'data/G1_1e9_1e2_0_1.csv' (FORMAT CSV)"
14+
duckdb data/groupby_large.duckdb -c "copy G1_1e9_1e2_5_0 to 'data/G1_1e9_1e2_5_0.csv' (FORMAT CSV)"
15+
16+
# expand join-small datasets to csv
17+
duckdb data/join_large.duckdb -c "copy J1_1e9_NA_0_0 to 'data/J1_1e9_NA_0_0.csv' (FORMAT CSV)"
18+
duckdb data/join_large.duckdb -c "copy J1_1e9_1e9_0_0 to 'data/J1_1e9_1e9_0_0.csv' (FORMAT CSV)"
19+
duckdb data/join_large.duckdb -c "copy J1_1e9_1e6_0_0 to 'data/J1_1e9_1e6_0_0.csv' (FORMAT CSV)"
20+
duckdb data/join_large.duckdb -c "copy J1_1e9_1e3_0_0 to 'data/J1_1e9_1e3_0_0.csv' (FORMAT CSV)"
21+
22+
23+
echo "Running all solutions on large (50GB) datasets"
24+
./run.sh
25+
26+
27+
###
28+
echo "done..."
29+
echo "removing data files"
30+
#rm data/*.csv
31+
#rm data/*.duckdb

_run/run_small_medium_groupby_join.sh

Lines changed: 59 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,59 @@
1+
# first download and expand small data
2+
3+
# get groupby small (0.5GB and 5GB datasets)
4+
aws s3 cp s3://duckdb-blobs/data/db-benchmark-data/groupby_small.duckdb data/groupby_small.duckdb
5+
# get join small (0.5GB and 5GB datasets)
6+
aws s3 cp s3://duckdb-blobs/data/db-benchmark-data/join_small.duckdb data/join_small.duckdb
7+
8+
9+
# expand groupby-small datasets to csv
10+
duckdb data/groupby_small.duckdb -c "copy G1_1e7_1e2_0_0 to 'data/G1_1e7_1e2_0_0.csv' (FORMAT CSV)"
11+
duckdb data/groupby_small.duckdb -c "copy G1_1e7_1e1_0_0 to 'data/G1_1e7_1e1_0_0.csv' (FORMAT CSV)"
12+
duckdb data/groupby_small.duckdb -c "copy G1_1e7_2e0_0_0 to 'data/G1_1e7_2e0_0_0.csv' (FORMAT CSV)"
13+
duckdb data/groupby_small.duckdb -c "copy G1_1e7_1e2_0_1 to 'data/G1_1e7_1e2_0_1.csv' (FORMAT CSV)"
14+
duckdb data/groupby_small.duckdb -c "copy G1_1e7_1e2_5_0 to 'data/G1_1e7_1e2_5_0.csv' (FORMAT CSV)"
15+
duckdb data/groupby_small.duckdb -c "copy G1_1e8_1e2_0_0 to 'data/G1_1e8_1e2_0_0.csv' (FORMAT CSV)"
16+
duckdb data/groupby_small.duckdb -c "copy G1_1e8_1e1_0_0 to 'data/G1_1e8_1e1_0_0.csv' (FORMAT CSV)"
17+
duckdb data/groupby_small.duckdb -c "copy G1_1e8_2e0_0_0 to 'data/G1_1e8_2e0_0_0.csv' (FORMAT CSV)"
18+
duckdb data/groupby_small.duckdb -c "copy G1_1e8_1e2_0_1 to 'data/G1_1e8_1e2_0_1.csv' (FORMAT CSV)"
19+
duckdb data/groupby_small.duckdb -c "copy G1_1e8_1e2_5_0 to 'data/G1_1e8_1e2_5_0.csv' (FORMAT CSV)"
20+
21+
# expand join-small datasets to csv
22+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e1_0_0 to 'data/J1_1e7_1e1_0_0.csv' (FORMAT CSV)"
23+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e4_5_0 to 'data/J1_1e7_1e4_5_0.csv' (FORMAT CSV)"
24+
duckdb data/join_small.duckdb -c "copy J1_1e7_NA_0_1 to 'data/J1_1e7_NA_0_1.csv' (FORMAT CSV)"
25+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e5_0_0 to 'data/J1_1e8_1e5_0_0.csv' (FORMAT CSV)"
26+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e8_5_0 to 'data/J1_1e8_1e8_5_0.csv' (FORMAT CSV)"
27+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e1_0_1 to 'data/J1_1e7_1e1_0_1.csv' (FORMAT CSV)"
28+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e7_0_0 to 'data/J1_1e7_1e7_0_0.csv' (FORMAT CSV)"
29+
duckdb data/join_small.duckdb -c "copy J1_1e7_NA_5_0 to 'data/J1_1e7_NA_5_0.csv' (FORMAT CSV)"
30+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e5_0_1 to 'data/J1_1e8_1e5_0_1.csv' (FORMAT CSV)"
31+
duckdb data/join_small.duckdb -c "copy J1_1e8_NA_0_0 to 'data/J1_1e8_NA_0_0.csv' (FORMAT CSV)"
32+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e1_5_0 to 'data/J1_1e7_1e1_5_0.csv' (FORMAT CSV)"
33+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e7_0_1 to 'data/J1_1e7_1e7_0_1.csv' (FORMAT CSV)"
34+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e2_0_0 to 'data/J1_1e8_1e2_0_0.csv' (FORMAT CSV)"
35+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e5_5_0 to 'data/J1_1e8_1e5_5_0.csv' (FORMAT CSV)"
36+
duckdb data/join_small.duckdb -c "copy J1_1e8_NA_0_1 to 'data/J1_1e8_NA_0_1.csv' (FORMAT CSV)"
37+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e4_0_0 to 'data/J1_1e7_1e4_0_0.csv' (FORMAT CSV)"
38+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e7_5_0 to 'data/J1_1e7_1e7_5_0.csv' (FORMAT CSV)"
39+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e2_0_1 to 'data/J1_1e8_1e2_0_1.csv' (FORMAT CSV)"
40+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e8_0_0 to 'data/J1_1e8_1e8_0_0.csv' (FORMAT CSV)"
41+
duckdb data/join_small.duckdb -c "copy J1_1e8_NA_5_0 to 'data/J1_1e8_NA_5_0.csv' (FORMAT CSV)"
42+
duckdb data/join_small.duckdb -c "copy J1_1e7_1e4_0_1 to 'data/J1_1e7_1e4_0_1.csv' (FORMAT CSV)"
43+
duckdb data/join_small.duckdb -c "copy J1_1e7_NA_0_0 to 'data/J1_1e7_NA_0_0.csv' (FORMAT CSV)"
44+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e2_5_0 to 'data/J1_1e8_1e2_5_0.csv' (FORMAT CSV)"
45+
duckdb data/join_small.duckdb -c "copy J1_1e8_1e8_0_1 to 'data/J1_1e8_1e8_0_1.csv' (FORMAT CSV)"
46+
47+
48+
cp _control/data_small.csv _control/data.csv
49+
50+
51+
echo "Running all solutions on small (0.5GB and 5GB) datasets"
52+
./run.sh
53+
54+
55+
###
56+
echo "done..."
57+
echo "removing small data files"
58+
rm data/*.csv
59+
rm data/*.duckdb

_setup_utils/format_and_mount.sh

Lines changed: 90 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,90 @@
1+
# script to format mount and copy data.
2+
3+
# remove a leftover instance mount
4+
rm -rf ~/db-benchmark-metal
5+
6+
# format the mount
7+
sudo mkfs -t xfs /dev/nvme0n1
8+
9+
mkdir ~/db-benchmark-metal
10+
# mount the nvme volumn
11+
sudo mount /dev/nvme0n1 ~/db-benchmark-metal
12+
# change ownsership of the volume
13+
sudo chown -R ubuntu ~/db-benchmark-metal/
14+
15+
git clone https://github.com/duckdblabs/db-benchmark.git ~/db-benchmark-metal
16+
17+
# if you have an EBS volume, you can generate the data once, save it on the ebs volume, and transfer it
18+
# each time.
19+
20+
if [[ $# -gt 0 ]]
21+
then
22+
echo "Creating data"
23+
mkdir -p ~/db-benchmark-metal/data/
24+
cd ~/db-benchmark-metal/data/
25+
echo "Creating 500mb group by datasets"
26+
Rscript ../_data/groupby-datagen.R 1e7 1e2 0 0
27+
Rscript ../_data/groupby-datagen.R 1e7 1e1 0 0
28+
Rscript ../_data/groupby-datagen.R 1e7 2e0 0 0
29+
Rscript ../_data/groupby-datagen.R 1e7 1e2 0 1
30+
Rscript ../_data/groupby-datagen.R 1e7 1e2 5 0
31+
echo "Creating 5gb group by datasets"
32+
Rscript ../_data/groupby-datagen.R 1e8 1e2 0 0
33+
Rscript ../_data/groupby-datagen.R 1e8 1e1 0 0
34+
Rscript ../_data/groupby-datagen.R 1e8 2e0 0 0
35+
Rscript ../_data/groupby-datagen.R 1e8 1e2 0 1
36+
Rscript ../_data/groupby-datagen.R 1e8 1e2 5 0
37+
echo "Creating 50gb group by datasets"
38+
Rscript ../_data/groupby-datagen.R 1e9 1e2 0 0
39+
Rscript ../_data/groupby-datagen.R 1e9 1e1 0 0
40+
Rscript ../_data/groupby-datagen.R 1e9 2e0 0 0
41+
Rscript ../_data/groupby-datagen.R 1e9 1e2 0 1
42+
Rscript ../_data/groupby-datagen.R 1e9 1e2 5 0
43+
echo "Creating 500mb join datasets"
44+
Rscript ../_data/join-datagen.R 1e7 0 0
45+
Rscript ../_data/join-datagen.R 1e7 5 0
46+
Rscript ../_data/join-datagen.R 1e7 0 1
47+
echo "Creating 5gb join datasets"
48+
Rscript ../_data/join-datagen.R 1e8 0 0
49+
Rscript ../_data/join-datagen.R 1e8 5 0
50+
Rscript ../_data/join-datagen.R 1e8 0 1
51+
echo "Creating 50gb join datasets"
52+
Rscript ../_data/join-datagen.R 1e9 0 0
53+
cd ..
54+
elif [[ ! -d "~/db-benchark/data" ]]
55+
then
56+
echo "no arguments passed. Copying data..."
57+
echo "ERROR: directory ~/db-benchmark/data does not exist"
58+
else
59+
mkdir -p ~/db-benchmark-metal/data/
60+
cd ~/db-benchmark-metal/data/
61+
echo "Copying data from ~/db-benchark/data"
62+
cp ~/db-benchmark/data/*.csv
63+
cd ~/db-benchmark-metal
64+
fi
65+
66+
67+
./_launcher/setup.sh
68+
69+
# setup all the solutions on db-benchmark-metal.
70+
# creates the necessary python virtual environments and creates the r-libraries
71+
# needed
72+
source path.env && python3 _utils/install_all_solutions.py all
73+
74+
75+
# setup mount for clickhouse spill
76+
sudo mkfs -t xfs /dev/nvme1n1
77+
sudo mkdir /var/lib/clickhouse-nvme-mount/
78+
sudo mount /dev/nvme1n1 /var/lib/clickhouse-nvme-mount/
79+
# not sure if below is necessary.
80+
sudo cp -a /var/lib/clickhouse/. /var/lib/clickhouse-nvme-mount/
81+
# change ownership of new mount to clickhouse
82+
sudo chown -R clickhouse:clickhouse /var/lib/clickhouse-nvme-mount/
83+
sudo chown -R clickhouse:clickhouse /dev/nvme1n1
84+
85+
# add config so clickhouse knows to use the mount to spill data
86+
sudo cp clickhouse/clickhouse-mount-config.xml /etc/clickhouse-server/config.d/data-paths.xml
87+
88+
echo "------------------------------------------"
89+
echo "------------------------------------------"
90+
echo "READY TO RUN BENCHMARK. ./run.sh"
File renamed without changes.

_setup_utils/sleep_and_run.sh

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,9 @@
1+
while [ -f run.lock ]
2+
do
3+
sleep 1800
4+
done
5+
6+
7+
rm run.lock
8+
9+
./run.sh

clickhouse/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
23.10.4.25
1+
24.8.4.13

collapse/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.0.3
1+
2.0.16

dask/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2023.10.0
1+
2024.9.0

datafusion/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
31.0.0
1+
41.0.0

datatable/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
1.14.9
1+
1.16.99

dplyr/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
1.1.3
1+
1.1.4

duckdb-latest/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
0.9.1.1
1+
1.0.99.9000

duckdb/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
1.0.0
1+
1.1.0

duckdb/setup-duckdb.sh

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -11,7 +11,7 @@ Rscript -e 'install.packages("DBI", lib="./duckdb/r-duckdb", repos = "http://clo
1111
cd duckdb
1212
git clone https://github.com/duckdb/duckdb-r.git
1313
cd duckdb-r
14-
git checkout v1.0.0
14+
git checkout v1.1.0
1515
cd ..
1616
ncores=`python3 -c 'import multiprocessing as mp; print(mp.cpu_count())'`
1717
MAKE="make -j$ncores" R CMD INSTALL -l "./r-duckdb" duckdb-r

juliadf/setup-juliadf.sh

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,11 +1,11 @@
11
# install julia
22

3-
wget https://julialang-s3.julialang.org/bin/linux/x64/1.10/julia-1.10.4-linux-x86_64.tar.gz
4-
tar -xvf julia-1.10.4-linux-x86_64.tar.gz
5-
sudo mv julia-1.10.4 /opt
6-
rm julia-1.10.4-linux-x86_64.tar.gz
3+
wget https://julialang-s3.julialang.org/bin/linux/x64/1.10/julia-1.10.5-linux-x86_64.tar.gz
4+
tar -xvf julia-1.10.5-linux-x86_64.tar.gz
5+
sudo mv julia-1.10.5 /opt
6+
rm julia-1.10.5-linux-x86_64.tar.gz
77
# put to paths
8-
echo 'export JULIA_HOME=/opt/julia-1.10.4' >> path.env
8+
echo 'export JULIA_HOME=/opt/julia-1.10.5' >> path.env
99
echo 'export PATH=$PATH:$JULIA_HOME/bin' >> path.env
1010
# note that cron job must have path updated as well
1111

juliads/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
0.7.18
1+
0.7.21

juliads/setup-juliads.sh

Lines changed: 7 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -1,20 +1,18 @@
1-
21
# install julia
3-
wget https://julialang-s3.julialang.org/bin/linux/x64/1.10/julia-1.10.4-linux-x86_64.tar.gz
4-
tar -xvf julia-1.10.4-linux-x86_64.tar.gz
5-
sudo mv julia-1.10.4 /opt
6-
rm julia-1.10.4-linux-x86_64.tar.gz
72

3+
wget https://julialang-s3.julialang.org/bin/linux/x64/1.10/julia-1.10.5-linux-x86_64.tar.gz
4+
tar -xvf julia-1.10.5-linux-x86_64.tar.gz
5+
sudo mv julia-1.10.5 /opt
6+
rm julia-1.10.5-linux-x86_64.tar.gz
87
# put to paths
9-
echo 'export JULIA_HOME=/opt/julia-1.10.4' >> path.env
8+
echo 'export JULIA_HOME=/opt/julia-1.10.5' >> path.env
109
echo 'export PATH=$PATH:$JULIA_HOME/bin' >> path.env
11-
echo "export JULIA_NUM_THREADS=40" >> path.env
1210
# note that cron job must have path updated as well
1311

1412
source path.env
1513

1614
# install julia InMemoryDatasets and csv packages
17-
julia -q -e 'using Pkg; Pkg.add(["InMemoryDatasets","DLMReader", "PooledArrays", "Arrow"])'
15+
julia -q -e 'using Pkg; Pkg.add(["InMemoryDatasets","DLMReader", "PooledArrays", "Arrow", "CSV"])'
1816
julia -q -e 'include("$(pwd())/_helpers/helpersds.jl"); pkgmeta = getpkgmeta("InMemoryDatasets"); println(string(pkgmeta["version"])); pkgmeta = getpkgmeta("DLMReader"); println(string(pkgmeta["version"]))'
1917

20-
./juliadf/ver-juliads.sh
18+
./juliadf/ver-juliadf.sh

logs.csv

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1790,4 +1790,4 @@ ip-172-31-31-147,1720089751,duckdb,1.0.0,1f98600c2c,join,J1_1e8_NA_5_0,172009121
17901790
ip-172-31-31-147,1720089751,duckdb,1.0.0,1f98600c2c,join,J1_1e8_NA_0_1,1720091226.51143,start,,,xlarge
17911791
ip-172-31-31-147,1720089751,duckdb,1.0.0,1f98600c2c,join,J1_1e8_NA_0_1,1720091262.47306,finish,0,0,xlarge
17921792
ip-172-31-31-147,1720529276,duckdb,1.0.0,1f98600c2c,join,J1_1e9_NA_0_0,1720529277.33854,start,,,xlarge
1793-
ip-172-31-31-147,1720529276,duckdb,1.0.0,1f98600c2c,join,J1_1e9_NA_0_0,1720530147.30009,finish,0,0,xlarge
1793+
ip-172-31-31-147,1720529276,duckdb,1.0.0,1f98600c2c,join,J1_1e9_NA_0_0,1720530147.30009,finish,0,0,xlarge

pandas/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
2.1.1
1+
2.2.2

polars/VERSION

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
1.1.0
1+
1.8.2

0 commit comments

Comments
 (0)