Replies: 8 comments 3 replies
-
Tested download with current command which does not kernel panic on upload:rclone --transfers 5 --use-mmap --buffer-size 0M sync mys3:san-is-testing-here/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/workspace/ test/ --links --progress --multi-thread-streams 1 15 minutes!!! |
Beta Was this translation helpful? Give feedback.
-
tested download using 1.63.1 removing
|
Beta Was this translation helpful? Give feedback.
-
RClone vs AWS Cli manual test after implementationTest with 1000 files with 100 MB size + 2 files with 10GB size
|
Beta Was this translation helpful? Give feedback.
-
After @matusdrobuliak66 comment above I did yet another trial: 5th trial 120.8 GBrclone 1.63.1
rclone sync mys3:sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/workspace/ test/ --progress --transfers 5 --use-mmap --buffer-size 0M --links --verbose
rclone sync test/ mys3:san-testing-again/aws-rclone/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --progress --transfers 5 --use-mmap --buffer-size 0M --links --verbose
AWS CLI S3 sync
time aws s3 sync s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ test-aws/ --delete
time aws s3 sync test-aws/ s3://san-testing-again/aws-testing/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --delete
Summary
--> we need to understand why the improvement is not visible in oSparc |
Beta Was this translation helpful? Give feedback.
-
here is yet another alternative: https://github.com/peak/s5cmd early trials showed the same as aws-cli on the tested dataset for download: 16m29.387s |
Beta Was this translation helpful? Give feedback.
-
g4dn.4xlarge with EBS disk at 1000Mbips transfer speed + 6000IOPS AWS CLI
time aws s3 sync s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ test-aws/ --delete
upload to s3://san-testing-again/aws-testing/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/
time aws s3 sync test-aws/ s3://san-testing-again/aws-testing/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --delete
tweaked AWS CLIfrom https://stackoverflow.com/questions/9664904/best-way-to-move-files-between-s3-buckets aws configure set default.s3.max_concurrent_requests 1000
aws configure set default.s3.max_queue_size 100000 "(The following commands will tell the AWS CLI to use 1,000 threads to execute jobs (each a small file or one part of a multipart copy) and look ahead 100,000 jobs)" command: time aws s3 sync s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ test-aws/ --delete
time aws s3 sync test-aws/ s3://san-testing-again/aws-testing-tweaked/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --delete
RClone 1.63.1
rclone sync mys3:sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/workspace/ test-rclone/ --progress --transfers 5 --use-mmap --buffer-size 0M --links --verbose
rclone sync test-rclone/ mys3:san-testing-again/aws-rclone/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --progress --transfers 5 --use-mmap --buffer-size 0M --links --verbose completed after 8m44.3s tweaked RClone 1.63.1removed checksum, mmap and buffer-size. rclone sync mys3:sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/workspace/ test-rclone/ --progress --transfers 5 --links --verbose --ignore-checksum
upload to san-testing-again/aws-rclone/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/
rclone sync test-rclone/ mys3:san-testing-again/aws-rclone-tweaked/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --progress --transfers 5 --links --verbose --ignore-checksum completed after 8m30.6s S5cmddownload from s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/
wget https://github.com/peak/s5cmd/releases/download/v2.2.2/s5cmd_2.2.2_Linux-64bit.tar.gz .
tar -xvf s5cmd_2.2.2_Linux-64bit.tar.gz
./s5cmd version
time ./s5cmd --stat --json sync --delete "s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/*" test-s5cmd/
time ./s5cmd --stat --json sync --delete test-s5cmd/ s3://san-testing-again/aws-s5cmd/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/
Summary
AWS-cli v2The AMI had an old version of the CLI curl "https://awscli.amazonaws.com/awscli-exe-linux-x86_64.zip" -o "awscliv2.zip"
unzip awscliv2.zip
sudo ./aws/install --bin-dir /usr/local/bin --install-dir /usr/local/aws-cli --update
/usr/local/bin/aws --version # shows now aws-cli/2.17.25 Python/3.11.9 Linux/6.5.0-1022-aws exe/x86_64.ubuntu.22
aws --version # shows aws-cli/1.22.34 Python/3.10.12 Linux/6.5.0-1022-aws botocore/1.23.34 very old!
time /usr/local/bin/aws s3 sync s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ test-aws-v2/ --delete
upload to s3://san-testing-again/aws-testing/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/
time /usr/local/bin/aws s3 sync test-aws-v2/ s3://san-testing-again/aws-testing-v2/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --delete
tweaked AWS CLI v2from https://stackoverflow.com/questions/9664904/best-way-to-move-files-between-s3-buckets aws configure set default.s3.max_concurrent_requests 1000 "(The following commands will tell the AWS CLI to use 1,000 threads to execute jobs (each a small file or one part of a multipart copy) command: time /usr/local/bin/aws s3 sync s3://sim4life-simcore/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ test-aws-v2-tweaked/ --delete
time /usr/local/bin/aws s3 sync test-aws-v2-tweaked/ s3://san-testing-again/aws-testing-v2-tweaked/72a6031e-b9d5-11ee-aeb0-02420a0bcae9/8b5290bb-7ded-409a-bdd5-4c95250b12cd/ --delete
|
Beta Was this translation helpful? Give feedback.
-
ResultsSAN original tests on bare metal (g4dn.xlarge)
MD test on oSparc after AWS-CLI PR came in master/stagingTest with 1000 files with 100 MB size + 2 files with 10GB size
SAN subsequent tests on bare metal (g4dn.xlarge)Same Dataset as done earlier but is now 120GB
SAN subsequent tests on bare metal (g4dn.4xlarge) with GP3 1000MiB/s - 6000IOPSSame dataset as done earlier but is now 120GB
GP3 disk costs
Conclusion
|
Beta Was this translation helpful? Give feedback.
-
RClone session
Created a EC2 g4dn.xlarge from the latest AMI without disk of 04.05.2024. then installed RClone using the script
sudo su
sudo apt install btop # to show what is going on while transferring
1.63.1
1.66.0
3rd trial with rclone 1.66.0
download
--> no changes
2. removing --mmap https://forum.rclone.org/t/what-is-the-impact-of-use-mmap-besides-reducing-memory-usage/40677/8
--> no changes
3.
--> 3x momentarily speed up, total time 4m43
upload
test bandwith
test aws CLI
download 3min13
2. with aws configure set default.s3.max_concurrent_requests 16 almot no changes
upload
3min38
Conclusion thus far:
Beta Was this translation helpful? Give feedback.
All reactions