Replies: 5 comments 24 replies
-
@myc724 Sorry for the regression. There are hundreds of changes between these two, can you try other releases between then to narrow down the range? |
Beta Was this translation helpful? Give feedback.
-
I found the issue with curl. I need to quotes and feed curl. curl "localhost:6061/debug/pprof/goroutine?debug=1"
|
Beta Was this translation helpful? Give feedback.
-
I search all these +1000 machines' juicefs log and didn't find any ERROR or FATAL. Also, can you elaborate what this change trying to achieve? Thanks! |
Beta Was this translation helpful? Give feedback.
-
These're from a worker machine which did not respond for a very long hours. The lines of trace seems different. curl "localhost:6061/debug/pprof/goroutine?debug=1"
|
Beta Was this translation helpful? Give feedback.
-
curl "localhost:9567/debug/pprof/goroutine?debug=1"
|
Beta Was this translation helpful? Give feedback.
-
We want to switch JuiceFS 0.16 to deploy TiKV. However, we have a test which deployed 1250 machines to read and compile 140GB of source code (1.2 million files) then write 73GB object files (114 thousand files) in parallel. After switching JuiceFS from 0.11.0-37 to 0.16 with identical format/mount configuration (eg. --block-size, --cache-size, etc.), the overall runtime dropped by a lot. Our study indicated that many JuiceFS client operations timeout (ie. read/write failures) forced our compiler processes to find different hosts to recompile the failure components. We’ve tried various knobs (eg, --get-timeout, --put-timeout, --io-retries, etc. ) but none of them seems to help.
We noticed that, for the format command, JuiceFS 0.11.0-37 default “--compress lz4” while JuiceFS 0.16 default “--compress none”. After applying lz4 compression, we found that, if we format a volume using the JuiceFS 0.16 but mount it to all machines using JuiceFS 0.11.0-37, this test can match the performance of the earlier and fastest run which fully deployed JuiceFS 0.11.0-37. We studied the mount command source code but didn’t find any obvious default setup change. Is there any default configuration being adjusted for the mount command from JuiceFS 0.11.0-37 to JuiceFS 0.16? Just want to see whether we can try to revert it to match for this test. Thanks!
the original run
juicefs version 0.11.0-37 ( 6127863)
the new run
juicefs version 0.16-dev (2021-08-02 5d93e58)
both runs using the same Redis for metadata as well as object-storage
Redis server v=6.2.2 sha=00000000:0 malloc=jemalloc-5.1.0 bits=64 build=3e62cb2e1fe34d7f
Beta Was this translation helpful? Give feedback.
All reactions