==> Synchronizing chroot copy [/home/alhp/workspace/chroot/root] -> [build_cfdf6060-e862-4678-a9c1-13c90f055610]...done ==> Making package: seaweedfs 3.92-2.1 (Tue Jul 1 11:52:08 2025) ==> Retrieving sources... -> Downloading seaweedfs-3.92.tar.gz... % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 406k 0 406k 0 0 510k 0 --:--:-- --:--:-- --:--:-- 510k 100 3113k 0 3113k 0 0 1763k 0 --:--:-- 0:00:01 --:--:-- 2790k 100 5377k 0 5377k 0 0 1943k 0 --:--:-- 0:00:02 --:--:-- 2520k 100 7580k 0 7580k 0 0 2005k 0 --:--:-- 0:00:03 --:--:-- 2404k 100 9936k 0 9936k 0 0 2075k 0 --:--:-- 0:00:04 --:--:-- 2387k 100 10.8M 0 10.8M 0 0 1921k 0 --:--:-- 0:00:05 --:--:-- 2146k 100 11.7M 0 11.7M 0 0 1778k 0 --:--:-- 0:00:06 --:--:-- 1784k 100 12.5M 0 12.5M 0 0 1648k 0 --:--:-- 0:00:07 --:--:-- 1486k 100 13.5M 0 13.5M 0 0 1576k 0 --:--:-- 0:00:08 --:--:-- 1253k 100 14.4M 0 14.4M 0 0 1517k 0 --:--:-- 0:00:09 --:--:-- 982k 100 15.7M 0 15.7M 0 0 1499k 0 --:--:-- 0:00:10 --:--:-- 1010k 100 16.9M 0 16.9M 0 0 1475k 0 --:--:-- 0:00:11 --:--:-- 1066k 100 18.0M 0 18.0M 0 0 1447k 0 --:--:-- 0:00:12 --:--:-- 1134k 100 19.5M 0 19.5M 0 0 1449k 0 --:--:-- 0:00:13 --:--:-- 1225k 100 21.0M 0 21.0M 0 0 1457k 0 --:--:-- 0:00:14 --:--:-- 1340k 100 22.5M 0 22.5M 0 0 1465k 0 --:--:-- 0:00:15 --:--:-- 1392k 100 24.3M 0 24.3M 0 0 1481k 0 --:--:-- 0:00:16 --:--:-- 1496k 100 26.1M 0 26.1M 0 0 1508k 0 --:--:-- 0:00:17 --:--:-- 1662k 100 28.0M 0 28.0M 0 0 1531k 0 --:--:-- 0:00:18 --:--:-- 1756k 100 28.6M 0 28.6M 0 0 1535k 0 --:--:-- 0:00:19 --:--:-- 1803k ==> Validating source files with sha256sums... seaweedfs-3.92.tar.gz ... Passed ==> Making package: seaweedfs 3.92-2.1 (Tue Jul 1 09:52:29 2025) ==> Checking runtime dependencies... ==> Installing missing dependencies... resolving dependencies... looking for conflicting packages... Package (1) New Version Net Change extra/mailcap 2.1.54-2 0.11 MiB Total Installed Size: 0.11 MiB :: Proceed with installation? [Y/n] checking keyring... checking package integrity... loading package files... checking for file conflicts... :: Processing package changes... installing mailcap... :: Running post-transaction hooks... (1/1) Arming ConditionNeedsUpdate... ==> Checking buildtime dependencies... ==> Installing missing dependencies... resolving dependencies... looking for conflicting packages... Package (1) New Version Net Change extra/go 2:1.24.4-1 237.84 MiB Total Installed Size: 237.84 MiB :: Proceed with installation? [Y/n] checking keyring... checking package integrity... loading package files... checking for file conflicts... :: Processing package changes... installing go... :: Running post-transaction hooks... (1/1) Arming ConditionNeedsUpdate... ==> Retrieving sources... -> Found seaweedfs-3.92.tar.gz ==> WARNING: Skipping all source file integrity checks. ==> Extracting sources... -> Extracting seaweedfs-3.92.tar.gz with bsdtar ==> Starting prepare()... ==> Starting build()... ==> Starting check()... ? github.com/seaweedfs/seaweedfs/weed [no test files] === RUN TestConcurrentAddRemoveNodes --- PASS: TestConcurrentAddRemoveNodes (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/cluster 0.005s === RUN TestAddServer I0701 09:55:49.492636 lock_ring.go:45 add server localhost:8080 I0701 09:55:49.492766 lock_ring.go:45 add server localhost:8081 I0701 09:55:49.492773 lock_ring.go:45 add server localhost:8082 I0701 09:55:49.492776 lock_ring.go:45 add server localhost:8083 I0701 09:55:49.492779 lock_ring.go:45 add server localhost:8084 I0701 09:55:49.492782 lock_ring.go:61 remove server localhost:8084 I0701 09:55:49.492786 lock_ring.go:61 remove server localhost:8082 I0701 09:55:49.492788 lock_ring.go:61 remove server localhost:8080 --- PASS: TestAddServer (0.11s) === RUN TestLockRing --- PASS: TestLockRing (0.22s) PASS ok github.com/seaweedfs/seaweedfs/weed/cluster/lock_manager 0.336s === RUN TestReadingTomlConfiguration database is map[connection_max:5000 enabled:true ports:[8001 8001 8002] server:192.168.1.1] servers is map[alpha:map[dc:eqdc10 ip:10.0.0.1] beta:map[dc:eqdc10 ip:10.0.0.2]] alpha ip is 10.0.0.1 --- PASS: TestReadingTomlConfiguration (0.00s) === RUN TestXYZ I0701 09:55:53.702922 volume_test.go:12 Last-Modified Mon, 08 Jul 2013 08:53:16 GMT --- PASS: TestXYZ (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/command 0.052s ? github.com/seaweedfs/seaweedfs/weed/command/scaffold [no test files] === RUN TestChunkGroup_doSearchChunks --- PASS: TestChunkGroup_doSearchChunks (0.00s) === RUN TestDoMaybeManifestize test 0 test 1 test 2 test 3 --- PASS: TestDoMaybeManifestize (0.00s) === RUN Test_removeGarbageChunks --- PASS: Test_removeGarbageChunks (0.00s) === RUN TestDoMinusChunks 2025/07/01 09:55:53 first deleted chunks: [file_id:"1" size:3 modified_ts_ns:100 source_file_id:"11" file_id:"2" offset:3 size:3 modified_ts_ns:200 file_id:"3" offset:6 size:3 modified_ts_ns:300 source_file_id:"33"] 2025/07/01 09:55:53 clusterA synced empty chunks event result: [] --- PASS: TestDoMinusChunks (0.00s) === RUN TestCompactFileChunksRealCase I0701 09:55:53.688354 filechunks2_test.go:84 before chunk 2,512f31f2c0700a [ 0, 25) I0701 09:55:53.688431 filechunks2_test.go:84 before chunk 6,512f2c2e24e9e8 [ 868352, 917585) I0701 09:55:53.688433 filechunks2_test.go:84 before chunk 7,514468dd5954ca [ 884736, 901120) I0701 09:55:53.688436 filechunks2_test.go:84 before chunk 5,5144463173fe77 [ 917504, 2297856) I0701 09:55:53.688438 filechunks2_test.go:84 before chunk 4,51444c7ab54e2d [ 2301952, 2367488) I0701 09:55:53.688440 filechunks2_test.go:84 before chunk 4,514450e643ad22 [ 2371584, 2420736) I0701 09:55:53.688441 filechunks2_test.go:84 before chunk 6,514456a5e9e4d7 [ 2449408, 2490368) I0701 09:55:53.688448 filechunks2_test.go:84 before chunk 3,51444f8d53eebe [ 2494464, 2555904) I0701 09:55:53.688450 filechunks2_test.go:84 before chunk 4,5144578b097c7e [ 2560000, 2596864) I0701 09:55:53.688451 filechunks2_test.go:84 before chunk 3,51445500b6b4ac [ 2637824, 2678784) I0701 09:55:53.688453 filechunks2_test.go:84 before chunk 1,51446285e52a61 [ 2695168, 2715648) I0701 09:55:53.688465 filechunks2_test.go:84 compacted chunk 2,512f31f2c0700a [ 0, 25) I0701 09:55:53.688468 filechunks2_test.go:84 compacted chunk 6,512f2c2e24e9e8 [ 868352, 917585) I0701 09:55:53.688470 filechunks2_test.go:84 compacted chunk 7,514468dd5954ca [ 884736, 901120) I0701 09:55:53.688472 filechunks2_test.go:84 compacted chunk 5,5144463173fe77 [ 917504, 2297856) I0701 09:55:53.688474 filechunks2_test.go:84 compacted chunk 4,51444c7ab54e2d [ 2301952, 2367488) I0701 09:55:53.688476 filechunks2_test.go:84 compacted chunk 4,514450e643ad22 [ 2371584, 2420736) I0701 09:55:53.688478 filechunks2_test.go:84 compacted chunk 6,514456a5e9e4d7 [ 2449408, 2490368) I0701 09:55:53.688480 filechunks2_test.go:84 compacted chunk 3,51444f8d53eebe [ 2494464, 2555904) I0701 09:55:53.688482 filechunks2_test.go:84 compacted chunk 4,5144578b097c7e [ 2560000, 2596864) I0701 09:55:53.688484 filechunks2_test.go:84 compacted chunk 3,51445500b6b4ac [ 2637824, 2678784) I0701 09:55:53.688488 filechunks2_test.go:84 compacted chunk 1,51446285e52a61 [ 2695168, 2715648) --- PASS: TestCompactFileChunksRealCase (0.00s) === RUN TestReadResolvedChunks resolved to 4 visible intervales [0,50) a 1 [50,150) b 2 [175,275) e 5 [275,300) d 4 --- PASS: TestReadResolvedChunks (0.00s) === RUN TestReadResolvedChunks2 resolved to 2 visible intervales [200,225) e 5 [225,250) c 3 --- PASS: TestReadResolvedChunks2 (0.00s) === RUN TestRandomizedReadResolvedChunks --- PASS: TestRandomizedReadResolvedChunks (0.01s) === RUN TestSequentialReadResolvedChunks visibles 13--- PASS: TestSequentialReadResolvedChunks (0.00s) === RUN TestActualReadResolvedChunks [0,2097152) 5,e7b96fef48 1634447487595823000 [2097152,4194304) 5,e5562640b9 1634447487595826000 [4194304,6291456) 5,df033e0fe4 1634447487595827000 [6291456,8388608) 7,eb08148a9b 1634447487595827000 [8388608,10485760) 7,e0f92d1604 1634447487595828000 [10485760,12582912) 7,e33cb63262 1634447487595828000 [12582912,14680064) 5,ea98e40e93 1634447487595829000 [14680064,16777216) 5,e165661172 1634447487595829000 [16777216,18874368) 3,e692097486 1634447487595830000 [18874368,20971520) 3,e28e2e3cbd 1634447487595830000 [20971520,23068672) 3,e443974d4e 1634447487595830000 [23068672,25165824) 2,e815bed597 1634447487595831000 [25165824,27140560) 5,e94715199e 1634447487595832000 --- PASS: TestActualReadResolvedChunks (0.00s) === RUN TestActualReadResolvedChunks2 [0,184320) 1,e7b96fef48 1 [184320,188416) 2,33562640b9 4 [188416,2285568) 4,df033e0fe4 3 --- PASS: TestActualReadResolvedChunks2 (0.00s) === RUN TestCompactFileChunks --- PASS: TestCompactFileChunks (0.00s) === RUN TestCompactFileChunks2 --- PASS: TestCompactFileChunks2 (0.00s) === RUN TestRandomFileChunksCompact --- PASS: TestRandomFileChunksCompact (0.00s) === RUN TestIntervalMerging 2025/07/01 09:55:53 ++++++++++ merged test case 0 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 0, interval start=0, stop=100, fileId=abc 2025/07/01 09:55:53 test case 0, interval start=100, stop=200, fileId=asdf 2025/07/01 09:55:53 test case 0, interval start=200, stop=300, fileId=fsad 2025/07/01 09:55:53 ++++++++++ merged test case 1 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 1, interval start=0, stop=200, fileId=asdf 2025/07/01 09:55:53 ++++++++++ merged test case 2 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 2, interval start=0, stop=70, fileId=b 2025/07/01 09:55:53 test case 2, interval start=70, stop=100, fileId=a 2025/07/01 09:55:53 ++++++++++ merged test case 3 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 3, interval start=0, stop=50, fileId=asdf 2025/07/01 09:55:53 test case 3, interval start=50, stop=300, fileId=xxxx 2025/07/01 09:55:53 ++++++++++ merged test case 4 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 4, interval start=0, stop=200, fileId=asdf 2025/07/01 09:55:53 test case 4, interval start=250, stop=500, fileId=xxxx 2025/07/01 09:55:53 ++++++++++ merged test case 5 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 5, interval start=0, stop=200, fileId=d 2025/07/01 09:55:53 test case 5, interval start=200, stop=220, fileId=c 2025/07/01 09:55:53 ++++++++++ merged test case 6 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 6, interval start=0, stop=100, fileId=xyz 2025/07/01 09:55:53 ++++++++++ merged test case 7 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 7, interval start=0, stop=2097152, fileId=3,029565bf3092 2025/07/01 09:55:53 test case 7, interval start=2097152, stop=5242880, fileId=6,029632f47ae2 2025/07/01 09:55:53 test case 7, interval start=5242880, stop=8388608, fileId=2,029734c5aa10 2025/07/01 09:55:53 test case 7, interval start=8388608, stop=11534336, fileId=5,02982f80de50 2025/07/01 09:55:53 test case 7, interval start=11534336, stop=14376529, fileId=7,0299ad723803 2025/07/01 09:55:53 ++++++++++ merged test case 8 ++++++++++++++++++++ 2025/07/01 09:55:53 test case 8, interval start=0, stop=77824, fileId=4,0b3df938e301 2025/07/01 09:55:53 test case 8, interval start=77824, stop=208896, fileId=4,0b3f0c7202f0 2025/07/01 09:55:53 test case 8, interval start=208896, stop=339968, fileId=2,0b4031a72689 2025/07/01 09:55:53 test case 8, interval start=339968, stop=471040, fileId=3,0b416a557362 2025/07/01 09:55:53 test case 8, interval start=471040, stop=472225, fileId=6,0b3e0650019c --- PASS: TestIntervalMerging (0.00s) === RUN TestChunksReading 2025/07/01 09:55:53 ++++++++++ read test case 0 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 0, chunk 0, offset=0, size=100, fileId=abc 2025/07/01 09:55:53 read case 0, chunk 1, offset=0, size=100, fileId=asdf 2025/07/01 09:55:53 read case 0, chunk 2, offset=0, size=50, fileId=fsad 2025/07/01 09:55:53 ++++++++++ read test case 1 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 1, chunk 0, offset=50, size=100, fileId=asdf 2025/07/01 09:55:53 ++++++++++ read test case 2 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 2, chunk 0, offset=20, size=30, fileId=b 2025/07/01 09:55:53 read case 2, chunk 1, offset=57, size=10, fileId=a 2025/07/01 09:55:53 ++++++++++ read test case 3 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 3, chunk 0, offset=0, size=50, fileId=asdf 2025/07/01 09:55:53 read case 3, chunk 1, offset=0, size=150, fileId=xxxx 2025/07/01 09:55:53 ++++++++++ read test case 4 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 4, chunk 0, offset=0, size=200, fileId=asdf 2025/07/01 09:55:53 read case 4, chunk 1, offset=0, size=150, fileId=xxxx 2025/07/01 09:55:53 ++++++++++ read test case 5 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 5, chunk 0, offset=0, size=200, fileId=c 2025/07/01 09:55:53 read case 5, chunk 1, offset=130, size=20, fileId=b 2025/07/01 09:55:53 ++++++++++ read test case 6 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 6, chunk 0, offset=0, size=100, fileId=xyz 2025/07/01 09:55:53 ++++++++++ read test case 7 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 7, chunk 0, offset=0, size=100, fileId=abc 2025/07/01 09:55:53 read case 7, chunk 1, offset=0, size=100, fileId=asdf 2025/07/01 09:55:53 ++++++++++ read test case 8 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 8, chunk 0, offset=0, size=90, fileId=abc 2025/07/01 09:55:53 read case 8, chunk 1, offset=0, size=100, fileId=asdf 2025/07/01 09:55:53 read case 8, chunk 2, offset=0, size=110, fileId=fsad 2025/07/01 09:55:53 ++++++++++ read test case 9 ++++++++++++++++++++ 2025/07/01 09:55:53 read case 9, chunk 0, offset=0, size=43175936, fileId=2,111fc2cbfac1 2025/07/01 09:55:53 read case 9, chunk 1, offset=0, size=9805824, fileId=2,112a36ea7f85 2025/07/01 09:55:53 read case 9, chunk 2, offset=0, size=19582976, fileId=4,112d5f31c5e7 2025/07/01 09:55:53 read case 9, chunk 3, offset=0, size=60690432, fileId=1,113245f0cdb6 2025/07/01 09:55:53 read case 9, chunk 4, offset=0, size=4014080, fileId=3,1141a70733b5 2025/07/01 09:55:53 read case 9, chunk 5, offset=0, size=16309588, fileId=1,114201d5bbdb --- PASS: TestChunksReading (0.00s) === RUN TestViewFromVisibleIntervals --- PASS: TestViewFromVisibleIntervals (0.00s) === RUN TestViewFromVisibleIntervals2 --- PASS: TestViewFromVisibleIntervals2 (0.00s) === RUN TestViewFromVisibleIntervals3 --- PASS: TestViewFromVisibleIntervals3 (0.00s) === RUN TestCompactFileChunks3 --- PASS: TestCompactFileChunks3 (0.00s) === RUN TestFilerConf --- PASS: TestFilerConf (0.00s) === RUN TestProtoMarshal e to: 234,2423423422 * 2342342354223234,2342342342"# 0Ø: text/jsonP --- PASS: TestProtoMarshal (0.00s) === RUN TestIntervalList_Overlay [0,25) 6 6 [25,50) 1 1 [50,150) 2 2 [175,210) 5 5 [210,225) 3 3 [225,250) 4 4 [0,25) 6 6 [25,50) 1 1 [50,150) 7 7 [175,210) 5 5 [210,225) 3 3 [225,250) 4 4 --- PASS: TestIntervalList_Overlay (0.00s) === RUN TestIntervalList_Overlay2 [0,50) 2 2 [50,100) 1 1 --- PASS: TestIntervalList_Overlay2 (0.00s) === RUN TestIntervalList_Overlay3 [0,60) 2 2 [60,100) 1 1 --- PASS: TestIntervalList_Overlay3 (0.00s) === RUN TestIntervalList_Overlay4 [0,100) 2 2 --- PASS: TestIntervalList_Overlay4 (0.00s) === RUN TestIntervalList_Overlay5 [0,110) 2 2 --- PASS: TestIntervalList_Overlay5 (0.00s) === RUN TestIntervalList_Overlay6 [50,110) 2 2 --- PASS: TestIntervalList_Overlay6 (0.00s) === RUN TestIntervalList_Overlay7 [50,90) 2 2 [90,100) 1 1 --- PASS: TestIntervalList_Overlay7 (0.00s) === RUN TestIntervalList_Overlay8 [50,60) 1 1 [60,90) 2 2 [90,100) 1 1 --- PASS: TestIntervalList_Overlay8 (0.00s) === RUN TestIntervalList_Overlay9 [50,60) 1 1 [60,100) 2 2 --- PASS: TestIntervalList_Overlay9 (0.00s) === RUN TestIntervalList_Overlay10 [50,60) 1 1 [60,110) 2 2 --- PASS: TestIntervalList_Overlay10 (0.00s) === RUN TestIntervalList_Overlay11 [0,90) 5 5 [90,100) 1 1 [100,110) 2 2 --- PASS: TestIntervalList_Overlay11 (0.00s) === RUN TestIntervalList_insertInterval1 [50,150) 2 2 [200,250) 3 3 --- PASS: TestIntervalList_insertInterval1 (0.00s) === RUN TestIntervalList_insertInterval2 [0,25) 3 3 [50,150) 2 2 --- PASS: TestIntervalList_insertInterval2 (0.00s) === RUN TestIntervalList_insertInterval3 [0,75) 3 3 [75,150) 2 2 [200,250) 4 4 --- PASS: TestIntervalList_insertInterval3 (0.00s) === RUN TestIntervalList_insertInterval4 [0,200) 3 3 [200,250) 4 4 --- PASS: TestIntervalList_insertInterval4 (0.00s) === RUN TestIntervalList_insertInterval5 [0,225) 5 5 [225,250) 4 4 --- PASS: TestIntervalList_insertInterval5 (0.00s) === RUN TestIntervalList_insertInterval6 [0,50) 1 1 [50,150) 2 2 [150,200) 1 1 [200,250) 4 4 [250,275) 1 1 --- PASS: TestIntervalList_insertInterval6 (0.00s) === RUN TestIntervalList_insertInterval7 [50,150) 2 2 [150,200) 1 1 [200,250) 4 4 [250,275) 1 1 --- PASS: TestIntervalList_insertInterval7 (0.00s) === RUN TestIntervalList_insertInterval8 [50,75) 2 2 [75,200) 3 3 [200,250) 4 4 [250,275) 3 3 --- PASS: TestIntervalList_insertInterval8 (0.00s) === RUN TestIntervalList_insertInterval9 [50,150) 3 3 [200,250) 4 4 --- PASS: TestIntervalList_insertInterval9 (0.00s) === RUN TestIntervalList_insertInterval10 [50,100) 2 2 [100,200) 5 5 [200,300) 4 4 --- PASS: TestIntervalList_insertInterval10 (0.00s) === RUN TestIntervalList_insertInterval11 [0,64) 1 1 [64,68) 2 2 [68,72) 4 4 [72,136) 3 3 --- PASS: TestIntervalList_insertInterval11 (0.00s) === RUN TestIntervalList_insertIntervalStruct [0,64) 1 {1 0 0} [64,68) 4 {4 0 0} [68,72) 2 {2 0 0} [72,136) 3 {3 0 0} --- PASS: TestIntervalList_insertIntervalStruct (0.00s) === RUN TestReaderAt --- PASS: TestReaderAt (0.00s) === RUN TestReaderAt0 --- PASS: TestReaderAt0 (0.00s) === RUN TestReaderAt1 --- PASS: TestReaderAt1 (0.00s) === RUN TestReaderAtGappedChunksDoNotLeak --- PASS: TestReaderAtGappedChunksDoNotLeak (0.00s) === RUN TestReaderAtSparseFileDoesNotLeak --- PASS: TestReaderAtSparseFileDoesNotLeak (0.00s) === RUN TestFilerRemoteStorage_FindRemoteStorageClient --- PASS: TestFilerRemoteStorage_FindRemoteStorageClient (0.00s) === RUN TestS3Conf --- PASS: TestS3Conf (0.00s) === RUN TestCheckDuplicateAccessKey --- PASS: TestCheckDuplicateAccessKey (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/filer 0.053s ? github.com/seaweedfs/seaweedfs/weed/filer/abstract_sql [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/arangodb [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/cassandra [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/cassandra2 [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/elastic/v7 [no test files] === RUN TestStore --- PASS: TestStore (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/filer/etcd 0.031s ? github.com/seaweedfs/seaweedfs/weed/filer/hbase [no test files] === RUN TestCreateAndFind I0701 09:55:53.723152 leveldb_store.go:47 filer store dir: /tmp/TestCreateAndFind151888640/001 I0701 09:55:53.723450 file_util.go:27 Folder /tmp/TestCreateAndFind151888640/001 Permission: -rwxr-xr-x I0701 09:55:53.724243 filer.go:155 create filer.store.id to 2144982562 --- PASS: TestCreateAndFind (0.03s) === RUN TestEmptyRoot I0701 09:55:53.733213 leveldb_store.go:47 filer store dir: /tmp/TestEmptyRoot2300128849/001 I0701 09:55:53.733234 file_util.go:27 Folder /tmp/TestEmptyRoot2300128849/001 Permission: -rwxr-xr-x I0701 09:55:53.733848 filer.go:155 create filer.store.id to -1343158228 --- PASS: TestEmptyRoot (0.01s) PASS ok github.com/seaweedfs/seaweedfs/weed/filer/leveldb 0.077s === RUN TestCreateAndFind I0701 09:55:53.723236 leveldb2_store.go:43 filer store leveldb2 dir: /tmp/TestCreateAndFind3728131723/001 I0701 09:55:53.723411 file_util.go:27 Folder /tmp/TestCreateAndFind3728131723/001 Permission: -rwxr-xr-x I0701 09:55:53.724500 filer.go:155 create filer.store.id to -478995870 --- PASS: TestCreateAndFind (0.04s) === RUN TestEmptyRoot I0701 09:55:53.729884 leveldb2_store.go:43 filer store leveldb2 dir: /tmp/TestEmptyRoot3121100251/001 I0701 09:55:53.729902 file_util.go:27 Folder /tmp/TestEmptyRoot3121100251/001 Permission: -rwxr-xr-x I0701 09:55:53.730584 filer.go:155 create filer.store.id to -836833710 --- PASS: TestEmptyRoot (0.01s) PASS ok github.com/seaweedfs/seaweedfs/weed/filer/leveldb2 0.079s === RUN TestCreateAndFind I0701 09:55:53.719585 leveldb3_store.go:50 filer store leveldb3 dir: /tmp/TestCreateAndFind755671425/001 I0701 09:55:53.719769 file_util.go:27 Folder /tmp/TestCreateAndFind755671425/001 Permission: -rwxr-xr-x I0701 09:55:53.720538 filer.go:155 create filer.store.id to 1417758920 --- PASS: TestCreateAndFind (0.02s) === RUN TestEmptyRoot I0701 09:55:53.724101 leveldb3_store.go:50 filer store leveldb3 dir: /tmp/TestEmptyRoot1821844971/001 I0701 09:55:53.724122 file_util.go:27 Folder /tmp/TestEmptyRoot1821844971/001 Permission: -rwxr-xr-x I0701 09:55:53.724699 filer.go:155 create filer.store.id to 1033439929 --- PASS: TestEmptyRoot (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/filer/leveldb3 0.063s ? github.com/seaweedfs/seaweedfs/weed/filer/mongodb [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/mysql [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/mysql2 [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/postgres [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/postgres2 [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/redis [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/redis2 [no test files] testing: warning: no tests to run PASS ok github.com/seaweedfs/seaweedfs/weed/filer/redis3 0.021s [no tests to run] ? github.com/seaweedfs/seaweedfs/weed/filer/redis_lua [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/redis_lua/stored_procedure [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/sqlite [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/store_test [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/tarantool [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/tikv [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer/ydb [no test files] ? github.com/seaweedfs/seaweedfs/weed/filer_client [no test files] === RUN TestShortHostname --- PASS: TestShortHostname (0.00s) === RUN TestInfo I0701 09:55:55.166946 glog_test.go:92 test --- PASS: TestInfo (0.00s) === RUN TestInfoDepth I0701 09:55:55.167009 glog_test.go:109 depth-test0 I0701 09:55:55.167011 glog_test.go:110 depth-test1 --- PASS: TestInfoDepth (0.00s) === RUN TestCopyStandardLogToPanic --- PASS: TestCopyStandardLogToPanic (0.00s) === RUN TestStandardLog I0701 09:55:55.167053 glog_test.go:163 test --- PASS: TestStandardLog (0.00s) === RUN TestHeader I0102 15:04:05.067890 glog_test.go:181 test --- PASS: TestHeader (0.00s) === RUN TestError E0701 09:55:55.167093 glog_test.go:202 test --- PASS: TestError (0.00s) === RUN TestWarning W0701 09:55:55.167105 glog_test.go:224 test --- PASS: TestWarning (0.00s) === RUN TestV I0701 09:55:55.167117 glog_test.go:243 test --- PASS: TestV (0.00s) === RUN TestVmoduleOn I0701 09:55:55.167136 glog_test.go:267 test --- PASS: TestVmoduleOn (0.00s) === RUN TestVmoduleOff --- PASS: TestVmoduleOff (0.00s) === RUN TestVmoduleGlob --- PASS: TestVmoduleGlob (0.00s) === RUN TestRollover I0701 09:55:55.167189 glog_test.go:339 x I0701 09:55:55.167514 glog_test.go:348 xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx I0701 09:55:56.169457 glog_test.go:361 x --- PASS: TestRollover (1.00s) === RUN TestLogBacktraceAt I0701 09:55:56.169669 glog_test.go:395 we want a stack trace here goroutine 2 [running]: github.com/seaweedfs/seaweedfs/weed/glog.stacks(0x0) /startdir/src/seaweedfs-3.92/weed/glog/glog.go:768 +0x85 github.com/seaweedfs/seaweedfs/weed/glog.(*loggingT).output(0x845060, 0x0, 0xc0000ae000, {0x6ec88e?, 0x1?}, 0x0?, 0x0) /startdir/src/seaweedfs-3.92/weed/glog/glog.go:677 +0xe5 github.com/seaweedfs/seaweedfs/weed/glog.(*loggingT).printDepth(0x845060, 0x0, 0xc00008be90?, {0xc00008be30, 0x1, 0x1}) /startdir/src/seaweedfs-3.92/weed/glog/glog.go:648 +0xea github.com/seaweedfs/seaweedfs/weed/glog.(*loggingT).print(...) /startdir/src/seaweedfs-3.92/weed/glog/glog.go:639 github.com/seaweedfs/seaweedfs/weed/glog.Info(...) /startdir/src/seaweedfs-3.92/weed/glog/glog.go:1061 github.com/seaweedfs/seaweedfs/weed/glog.TestLogBacktraceAt(0xc000003500) /startdir/src/seaweedfs-3.92/weed/glog/glog_test.go:395 +0x438 testing.tRunner(0xc000003500, 0x663078) /usr/lib/go/src/testing/testing.go:1792 +0xf4 created by testing.(*T).Run in goroutine 1 /usr/lib/go/src/testing/testing.go:1851 +0x413 --- PASS: TestLogBacktraceAt (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/glog 1.016s === RUN TestGetActionsUserPath --- PASS: TestGetActionsUserPath (0.00s) === RUN TestGetActionsWildcardPath --- PASS: TestGetActionsWildcardPath (0.00s) === RUN TestGetActionsInvalidAction --- PASS: TestGetActionsInvalidAction (0.00s) === RUN TestCreateUser --- PASS: TestCreateUser (0.00s) === RUN TestListUsers --- PASS: TestListUsers (0.00s) === RUN TestListAccessKeys --- PASS: TestListAccessKeys (0.00s) === RUN TestGetUser --- PASS: TestGetUser (0.00s) === RUN TestCreatePolicy --- PASS: TestCreatePolicy (0.00s) === RUN TestPutUserPolicy --- PASS: TestPutUserPolicy (0.00s) === RUN TestPutUserPolicyError E0701 09:55:56.198303 iamapi_management_handlers.go:508 PutUserPolicy: the user with name InvalidUser cannot be found E0701 09:55:56.198567 iamapi_handlers.go:29 Response the user with name InvalidUser cannot be found --- PASS: TestPutUserPolicyError (0.00s) === RUN TestGetUserPolicy --- PASS: TestGetUserPolicy (0.00s) === RUN TestUpdateUser --- PASS: TestUpdateUser (0.00s) === RUN TestDeleteUser --- PASS: TestDeleteUser (0.00s) === RUN TestHandleImplicitUsername --- PASS: TestHandleImplicitUsername (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/iamapi 0.059s === RUN TestCropping --- PASS: TestCropping (0.20s) === RUN TestXYZ --- PASS: TestXYZ (0.73s) === RUN TestResizing --- PASS: TestResizing (0.06s) PASS ok github.com/seaweedfs/seaweedfs/weed/images 1.005s === RUN TestInodeEntry_removeOnePath === RUN TestInodeEntry_removeOnePath/actual_case === RUN TestInodeEntry_removeOnePath/empty === RUN TestInodeEntry_removeOnePath/single === RUN TestInodeEntry_removeOnePath/first === RUN TestInodeEntry_removeOnePath/middle === RUN TestInodeEntry_removeOnePath/last === RUN TestInodeEntry_removeOnePath/not_found --- PASS: TestInodeEntry_removeOnePath (0.00s) --- PASS: TestInodeEntry_removeOnePath/actual_case (0.00s) --- PASS: TestInodeEntry_removeOnePath/empty (0.00s) --- PASS: TestInodeEntry_removeOnePath/single (0.00s) --- PASS: TestInodeEntry_removeOnePath/first (0.00s) --- PASS: TestInodeEntry_removeOnePath/middle (0.00s) --- PASS: TestInodeEntry_removeOnePath/last (0.00s) --- PASS: TestInodeEntry_removeOnePath/not_found (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/mount 0.033s ? github.com/seaweedfs/seaweedfs/weed/mount/meta_cache [no test files] === RUN Test_PageChunkWrittenIntervalList --- PASS: Test_PageChunkWrittenIntervalList (0.00s) === RUN Test_PageChunkWrittenIntervalList1 --- PASS: Test_PageChunkWrittenIntervalList1 (0.00s) === RUN TestUploadPipeline --- PASS: TestUploadPipeline (23.50s) PASS ok github.com/seaweedfs/seaweedfs/weed/mount/page_writer 23.512s ? github.com/seaweedfs/seaweedfs/weed/mount/unmount [no test files] ? github.com/seaweedfs/seaweedfs/weed/mq/agent [no test files] ? github.com/seaweedfs/seaweedfs/weed/mq/broker [no test files] ? github.com/seaweedfs/seaweedfs/weed/mq/client/agent_client [no test files] ? github.com/seaweedfs/seaweedfs/weed/mq/client/pub_client [no test files] ? github.com/seaweedfs/seaweedfs/weed/mq/client/sub_client [no test files] ? github.com/seaweedfs/seaweedfs/weed/mq/logstore [no test files] === RUN Test_allocateOneBroker === RUN Test_allocateOneBroker/test_only_one_broker I0701 09:55:56.758839 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 1, followerCount: 1, assignments: [partition:{ring_size:2520 range_stop:2520 unix_time_ns:1751363756758823090}] I0701 09:55:56.759271 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 1, followerCount: 1, assignments: [partition:{ring_size:2520 range_stop:2520 unix_time_ns:1751363756758823090} leader_broker:"localhost:17777"] hasChanges: true I0701 09:55:56.759286 allocate.go:33 allocate topic partitions 1: [partition:{ring_size:2520 range_stop:2520 unix_time_ns:1751363756758823090} leader_broker:"localhost:17777"] --- PASS: Test_allocateOneBroker (0.00s) --- PASS: Test_allocateOneBroker/test_only_one_broker (0.00s) === RUN TestEnsureAssignmentsToActiveBrokersX === RUN TestEnsureAssignmentsToActiveBrokersX/test_empty_leader test empty leader before [partition:{} follower_broker:"localhost:2"] I0701 09:55:56.759460 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} follower_broker:"localhost:2"] I0701 09:55:56.759490 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:6" follower_broker:"localhost:2"] hasChanges: true test empty leader after [partition:{} leader_broker:"localhost:6" follower_broker:"localhost:2"] === RUN TestEnsureAssignmentsToActiveBrokersX/test_empty_follower test empty follower before [partition:{} leader_broker:"localhost:1"] I0701 09:55:56.759519 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:1"] I0701 09:55:56.759547 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:3"] hasChanges: true test empty follower after [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:3"] === RUN TestEnsureAssignmentsToActiveBrokersX/test_dead_follower test dead follower before [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:200"] I0701 09:55:56.759575 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:200"] I0701 09:55:56.759600 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:4"] hasChanges: true test dead follower after [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:4"] === RUN TestEnsureAssignmentsToActiveBrokersX/test_dead_leader_and_follower test dead leader and follower before [partition:{} leader_broker:"localhost:100" follower_broker:"localhost:200"] I0701 09:55:56.759627 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:100" follower_broker:"localhost:200"] I0701 09:55:56.759654 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 6, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:3" follower_broker:"localhost:4"] hasChanges: true test dead leader and follower after [partition:{} leader_broker:"localhost:3" follower_broker:"localhost:4"] === RUN TestEnsureAssignmentsToActiveBrokersX/test_low_active_brokers test low active brokers before [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:2"] I0701 09:55:56.759677 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 2, followerCount: 3, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:2"] I0701 09:55:56.759704 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 2, followerCount: 3, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:2"] hasChanges: false test low active brokers after [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:2"] === RUN TestEnsureAssignmentsToActiveBrokersX/test_low_active_brokers_with_one_follower test low active brokers with one follower before [partition:{} leader_broker:"localhost:1"] I0701 09:55:56.759728 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 2, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:1"] I0701 09:55:56.759752 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 2, followerCount: 1, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:1"] hasChanges: true test low active brokers with one follower after [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:1"] === RUN TestEnsureAssignmentsToActiveBrokersX/test_single_active_broker test single active broker before [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:2"] I0701 09:55:56.759775 allocate.go:81 EnsureAssignmentsToActiveBrokers: activeBrokers: 1, followerCount: 3, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:2"] I0701 09:55:56.759802 allocate.go:125 EnsureAssignmentsToActiveBrokers: activeBrokers: 1, followerCount: 3, assignments: [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:1"] hasChanges: true test single active broker after [partition:{} leader_broker:"localhost:1" follower_broker:"localhost:1"] --- PASS: TestEnsureAssignmentsToActiveBrokersX (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_empty_leader (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_empty_follower (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_dead_follower (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_dead_leader_and_follower (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_low_active_brokers (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_low_active_brokers_with_one_follower (0.00s) --- PASS: TestEnsureAssignmentsToActiveBrokersX/test_single_active_broker (0.00s) === RUN TestBalanceTopicPartitionOnBrokers === RUN TestBalanceTopicPartitionOnBrokers/test --- PASS: TestBalanceTopicPartitionOnBrokers (0.00s) --- PASS: TestBalanceTopicPartitionOnBrokers/test (0.00s) === RUN Test_findMissingPartitions === RUN Test_findMissingPartitions/one_partition === RUN Test_findMissingPartitions/two_partitions === RUN Test_findMissingPartitions/four_partitions,_missing_last_two === RUN Test_findMissingPartitions/four_partitions,_missing_first_two === RUN Test_findMissingPartitions/four_partitions,_missing_middle_two === RUN Test_findMissingPartitions/four_partitions,_missing_three --- PASS: Test_findMissingPartitions (0.00s) --- PASS: Test_findMissingPartitions/one_partition (0.00s) --- PASS: Test_findMissingPartitions/two_partitions (0.00s) --- PASS: Test_findMissingPartitions/four_partitions,_missing_last_two (0.00s) --- PASS: Test_findMissingPartitions/four_partitions,_missing_first_two (0.00s) --- PASS: Test_findMissingPartitions/four_partitions,_missing_middle_two (0.00s) --- PASS: Test_findMissingPartitions/four_partitions,_missing_three (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/mq/pub_balancer 0.036s === RUN TestEnumScalarType === RUN TestEnumScalarType/Boolean === RUN TestEnumScalarType/Integer === RUN TestEnumScalarType/Long === RUN TestEnumScalarType/Float === RUN TestEnumScalarType/Double === RUN TestEnumScalarType/Bytes === RUN TestEnumScalarType/String --- PASS: TestEnumScalarType (0.00s) --- PASS: TestEnumScalarType/Boolean (0.00s) --- PASS: TestEnumScalarType/Integer (0.00s) --- PASS: TestEnumScalarType/Long (0.00s) --- PASS: TestEnumScalarType/Float (0.00s) --- PASS: TestEnumScalarType/Double (0.00s) --- PASS: TestEnumScalarType/Bytes (0.00s) --- PASS: TestEnumScalarType/String (0.00s) === RUN TestField --- PASS: TestField (0.00s) === RUN TestRecordType fields: < name: "field_key" field_index: 1 type: < scalar_type: INT32 > > fields: < name: "field_record" field_index: 2 type: < record_type: < fields: < name: "field_1" field_index: 1 type: < scalar_type: INT32 > > fields: < name: "field_2" field_index: 2 type: < scalar_type: STRING > > > > > {"fields":[{"name":"field_key","field_index":1,"type":{"Kind":{"ScalarType":1}}},{"name":"field_record","field_index":2,"type":{"Kind":{"RecordType":{"fields":[{"name":"field_1","field_index":1,"type":{"Kind":{"ScalarType":1}}},{"name":"field_2","field_index":2,"type":{"Kind":{"ScalarType":7}}}]}}}}]} --- PASS: TestRecordType (0.00s) === RUN TestStructToSchema === RUN TestStructToSchema/scalar_type === RUN TestStructToSchema/simple_struct_type === RUN TestStructToSchema/simple_list === RUN TestStructToSchema/simple_[]byte === RUN TestStructToSchema/nested_simpe_structs === RUN TestStructToSchema/nested_struct_type --- PASS: TestStructToSchema (0.00s) --- PASS: TestStructToSchema/scalar_type (0.00s) --- PASS: TestStructToSchema/simple_struct_type (0.00s) --- PASS: TestStructToSchema/simple_list (0.00s) --- PASS: TestStructToSchema/simple_[]byte (0.00s) --- PASS: TestStructToSchema/nested_simpe_structs (0.00s) --- PASS: TestStructToSchema/nested_struct_type (0.00s) === RUN TestToParquetLevels === RUN TestToParquetLevels/nested_type --- PASS: TestToParquetLevels (0.00s) --- PASS: TestToParquetLevels/nested_type (0.00s) === RUN TestWriteReadParquet RecordType: fields:{name:"Address" type:{record_type:{fields:{name:"City" type:{scalar_type:STRING}} fields:{name:"Street" type:{scalar_type:STRING}}}}} fields:{name:"Company" type:{scalar_type:STRING}} fields:{name:"CreatedAt" type:{scalar_type:INT64}} fields:{name:"ID" type:{scalar_type:INT64}} fields:{name:"Person" type:{record_type:{fields:{name:"emails" type:{list_type:{element_type:{scalar_type:STRING}}}} fields:{name:"zName" type:{scalar_type:STRING}}}}} ParquetSchema: message example { optional group Address { optional binary City; optional binary Street; } optional binary Company; optional int64 CreatedAt; optional int64 ID; optional group Person { repeated binary emails; optional binary zName; } } Go Type: struct { Address *struct { City *[]uint8; Street *[]uint8 }; Company *[]uint8; CreatedAt *int64; ID *int64; Person *struct { Emails []*[]uint8; ZName *[]uint8 } } Write RecordValue: fields:{key:"Company" value:{string_value:"company_0"}} fields:{key:"CreatedAt" value:{int64_value:2}} fields:{key:"ID" value:{int64_value:1}} fields:{key:"Person" value:{record_value:{fields:{key:"emails" value:{list_value:{values:{string_value:"john_0@a.com"} values:{string_value:"john_0@b.com"} values:{string_value:"john_0@c.com"} values:{string_value:"john_0@d.com"} values:{string_value:"john_0@e.com"}}}} fields:{key:"zName" value:{string_value:"john_0"}}}}} Build Row: [C:0 D:0 R:0 V: C:1 D:0 R:0 V: C:2 D:1 R:0 V:company_0 C:3 D:1 R:0 V:2 C:4 D:1 R:0 V:1 C:5 D:2 R:0 V:john_0@a.com C:5 D:2 R:1 V:john_0@b.com C:5 D:2 R:1 V:john_0@c.com C:5 D:2 R:1 V:john_0@d.com C:5 D:2 R:1 V:john_0@e.com C:6 D:2 R:0 V:john_0] Write RecordValue: fields:{key:"Company" value:{string_value:"company_1"}} fields:{key:"CreatedAt" value:{int64_value:4}} fields:{key:"ID" value:{int64_value:2}} fields:{key:"Person" value:{record_value:{fields:{key:"emails" value:{list_value:{values:{string_value:"john_1@a.com"} values:{string_value:"john_1@b.com"} values:{string_value:"john_1@c.com"} values:{string_value:"john_1@d.com"} values:{string_value:"john_1@e.com"}}}} fields:{key:"zName" value:{string_value:"john_1"}}}}} Build Row: [C:0 D:0 R:0 V: C:1 D:0 R:0 V: C:2 D:1 R:0 V:company_1 C:3 D:1 R:0 V:4 C:4 D:1 R:0 V:2 C:5 D:2 R:0 V:john_1@a.com C:5 D:2 R:1 V:john_1@b.com C:5 D:2 R:1 V:john_1@c.com C:5 D:2 R:1 V:john_1@d.com C:5 D:2 R:1 V:john_1@e.com C:6 D:2 R:0 V:john_1] Write RecordValue: fields:{key:"Company" value:{string_value:"company_2"}} fields:{key:"CreatedAt" value:{int64_value:6}} fields:{key:"ID" value:{int64_value:3}} fields:{key:"Person" value:{record_value:{fields:{key:"emails" value:{list_value:{values:{string_value:"john_2@a.com"} values:{string_value:"john_2@b.com"} values:{string_value:"john_2@c.com"} values:{string_value:"john_2@d.com"} values:{string_value:"john_2@e.com"}}}} fields:{key:"zName" value:{string_value:"john_2"}}}}} Build Row: [C:0 D:0 R:0 V: C:1 D:0 R:0 V: C:2 D:1 R:0 V:company_2 C:3 D:1 R:0 V:6 C:4 D:1 R:0 V:3 C:5 D:2 R:0 V:john_2@a.com C:5 D:2 R:1 V:john_2@b.com C:5 D:2 R:1 V:john_2@c.com C:5 D:2 R:1 V:john_2@d.com C:5 D:2 R:1 V:john_2@e.com C:6 D:2 R:0 V:john_2] Read RecordValue: fields:{key:"Address" value:{record_value:{fields:{key:"City" value:{string_value:""}} fields:{key:"Street" value:{string_value:""}}}}} fields:{key:"Company" value:{string_value:"company_0"}} fields:{key:"CreatedAt" value:{int64_value:2}} fields:{key:"ID" value:{int64_value:1}} fields:{key:"Person" value:{record_value:{fields:{key:"emails" value:{list_value:{values:{string_value:"john_0@a.com"} values:{string_value:"john_0@b.com"} values:{string_value:"john_0@c.com"} values:{string_value:"john_0@d.com"} values:{string_value:"john_0@e.com"}}}} fields:{key:"zName" value:{string_value:"john_0"}}}}} Read RecordValue: fields:{key:"Address" value:{record_value:{fields:{key:"City" value:{string_value:""}} fields:{key:"Street" value:{string_value:""}}}}} fields:{key:"Company" value:{string_value:"company_1"}} fields:{key:"CreatedAt" value:{int64_value:4}} fields:{key:"ID" value:{int64_value:2}} fields:{key:"Person" value:{record_value:{fields:{key:"emails" value:{list_value:{values:{string_value:"john_1@a.com"} values:{string_value:"john_1@b.com"} values:{string_value:"john_1@c.com"} values:{string_value:"john_1@d.com"} values:{string_value:"john_1@e.com"}}}} fields:{key:"zName" value:{string_value:"john_1"}}}}} Read RecordValue: fields:{key:"Address" value:{record_value:{fields:{key:"City" value:{string_value:""}} fields:{key:"Street" value:{string_value:""}}}}} fields:{key:"Company" value:{string_value:"company_2"}} fields:{key:"CreatedAt" value:{int64_value:6}} fields:{key:"ID" value:{int64_value:3}} fields:{key:"Person" value:{record_value:{fields:{key:"emails" value:{list_value:{values:{string_value:"john_2@a.com"} values:{string_value:"john_2@b.com"} values:{string_value:"john_2@c.com"} values:{string_value:"john_2@d.com"} values:{string_value:"john_2@e.com"}}}} fields:{key:"zName" value:{string_value:"john_2"}}}}} total: 3 --- PASS: TestWriteReadParquet (0.02s) PASS ok github.com/seaweedfs/seaweedfs/weed/mq/schema 0.031s === RUN TestMessageSerde serialized size 368 --- PASS: TestMessageSerde (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/mq/segment 0.004s === RUN TestRingBuffer --- PASS: TestRingBuffer (0.00s) === RUN TestInflightMessageTracker --- PASS: TestInflightMessageTracker (0.00s) === RUN TestInflightMessageTracker2 --- PASS: TestInflightMessageTracker2 (0.00s) === RUN TestInflightMessageTracker3 --- PASS: TestInflightMessageTracker3 (0.00s) === RUN TestInflightMessageTracker4 --- PASS: TestInflightMessageTracker4 (0.00s) === RUN TestAddConsumerInstance &{isAssign:true partition:{RangeStart:0 RangeStop:1 RingSize:3 UnixTimeNs:0} consumer:first ts:{wall:13983969361740280973 ext:514681371 loc:0x1ab6a40}} &{isAssign:true partition:{RangeStart:1 RangeStop:2 RingSize:3 UnixTimeNs:0} consumer:first ts:{wall:13983969361740288156 ext:514688544 loc:0x1ab6a40}} --- PASS: TestAddConsumerInstance (1.00s) === RUN TestMultipleConsumerInstances &{isAssign:true partition:{RangeStart:0 RangeStop:1 RingSize:3 UnixTimeNs:0} consumer:first ts:{wall:13983969362816733559 ext:1517392133 loc:0x1ab6a40}} &{isAssign:true partition:{RangeStart:1 RangeStop:2 RingSize:3 UnixTimeNs:0} consumer:second ts:{wall:13983969362816736605 ext:1517395179 loc:0x1ab6a40}} &{isAssign:true partition:{RangeStart:2 RangeStop:3 RingSize:3 UnixTimeNs:0} consumer:third ts:{wall:13983969362816740692 ext:1517399266 loc:0x1ab6a40}} --- PASS: TestMultipleConsumerInstances (1.00s) === RUN TestConfirmAdjustment &{isAssign:true partition:{RangeStart:1 RangeStop:2 RingSize:3 UnixTimeNs:0} consumer:first ts:{wall:13983969363887577801 ext:2514494551 loc:0x1ab6a40}} &{isAssign:true partition:{RangeStart:2 RangeStop:3 RingSize:3 UnixTimeNs:0} consumer:third ts:{wall:13983969363887581999 ext:2514498749 loc:0x1ab6a40}} &{isAssign:true partition:{RangeStart:0 RangeStop:1 RingSize:3 UnixTimeNs:0} consumer:second ts:{wall:13983969363887583322 ext:2514500072 loc:0x1ab6a40}} &{isAssign:true partition:{RangeStart:2 RangeStop:3 RingSize:3 UnixTimeNs:0} consumer:first ts:{wall:13983969366037806927 ext:4517240029 loc:0x1ab6a40}} --- PASS: TestConfirmAdjustment (5.00s) === RUN Test_doBalanceSticky === RUN Test_doBalanceSticky/1_consumer_instance,_1_partition === RUN Test_doBalanceSticky/2_consumer_instances,_1_partition === RUN Test_doBalanceSticky/1_consumer_instance,_2_partitions === RUN Test_doBalanceSticky/2_consumer_instances,_2_partitions === RUN Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_deleted_consumer_instance === RUN Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_new_consumer_instance === RUN Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_new_partition === RUN Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_new_partition,_1_new_consumer_instance --- PASS: Test_doBalanceSticky (0.00s) --- PASS: Test_doBalanceSticky/1_consumer_instance,_1_partition (0.00s) --- PASS: Test_doBalanceSticky/2_consumer_instances,_1_partition (0.00s) --- PASS: Test_doBalanceSticky/1_consumer_instance,_2_partitions (0.00s) --- PASS: Test_doBalanceSticky/2_consumer_instances,_2_partitions (0.00s) --- PASS: Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_deleted_consumer_instance (0.00s) --- PASS: Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_new_consumer_instance (0.00s) --- PASS: Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_new_partition (0.00s) --- PASS: Test_doBalanceSticky/2_consumer_instances,_2_partitions,_1_new_partition,_1_new_consumer_instance (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/mq/sub_coordinator 7.020s ? github.com/seaweedfs/seaweedfs/weed/mq/topic [no test files] ? github.com/seaweedfs/seaweedfs/weed/notification [no test files] ? github.com/seaweedfs/seaweedfs/weed/notification/aws_sqs [no test files] ? github.com/seaweedfs/seaweedfs/weed/notification/gocdk_pub_sub [no test files] ? github.com/seaweedfs/seaweedfs/weed/notification/google_pub_sub [no test files] ? github.com/seaweedfs/seaweedfs/weed/notification/kafka [no test files] ? github.com/seaweedfs/seaweedfs/weed/notification/log [no test files] === RUN TestCaching vid 123 locations = [{a.com:8080 0}] --- PASS: TestCaching (2.01s) === RUN TestCreateNeedleFromRequest needle: 0f084d17353afda0 Size:0, DataSize:0, Name:t.txt, Mime:text/plain; charset=utf-8 Compressed:true, originalSize: 1422 W0701 09:55:59.846843 upload_content.go:192 uploading 0 to http://localhost:8080/389,0f084d17353afda0: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF needle: 0f084d17353afda0 Size:0, DataSize:0, Name:t.txt, Mime:text/plain; charset=utf-8 Compressed:true, originalSize: 1422 W0701 09:56:00.328140 upload_content.go:192 uploading 1 to http://localhost:8080/389,0f084d17353afda0: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF needle: 0f084d17353afda0 Size:0, DataSize:0, Name:t.txt, Mime:text/plain; charset=utf-8 Compressed:true, originalSize: 1422 W0701 09:56:01.046809 upload_content.go:192 uploading 2 to http://localhost:8080/389,0f084d17353afda0: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF err: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF uploadResult: needle: 0f084d17353afda0 Size:0, DataSize:0, Name:t.txt, Mime:text/plain Compressed:true, dataSize:803 originalSize:1422 W0701 09:56:01.046944 upload_content.go:192 uploading 0 to http://localhost:8080/389,0f084d17353afda0: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF needle: 0f084d17353afda0 Size:0, DataSize:0, Name:t.txt, Mime:text/plain Compressed:true, dataSize:803 originalSize:1422 W0701 09:56:01.524738 upload_content.go:192 uploading 1 to http://localhost:8080/389,0f084d17353afda0: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF needle: 0f084d17353afda0 Size:0, DataSize:0, Name:t.txt, Mime:text/plain Compressed:true, dataSize:803 originalSize:1422 W0701 09:56:02.239663 upload_content.go:192 uploading 2 to http://localhost:8080/389,0f084d17353afda0: upload t.txt 803 bytes to http://localhost:8080/389,0f084d17353afda0: EOF --- PASS: TestCreateNeedleFromRequest (2.40s) PASS ok github.com/seaweedfs/seaweedfs/weed/operation 4.427s === RUN TestJsonpMarshalUnmarshal marshalled: { "backendType": "aws", "backendId": "", "key": "", "offset": "0", "fileSize": "12", "modifiedTime": "0", "extension": "" } unmarshalled: backend_type:"aws" backend_id:"temp" file_size:12 --- PASS: TestJsonpMarshalUnmarshal (0.00s) === RUN TestServerAddresses_ToAddressMapOrSrv_shouldRemovePrefix --- PASS: TestServerAddresses_ToAddressMapOrSrv_shouldRemovePrefix (0.00s) === RUN TestServerAddresses_ToAddressMapOrSrv_shouldHandleIPPortList --- PASS: TestServerAddresses_ToAddressMapOrSrv_shouldHandleIPPortList (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/pb 0.021s === RUN TestFileIdSize 24 14 --- PASS: TestFileIdSize (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/pb/filer_pb 0.014s ? github.com/seaweedfs/seaweedfs/weed/pb/iam_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/master_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/message_fbs [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/mount_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/mq_agent_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/mq_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/remote_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/s3_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/schema_pb [no test files] ? github.com/seaweedfs/seaweedfs/weed/pb/volume_server_pb [no test files] === RUN TestGjson { "quiz": { "sport": { "q1": { "question": "Which one is correct team name in NBA?", "options": [ "New York Bulls", "Los Angeles Kings", "Golden State Warriros", "Huston Rocket" ], "answer": "Huston Rocket" } }, "maths": { "q1": { "question": "5 + 7 = ?", "options": [ "10", "11", "12", "13" ], "answer": "12" }, "q2": { "question": "12 - 8 = ?", "options": [ "1", "2", "3", "4" ], "answer": "4" } } } } +++++++++++ 12 5 { "sport": { "q1": { "question": "Which one is correct team name in NBA?", "options": [ "New York Bulls", "Los Angeles Kings", "Golden State Warriros", "Huston Rocket" ], "answer": "Huston Rocket" } }, "maths": { "q1": { "question": "5 + 7 = ?", "options": [ "10", "11", "12", "13" ], "answer": "12" }, "q2": { "question": "12 - 8 = ?", "options": [ "1", "2", "3", "4" ], "answer": "4" } } } 0 0 ----------- { "fruit": "Apple", "size": "Large", "quiz": "Red" } +++++++++++ 51 3 Red 13 3 Apple ----------- --- PASS: TestGjson (0.00s) === RUN TestJsonQueryRow {fruit:"Bl\"ue",size:6} --- PASS: TestJsonQueryRow (0.00s) === RUN TestJsonQueryNumber {fruit:"Bl\"ue",quiz:"green"} --- PASS: TestJsonQueryNumber (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/query/json 0.010s ? github.com/seaweedfs/seaweedfs/weed/query/sqltypes [no test files] ? github.com/seaweedfs/seaweedfs/weed/remote_storage [no test files] ? github.com/seaweedfs/seaweedfs/weed/remote_storage/azure [no test files] ? github.com/seaweedfs/seaweedfs/weed/remote_storage/gcs [no test files] ? github.com/seaweedfs/seaweedfs/weed/remote_storage/s3 [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/repl_util [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink/azuresink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink/b2sink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink/filersink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink/gcssink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink/localsink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sink/s3sink [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/source [no test files] ? github.com/seaweedfs/seaweedfs/weed/replication/sub [no test files] === RUN TestIdentityListFileFormat { "identities": [ { "name": "some_name", "credentials": [ { "accessKey": "some_access_key1", "secretKey": "some_secret_key2" } ], "actions": [ "Admin", "Read", "Write" ], "account": null }, { "name": "some_read_only_user", "credentials": [ { "accessKey": "some_access_key1", "secretKey": "some_secret_key1" } ], "actions": [ "Read" ], "account": null }, { "name": "some_normal_user", "credentials": [ { "accessKey": "some_access_key2", "secretKey": "some_secret_key2" } ], "actions": [ "Read", "Write" ], "account": null } ], "accounts": [] } --- PASS: TestIdentityListFileFormat (0.00s) === RUN TestCanDo --- PASS: TestCanDo (0.00s) === RUN TestLoadS3ApiConfiguration --- PASS: TestLoadS3ApiConfiguration (0.00s) === RUN TestIsRequestPresignedSignatureV4 --- PASS: TestIsRequestPresignedSignatureV4 (0.00s) === RUN TestIsReqAuthenticated --- PASS: TestIsReqAuthenticated (0.00s) === RUN TestCheckaAnonymousRequestAuthType --- PASS: TestCheckaAnonymousRequestAuthType (0.00s) === RUN TestCheckAdminRequestAuthType --- PASS: TestCheckAdminRequestAuthType (0.00s) === RUN TestGetStringToSignPUT --- PASS: TestGetStringToSignPUT (0.00s) === RUN TestGetStringToSignGETEmptyStringHash --- PASS: TestGetStringToSignGETEmptyStringHash (0.00s) === RUN TestBuildBucketMetadata W0701 09:56:02.204823 bucket_metadata.go:106 Invalid ownership: , bucket: ownershipEmptyStr W0701 09:56:02.205144 bucket_metadata.go:117 owner[id=xxxxx] is invalid, bucket: acpEmptyObject --- PASS: TestBuildBucketMetadata (0.00s) === RUN TestGetBucketMetadata --- PASS: TestGetBucketMetadata (2.00s) === RUN TestNewSignV4ChunkedReaderstreamingAws4HmacSha256Payload --- PASS: TestNewSignV4ChunkedReaderstreamingAws4HmacSha256Payload (0.00s) === RUN TestNewSignV4ChunkedReaderStreamingUnsignedPayloadTrailer --- PASS: TestNewSignV4ChunkedReaderStreamingUnsignedPayloadTrailer (0.00s) === RUN TestInitiateMultipartUploadResult --- PASS: TestInitiateMultipartUploadResult (0.00s) === RUN TestListPartsResult --- PASS: TestListPartsResult (0.00s) === RUN Test_parsePartNumber === RUN Test_parsePartNumber/first === RUN Test_parsePartNumber/second --- PASS: Test_parsePartNumber (0.00s) --- PASS: Test_parsePartNumber/first (0.00s) --- PASS: Test_parsePartNumber/second (0.00s) === RUN TestGetAccountId --- PASS: TestGetAccountId (0.00s) === RUN TestExtractAcl --- PASS: TestExtractAcl (0.00s) === RUN TestParseAndValidateAclHeaders W0701 09:56:04.214274 s3api_acl_helper.go:292 invalid canonical grantee! account id[notExistsAccount] is not exists W0701 09:56:04.214281 s3api_acl_helper.go:281 invalid group grantee! group name[http:sfasf] is not valid --- PASS: TestParseAndValidateAclHeaders (0.00s) === RUN TestDetermineReqGrants --- PASS: TestDetermineReqGrants (0.00s) === RUN TestAssembleEntryWithAcp --- PASS: TestAssembleEntryWithAcp (0.00s) === RUN TestGrantEquals --- PASS: TestGrantEquals (0.00s) === RUN TestSetAcpOwnerHeader --- PASS: TestSetAcpOwnerHeader (0.00s) === RUN TestSetAcpGrantsHeader --- PASS: TestSetAcpGrantsHeader (0.00s) === RUN TestListBucketsHandler --- PASS: TestListBucketsHandler (0.00s) === RUN TestLimit --- PASS: TestLimit (0.00s) === RUN TestProcessMetadata --- PASS: TestProcessMetadata (0.00s) === RUN TestProcessMetadataBytes --- PASS: TestProcessMetadataBytes (0.01s) === RUN TestListObjectsHandler --- PASS: TestListObjectsHandler (0.00s) === RUN Test_normalizePrefixMarker === RUN Test_normalizePrefixMarker/prefix_is_a_directory === RUN Test_normalizePrefixMarker/normal_case === RUN Test_normalizePrefixMarker/empty_prefix === RUN Test_normalizePrefixMarker/empty_directory --- PASS: Test_normalizePrefixMarker (0.00s) --- PASS: Test_normalizePrefixMarker/prefix_is_a_directory (0.00s) --- PASS: Test_normalizePrefixMarker/normal_case (0.00s) --- PASS: Test_normalizePrefixMarker/empty_prefix (0.00s) --- PASS: Test_normalizePrefixMarker/empty_directory (0.00s) === RUN TestRemoveDuplicateSlashes === RUN TestRemoveDuplicateSlashes/empty === RUN TestRemoveDuplicateSlashes/slash === RUN TestRemoveDuplicateSlashes/object === RUN TestRemoveDuplicateSlashes/correct_path === RUN TestRemoveDuplicateSlashes/path_with_duplicates --- PASS: TestRemoveDuplicateSlashes (0.00s) --- PASS: TestRemoveDuplicateSlashes/empty (0.00s) --- PASS: TestRemoveDuplicateSlashes/slash (0.00s) --- PASS: TestRemoveDuplicateSlashes/object (0.00s) --- PASS: TestRemoveDuplicateSlashes/correct_path (0.00s) --- PASS: TestRemoveDuplicateSlashes/path_with_duplicates (0.00s) === RUN TestS3ApiServer_toFilerUrl === RUN TestS3ApiServer_toFilerUrl/simple === RUN TestS3ApiServer_toFilerUrl/double_prefix === RUN TestS3ApiServer_toFilerUrl/triple_prefix === RUN TestS3ApiServer_toFilerUrl/empty_prefix --- PASS: TestS3ApiServer_toFilerUrl (0.00s) --- PASS: TestS3ApiServer_toFilerUrl/simple (0.00s) --- PASS: TestS3ApiServer_toFilerUrl/double_prefix (0.00s) --- PASS: TestS3ApiServer_toFilerUrl/triple_prefix (0.00s) --- PASS: TestS3ApiServer_toFilerUrl/empty_prefix (0.00s) === RUN TestGetRequestDataReader_ChunkedEncodingWithoutIAM W0701 09:56:04.222562 auth_credentials.go:131 fail to load config: read S3 config: getOrCreateConnection : fail to dial : grpc: no transport security set (use grpc.WithTransportCredentials(insecure.NewCredentials()) explicitly or set credentials) === RUN TestGetRequestDataReader_ChunkedEncodingWithoutIAM/RegularRequest s3api_put_object_helper_test.go:80: Test case: RegularRequest - Regular requests without chunked encoding should pass through unchanged === RUN TestGetRequestDataReader_ChunkedEncodingWithoutIAM/StreamingSignedWithoutIAM s3api_put_object_helper_test.go:80: Test case: StreamingSignedWithoutIAM - Streaming signed requests should fail when IAM is disabled === RUN TestGetRequestDataReader_ChunkedEncodingWithoutIAM/StreamingUnsignedWithoutIAM s3api_put_object_helper_test.go:80: Test case: StreamingUnsignedWithoutIAM - Streaming unsigned requests should be processed even when IAM is disabled --- PASS: TestGetRequestDataReader_ChunkedEncodingWithoutIAM (0.00s) --- PASS: TestGetRequestDataReader_ChunkedEncodingWithoutIAM/RegularRequest (0.00s) --- PASS: TestGetRequestDataReader_ChunkedEncodingWithoutIAM/StreamingSignedWithoutIAM (0.00s) --- PASS: TestGetRequestDataReader_ChunkedEncodingWithoutIAM/StreamingUnsignedWithoutIAM (0.00s) === RUN TestGetRequestDataReader_AuthTypeDetection W0701 09:56:04.222666 auth_credentials.go:131 fail to load config: read S3 config: getOrCreateConnection : fail to dial : grpc: no transport security set (use grpc.WithTransportCredentials(insecure.NewCredentials()) explicitly or set credentials) === RUN TestGetRequestDataReader_AuthTypeDetection/ChunkedDataWithChecksum --- PASS: TestGetRequestDataReader_AuthTypeDetection (0.00s) --- PASS: TestGetRequestDataReader_AuthTypeDetection/ChunkedDataWithChecksum (0.00s) === RUN TestGetRequestDataReader_IAMEnabled W0701 09:56:04.222717 auth_credentials.go:131 fail to load config: read S3 config: getOrCreateConnection : fail to dial : grpc: no transport security set (use grpc.WithTransportCredentials(insecure.NewCredentials()) explicitly or set credentials) === RUN TestGetRequestDataReader_IAMEnabled/StreamingUnsignedWithIAMEnabled --- PASS: TestGetRequestDataReader_IAMEnabled (0.00s) --- PASS: TestGetRequestDataReader_IAMEnabled/StreamingUnsignedWithIAMEnabled (0.00s) === RUN TestAuthTypeDetection === RUN TestAuthTypeDetection/StreamingUnsigned === RUN TestAuthTypeDetection/StreamingSigned === RUN TestAuthTypeDetection/Regular --- PASS: TestAuthTypeDetection (0.00s) --- PASS: TestAuthTypeDetection/StreamingUnsigned (0.00s) --- PASS: TestAuthTypeDetection/StreamingSigned (0.00s) --- PASS: TestAuthTypeDetection/Regular (0.00s) === RUN TestCopyObjectResponse 2025-07-01T09:56:04.222804981Z12345678 --- PASS: TestCopyObjectResponse (0.00s) === RUN TestCopyPartResponse 2025-07-01T09:56:04.22284185Z12345678 --- PASS: TestCopyPartResponse (0.00s) === RUN TestXMLUnmarshall --- PASS: TestXMLUnmarshall (0.00s) === RUN TestXMLMarshall --- PASS: TestXMLMarshall (0.00s) === RUN TestValidateTags --- PASS: TestValidateTags (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/s3api 2.045s === RUN TestPostPolicyForm --- PASS: TestPostPolicyForm (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/s3api/policy 0.007s ? github.com/seaweedfs/seaweedfs/weed/s3api/s3_constants [no test files] === RUN Test_verifyBucketName --- PASS: Test_verifyBucketName (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/s3api/s3bucket 0.004s ? github.com/seaweedfs/seaweedfs/weed/s3api/s3err [no test files] ? github.com/seaweedfs/seaweedfs/weed/security [no test files] === RUN TestSequencer I0701 09:56:02.190351 snowflake_sequencer.go:21 use snowflake seq id generator, nodeid:for_test hex_of_nodeid: 1 1aec374d13401000 1aec374d13401001 1aec374d13401002 1aec374d13401003 1aec374d13401004 1aec374d13401005 1aec374d13401006 1aec374d13401007 1aec374d13401008 1aec374d13401009 1aec374d1340100a 1aec374d1340100b 1aec374d1340100c 1aec374d1340100d 1aec374d1340100e 1aec374d1340100f 1aec374d13401010 1aec374d13401011 1aec374d13401012 1aec374d13401013 1aec374d13401014 1aec374d13401015 1aec374d13401016 1aec374d13401017 1aec374d13401018 1aec374d13401019 1aec374d1340101a 1aec374d1340101b 1aec374d1340101c 1aec374d1340101d 1aec374d1340101e 1aec374d1340101f 1aec374d13401020 1aec374d13401021 1aec374d13401022 1aec374d13401023 1aec374d13401024 1aec374d13401025 1aec374d13401026 1aec374d13401027 1aec374d13401028 1aec374d13401029 1aec374d1340102a 1aec374d1340102b 1aec374d1340102c 1aec374d1340102d 1aec374d1340102e 1aec374d1340102f 1aec374d13401030 1aec374d13401031 1aec374d13401032 1aec374d13401033 1aec374d13401034 1aec374d13401035 1aec374d13401036 1aec374d13401037 1aec374d13401038 1aec374d13401039 1aec374d1340103a 1aec374d1340103b 1aec374d1340103c 1aec374d1340103d 1aec374d1340103e 1aec374d1340103f 1aec374d13401040 1aec374d13401041 1aec374d13401042 1aec374d13401043 1aec374d13401044 1aec374d13401045 1aec374d13401046 1aec374d13401047 1aec374d13401048 1aec374d13401049 1aec374d1340104a 1aec374d1340104b 1aec374d1340104c 1aec374d1340104d 1aec374d1340104e 1aec374d1340104f 1aec374d13401050 1aec374d13401051 1aec374d13401052 1aec374d13401053 1aec374d13401054 1aec374d13401055 1aec374d13401056 1aec374d13401057 1aec374d13401058 1aec374d13401059 1aec374d1340105a 1aec374d1340105b 1aec374d1340105c 1aec374d1340105d 1aec374d1340105e 1aec374d1340105f 1aec374d13401060 1aec374d13401061 1aec374d13401062 1aec374d13401063 --- PASS: TestSequencer (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/sequence 0.010s === RUN TestParseURL --- PASS: TestParseURL (0.00s) === RUN TestPtrie matched1 /topics/abc matched1 /topics/abc/d matched2 /topics/abc matched2 /topics/abc/d --- PASS: TestPtrie (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/server 0.028s ? github.com/seaweedfs/seaweedfs/weed/server/constants [no test files] === RUN TestToBreadcrumb === RUN TestToBreadcrumb/empty === RUN TestToBreadcrumb/test1 === RUN TestToBreadcrumb/test2 === RUN TestToBreadcrumb/test3 --- PASS: TestToBreadcrumb (0.00s) --- PASS: TestToBreadcrumb/empty (0.00s) --- PASS: TestToBreadcrumb/test1 (0.00s) --- PASS: TestToBreadcrumb/test2 (0.00s) --- PASS: TestToBreadcrumb/test3 (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/server/filer_ui 0.009s ? github.com/seaweedfs/seaweedfs/weed/server/master_ui [no test files] ? github.com/seaweedfs/seaweedfs/weed/server/volume_server_ui [no test files] ? github.com/seaweedfs/seaweedfs/weed/sftpd [no test files] ? github.com/seaweedfs/seaweedfs/weed/sftpd/auth [no test files] ? github.com/seaweedfs/seaweedfs/weed/sftpd/user [no test files] ? github.com/seaweedfs/seaweedfs/weed/sftpd/utils [no test files] === RUN TestCollectCollectionsForVolumeIds --- PASS: TestCollectCollectionsForVolumeIds (0.00s) === RUN TestParseReplicaPlacementArg using master default replica placement "123" for EC volumes using replica placement "021" for EC volumes --- PASS: TestParseReplicaPlacementArg (0.00s) === RUN TestEcDistribution => 192.168.1.5:8080 27010 => 192.168.1.6:8080 17420 => 192.168.1.1:8080 17330 => 192.168.1.4:8080 1900 => 192.168.1.2:8080 1540 --- PASS: TestEcDistribution (0.00s) === RUN TestPickRackToBalanceShardsInto --- PASS: TestPickRackToBalanceShardsInto (0.00s) === RUN TestPickEcNodeToBalanceShardsInto --- PASS: TestPickEcNodeToBalanceShardsInto (0.00s) === RUN TestCountFreeShardSlots === RUN TestCountFreeShardSlots/topology_#1,_free_HDD_shards === RUN TestCountFreeShardSlots/topology_#1,_no_free_SSD_shards_available === RUN TestCountFreeShardSlots/topology_#2,_no_negative_free_HDD_shards === RUN TestCountFreeShardSlots/topology_#2,_no_free_SSD_shards_available --- PASS: TestCountFreeShardSlots (0.00s) --- PASS: TestCountFreeShardSlots/topology_#1,_free_HDD_shards (0.00s) --- PASS: TestCountFreeShardSlots/topology_#1,_no_free_SSD_shards_available (0.00s) --- PASS: TestCountFreeShardSlots/topology_#2,_no_negative_free_HDD_shards (0.00s) --- PASS: TestCountFreeShardSlots/topology_#2,_no_free_SSD_shards_available (0.00s) === RUN TestCommandEcBalanceSmall balanceEcVolumes c1 dn1 moves ec shard 1.1 to dn2 dn1 moves ec shard 1.2 to dn2 dn1 moves ec shard 1.3 to dn2 dn1 moves ec shard 1.4 to dn2 dn1 moves ec shard 1.5 to dn2 dn1 moves ec shard 1.6 to dn2 dn1 moves ec shard 1.0 to dn2 dn2 moves ec shard 2.5 to dn1 dn2 moves ec shard 2.6 to dn1 dn2 moves ec shard 2.0 to dn1 dn2 moves ec shard 2.1 to dn1 dn2 moves ec shard 2.2 to dn1 dn2 moves ec shard 2.3 to dn1 dn2 moves ec shard 2.4 to dn1 --- PASS: TestCommandEcBalanceSmall (0.00s) === RUN TestCommandEcBalanceNothingToMove balanceEcVolumes c1 --- PASS: TestCommandEcBalanceNothingToMove (0.00s) === RUN TestCommandEcBalanceAddNewServers balanceEcVolumes c1 --- PASS: TestCommandEcBalanceAddNewServers (0.00s) === RUN TestCommandEcBalanceAddNewRacks balanceEcVolumes c1 dn2 moves ec shard 1.9 to dn3 dn2 moves ec shard 1.10 to dn4 dn1 moves ec shard 1.0 to dn3 dn2 moves ec shard 1.7 to dn4 dn2 moves ec shard 1.8 to dn3 dn1 moves ec shard 1.1 to dn4 dn1 moves ec shard 1.2 to dn3 dn1 moves ec shard 2.9 to dn4 dn2 moves ec shard 2.2 to dn3 dn2 moves ec shard 2.3 to dn4 dn1 moves ec shard 2.7 to dn3 dn2 moves ec shard 2.0 to dn4 dn2 moves ec shard 2.1 to dn3 dn1 moves ec shard 2.8 to dn4 --- PASS: TestCommandEcBalanceAddNewRacks (0.00s) === RUN TestCommandEcBalanceVolumeEvenButRackUneven balanceEcVolumes c1 dn_shared moves ec shards 1.0 to dn3 --- PASS: TestCommandEcBalanceVolumeEvenButRackUneven (0.00s) === RUN TestCircuitBreakerShell --- PASS: TestCircuitBreakerShell (0.00s) === RUN TestIsGoodMove replication: 100 expected false name: test 100 move to wrong data centers replication: 100 expected true name: test 100 move to spread into proper data centers replication: 001 expected false name: test move to the same node replication: 001 expected false name: test move to the same rack, but existing node replication: 001 expected true name: test move to the same rack, a new node replication: 010 expected false name: test 010 move all to the same rack replication: 010 expected true name: test 010 move to spread racks replication: 010 expected true name: test 010 move to spread racks replication: 011 expected true name: test 011 switch which rack has more replicas replication: 011 expected true name: test 011 move the lonely replica to another racks replication: 011 expected false name: test 011 move to wrong racks replication: 011 expected false name: test 011 move all to the same rack --- PASS: TestIsGoodMove (0.00s) === RUN TestBalance hdd 0.10 0.21:0.06 moving volume 31 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.20:0.06 moving volume 29 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.20:0.06 moving volume 30 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.20:0.06 moving volume 27 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.19:0.06 moving volume 28 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.19:0.06 moving volume collection4_7 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.19:0.06 moving volume collection0_25 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.18:0.06 moving volume collection3_9 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.18:0.06 moving volume collection1_80 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.18:0.06 moving volume collection1_69 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.18:0.06 moving volume 4 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.17:0.06 moving volume collection1_84 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.17:0.07 moving volume 2 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.17:0.07 moving volume collection1_63 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.17:0.07 moving volume 6 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.17:0.07 moving volume collection1_74 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.16:0.07 moving volume 3 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.16:0.07 moving volume collection1_85 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.16:0.07 moving volume collection1_54 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.16:0.07 moving volume collection1_81 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.15:0.07 moving volume collection1_97 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.15:0.07 moving volume collection1_56 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.15:0.07 moving volume collection1_174 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.15:0.07 moving volume collection2_380 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.15:0.07 moving volume collection1_105 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.14:0.07 moving volume collection1_215 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.14:0.07 moving volume collection0_24 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.14:0.07 moving volume collection1_173 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.14:0.07 moving volume collection1_107 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.07 moving volume 5 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_136 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_238 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_240 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection0_26 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_167 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_66 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_65 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_57 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_62 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_67 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_138 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_70 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_90 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_72 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_71 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_75 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_58 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_177 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.08 moving volume collection1_87 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.13:0.09 moving volume collection1_73 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_77 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_116 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_83 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_91 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_79 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_64 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_61 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_76 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_59 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_139 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_96 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_144 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_95 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_92 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_86 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_60 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.09 moving volume collection1_55 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection2_379 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_94 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_82 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_128 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_89 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_53 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection2_357 192.168.1.2:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_99 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.12:0.10 moving volume collection1_111 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection1_176 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection4_7 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection3_9 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_169 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume 1 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_197 192.168.1.4:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume 4 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume 2 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_126 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection2_381 192.168.1.2:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_165 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume 6 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume 3 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_232 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection0_25 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection2_345 192.168.1.4:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_135 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection1_68 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_117 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection1_74 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection2_378 192.168.1.1:8080 => 192.168.1.5:8080 hdd 0.10 0.11:0.10 moving volume collection1_194 192.168.1.1:8080 => 192.168.1.6:8080 hdd 0.10 0.11:0.10 moving volume collection1_179 192.168.1.2:8080 => 192.168.1.5:8080 --- PASS: TestBalance (0.01s) === RUN TestVolumeSelection collect volumes quiet for: 0 seconds --- PASS: TestVolumeSelection (0.00s) === RUN TestDeleteEmptySelection --- PASS: TestDeleteEmptySelection (0.00s) === RUN TestShouldSkipVolume --- PASS: TestShouldSkipVolume (0.00s) === RUN TestSatisfyReplicaPlacementComplicated replication: 100 expected false name: test 100 negative replication: 100 expected true name: test 100 positive replication: 022 expected true name: test 022 positive replication: 022 expected false name: test 022 negative replication: 210 expected true name: test 210 moved from 200 positive replication: 210 expected false name: test 210 moved from 200 negative extra dc replication: 210 expected false name: test 210 moved from 200 negative extra data node --- PASS: TestSatisfyReplicaPlacementComplicated (0.00s) === RUN TestSatisfyReplicaPlacement01x replication: 011 expected true name: test 011 same existing rack replication: 011 expected false name: test 011 negative replication: 011 expected true name: test 011 different existing racks replication: 011 expected false name: test 011 different existing racks negative --- PASS: TestSatisfyReplicaPlacement01x (0.00s) === RUN TestSatisfyReplicaPlacement00x replication: 001 expected true name: test 001 replication: 002 expected true name: test 002 positive replication: 002 expected false name: test 002 negative, repeat the same node replication: 002 expected false name: test 002 negative, enough node already --- PASS: TestSatisfyReplicaPlacement00x (0.00s) === RUN TestSatisfyReplicaPlacement100 replication: 100 expected true name: test 100 --- PASS: TestSatisfyReplicaPlacement100 (0.00s) === RUN TestMisplacedChecking replication: 001 expected true name: test 001 replication: 010 expected false name: test 010 replication: 011 expected false name: test 011 replication: 110 expected true name: test 110 replication: 100 expected true name: test 100 --- PASS: TestMisplacedChecking (0.00s) === RUN TestPickingMisplacedVolumeToDelete replication: 001 name: test 001 command_volume_fix_replication_test.go:435: test 001: picked dn2 001 replication: 100 name: test 100 command_volume_fix_replication_test.go:435: test 100: picked dn2 100 --- PASS: TestPickingMisplacedVolumeToDelete (0.00s) === RUN TestSatisfyReplicaCurrentLocation === RUN TestSatisfyReplicaCurrentLocation/test_001 === RUN TestSatisfyReplicaCurrentLocation/test_010 === RUN TestSatisfyReplicaCurrentLocation/test_011 === RUN TestSatisfyReplicaCurrentLocation/test_110 === RUN TestSatisfyReplicaCurrentLocation/test_100 --- PASS: TestSatisfyReplicaCurrentLocation (0.00s) --- PASS: TestSatisfyReplicaCurrentLocation/test_001 (0.00s) --- PASS: TestSatisfyReplicaCurrentLocation/test_010 (0.00s) --- PASS: TestSatisfyReplicaCurrentLocation/test_011 (0.00s) --- PASS: TestSatisfyReplicaCurrentLocation/test_110 (0.00s) --- PASS: TestSatisfyReplicaCurrentLocation/test_100 (0.00s) === RUN TestParsing --- PASS: TestParsing (0.08s) === RUN TestVolumeServerEvacuate moving volume collection0_15 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection0_21 192.168.1.4:8080 => 192.168.1.6:8080 moving volume collection0_22 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection0_23 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection0_24 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection0_25 192.168.1.4:8080 => 192.168.1.2:8080 moving volume 27 192.168.1.4:8080 => 192.168.1.2:8080 moving volume 28 192.168.1.4:8080 => 192.168.1.2:8080 moving volume 29 192.168.1.4:8080 => 192.168.1.2:8080 moving volume 30 192.168.1.4:8080 => 192.168.1.2:8080 moving volume 31 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_33 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_38 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_51 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_52 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_54 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_63 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_69 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_74 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_80 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_84 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_85 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_97 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_98 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_105 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_106 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_112 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_116 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_119 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_128 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_133 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_136 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_138 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_140 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_144 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_161 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_173 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_174 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_197 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection1_219 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_263 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_272 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_291 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_299 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_301 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_302 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_339 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_345 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_355 192.168.1.4:8080 => 192.168.1.2:8080 moving volume collection2_373 192.168.1.4:8080 => 192.168.1.2:8080 --- PASS: TestVolumeServerEvacuate (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/shell 0.236s === RUN TestRobinCounter --- PASS: TestRobinCounter (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/stats 0.008s === RUN TestUnUsedSpace --- PASS: TestUnUsedSpace (0.00s) === RUN TestFirstInvalidIndex I0701 09:56:02.196891 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.197100 volume_loading.go:157 loading memory index /tmp/TestFirstInvalidIndex3095463392/001/1.idx to memory --- PASS: TestFirstInvalidIndex (0.00s) === RUN TestFastLoadingNeedleMapMetrics I0701 09:56:02.210505 needle_map_metric_test.go:26 FileCount expected 10000 actual 11981 I0701 09:56:02.210519 needle_map_metric_test.go:27 DeletedSize expected 1669 actual 1669 I0701 09:56:02.210522 needle_map_metric_test.go:28 ContentSize expected 10000 actual 10000 I0701 09:56:02.210524 needle_map_metric_test.go:29 DeletedCount expected 1669 actual 3650 I0701 09:56:02.210526 needle_map_metric_test.go:30 MaxFileKey expected 10000 actual 10000 --- PASS: TestFastLoadingNeedleMapMetrics (0.01s) === RUN TestBinarySearch --- PASS: TestBinarySearch (0.00s) === RUN TestSortVolumeInfos --- PASS: TestSortVolumeInfos (0.00s) === RUN TestReadNeedMetaWithWritesAndUpdates I0701 09:56:02.210707 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.210717 volume_loading.go:157 loading memory index /tmp/TestReadNeedMetaWithWritesAndUpdates866643764/001/1.idx to memory --- PASS: TestReadNeedMetaWithWritesAndUpdates (0.00s) === RUN TestReadNeedMetaWithDeletesThenWrites I0701 09:56:02.211068 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.211074 volume_loading.go:157 loading memory index /tmp/TestReadNeedMetaWithDeletesThenWrites3188694118/001/1.idx to memory --- PASS: TestReadNeedMetaWithDeletesThenWrites (0.00s) === RUN TestMakeDiff --- PASS: TestMakeDiff (0.00s) === RUN TestMemIndexCompaction I0701 09:56:02.211342 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.211352 volume_loading.go:157 loading memory index /tmp/TestMemIndexCompaction3831390443/001/1.idx to memory I0701 09:56:02.343917 needle_map_memory.go:111 loading idx from offset 0 for file: /tmp/TestMemIndexCompaction3831390443/001/1.cpx volume_vacuum_test.go:92: compaction speed: 56057161.90 bytes/s I0701 09:56:02.437449 volume_vacuum.go:114 Committing volume 1 vacuuming... I0701 09:56:02.518271 needle_map_memory.go:111 loading idx from offset 9717 for file: /tmp/TestMemIndexCompaction3831390443/001/1.cpx I0701 09:56:02.524198 volume_loading.go:98 readSuperBlock volume 1 version 3 I0701 09:56:02.524218 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.524226 volume_loading.go:154 updating memory compact index /tmp/TestMemIndexCompaction3831390443/001/1.idx volume_vacuum_test.go:110: realRecordCount:29717, v.FileCount():29717 mm.DeletedCount():9828 I0701 09:56:02.524270 volume_loading.go:98 readSuperBlock volume 1 version 3 I0701 09:56:02.524277 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.524287 volume_loading.go:157 loading memory index /tmp/TestMemIndexCompaction3831390443/001/1.idx to memory --- PASS: TestMemIndexCompaction (0.34s) === RUN TestLDBIndexCompaction I0701 09:56:02.552468 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:02.552481 volume_loading.go:172 loading leveldb index /tmp/TestLDBIndexCompaction2361292799/001/1.ldb I0701 09:56:02.553785 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestLDBIndexCompaction2361292799/001/1.ldb, watermark 0, num of entries:0 I0701 09:56:02.554736 needle_map_leveldb.go:66 Loading /tmp/TestLDBIndexCompaction2361292799/001/1.ldb... , watermark: 0 I0701 09:56:02.885255 needle_map_leveldb.go:338 loading idx to leveldb from offset 0 for file: /tmp/TestLDBIndexCompaction2361292799/001/1.cpx volume_vacuum_test.go:92: compaction speed: 24610863.48 bytes/s I0701 09:56:03.443235 volume_vacuum.go:114 Committing volume 1 vacuuming... I0701 09:56:03.565631 needle_map_leveldb.go:338 loading idx to leveldb from offset 9724 for file: /tmp/TestLDBIndexCompaction2361292799/001/1.cpx I0701 09:56:03.685436 volume_loading.go:98 readSuperBlock volume 1 version 3 I0701 09:56:03.685466 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.685481 volume_loading.go:169 updating leveldb index /tmp/TestLDBIndexCompaction2361292799/001/1.ldb volume_vacuum_test.go:105: watermark from levelDB: 20000, realWatermark: 20000, nm.recordCount: 29724, realRecordCount:29724, fileCount=29724, deletedcount:9712 I0701 09:56:03.708501 volume_loading.go:98 readSuperBlock volume 1 version 3 I0701 09:56:03.708521 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.708535 volume_loading.go:172 loading leveldb index /tmp/TestLDBIndexCompaction2361292799/001/1.ldb I0701 09:56:03.710150 needle_map_leveldb.go:66 Loading /tmp/TestLDBIndexCompaction2361292799/001/1.ldb... , watermark: 20000 --- PASS: TestLDBIndexCompaction (1.24s) === RUN TestSearchVolumesWithDeletedNeedles I0701 09:56:03.792012 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.792024 volume_loading.go:157 loading memory index /tmp/TestSearchVolumesWithDeletedNeedles2651747654/001/1.idx to memory offset: 9960, isLast: false --- PASS: TestSearchVolumesWithDeletedNeedles (0.00s) === RUN TestDestroyEmptyVolumeWithOnlyEmpty I0701 09:56:03.792384 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.792396 volume_loading.go:157 loading memory index /tmp/TestDestroyEmptyVolumeWithOnlyEmpty77043775/001/1.idx to memory --- PASS: TestDestroyEmptyVolumeWithOnlyEmpty (0.00s) === RUN TestDestroyEmptyVolumeWithoutOnlyEmpty I0701 09:56:03.792612 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.792635 volume_loading.go:157 loading memory index /tmp/TestDestroyEmptyVolumeWithoutOnlyEmpty1871748140/001/1.idx to memory --- PASS: TestDestroyEmptyVolumeWithoutOnlyEmpty (0.00s) === RUN TestDestroyNonemptyVolumeWithOnlyEmpty I0701 09:56:03.792824 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.792828 volume_loading.go:157 loading memory index /tmp/TestDestroyNonemptyVolumeWithOnlyEmpty2426314703/001/1.idx to memory --- PASS: TestDestroyNonemptyVolumeWithOnlyEmpty (0.00s) === RUN TestDestroyNonemptyVolumeWithoutOnlyEmpty I0701 09:56:03.793012 volume_loading.go:139 checking volume data integrity for volume 1 I0701 09:56:03.793021 volume_loading.go:157 loading memory index /tmp/TestDestroyNonemptyVolumeWithoutOnlyEmpty2745812627/001/1.idx to memory --- PASS: TestDestroyNonemptyVolumeWithoutOnlyEmpty (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage 1.614s ? github.com/seaweedfs/seaweedfs/weed/storage/backend [no test files] === RUN TestMemoryMapMaxSizeReadWrite --- PASS: TestMemoryMapMaxSizeReadWrite (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage/backend/memory_map 0.001s ? github.com/seaweedfs/seaweedfs/weed/storage/backend/rclone_backend [no test files] ? github.com/seaweedfs/seaweedfs/weed/storage/backend/s3_backend [no test files] === RUN TestEncodingDecoding I0701 09:56:03.826048 ec_encoder.go:81 encodeDatFile 1.dat size:2590912 --- PASS: TestEncodingDecoding (0.26s) === RUN TestLocateData [{BlockIndex:5 InnerBlockOffset:100 Size:9900 IsLargeBlock:true LargeBlockRowsCount:1} {BlockIndex:6 InnerBlockOffset:0 Size:10000 IsLargeBlock:true LargeBlockRowsCount:1} {BlockIndex:7 InnerBlockOffset:0 Size:10000 IsLargeBlock:true LargeBlockRowsCount:1} {BlockIndex:8 InnerBlockOffset:0 Size:10000 IsLargeBlock:true LargeBlockRowsCount:1} {BlockIndex:9 InnerBlockOffset:0 Size:10000 IsLargeBlock:true LargeBlockRowsCount:1} {BlockIndex:0 InnerBlockOffset:0 Size:1 IsLargeBlock:false LargeBlockRowsCount:1}] --- PASS: TestLocateData (0.00s) === RUN TestLocateData2 --- PASS: TestLocateData2 (0.00s) === RUN TestLocateData3 {BlockIndex:8876 InnerBlockOffset:912752 Size:112568 IsLargeBlock:false LargeBlockRowsCount:2} --- PASS: TestLocateData3 (0.00s) === RUN TestPositioning offset: 31300679656 size: 1167 offset: 11513014944 size: 66044 offset: 26311863528 size: 26823 interval: {BlockIndex:14852 InnerBlockOffset:994536 Size:26856 IsLargeBlock:false LargeBlockRowsCount:1}, shardId: 2, shardOffset: 2631871720 --- PASS: TestPositioning (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage/erasure_coding 0.273s ? github.com/seaweedfs/seaweedfs/weed/storage/idx [no test files] === RUN TestParseFileIdFromString --- PASS: TestParseFileIdFromString (0.00s) === RUN TestParseKeyHash --- PASS: TestParseKeyHash (0.00s) === RUN TestAppend --- PASS: TestAppend (0.00s) === RUN TestWriteNeedle_CompatibilityWithLegacy === RUN TestWriteNeedle_CompatibilityWithLegacy/Version1 === RUN TestWriteNeedle_CompatibilityWithLegacy/Version2 === RUN TestWriteNeedle_CompatibilityWithLegacy/Version3 --- PASS: TestWriteNeedle_CompatibilityWithLegacy (0.00s) --- PASS: TestWriteNeedle_CompatibilityWithLegacy/Version1 (0.00s) --- PASS: TestWriteNeedle_CompatibilityWithLegacy/Version2 (0.00s) --- PASS: TestWriteNeedle_CompatibilityWithLegacy/Version3 (0.00s) === RUN TestNewVolumeId volume_id_test.go:11: a is not legal volume id, strconv.ParseUint: parsing "a": invalid syntax --- PASS: TestNewVolumeId (0.00s) === RUN TestVolumeId_String --- PASS: TestVolumeId_String (0.00s) === RUN TestVolumeId_Next --- PASS: TestVolumeId_Next (0.00s) === RUN TestTTLReadWrite --- PASS: TestTTLReadWrite (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage/needle 0.006s === RUN TestMemoryUsage Each 12.92 Bytes Alloc = 20 MiB TotalAlloc = 46 MiB Sys = 37 MiB NumGC = 8 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 1.220322102s Each 12.70 Bytes Alloc = 41 MiB TotalAlloc = 92 MiB Sys = 65 MiB NumGC = 10 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 1.62411105s Each 12.63 Bytes Alloc = 61 MiB TotalAlloc = 137 MiB Sys = 89 MiB NumGC = 12 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 1.155560628s Each 12.60 Bytes Alloc = 81 MiB TotalAlloc = 183 MiB Sys = 118 MiB NumGC = 13 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 1.186372152s Each 12.57 Bytes Alloc = 101 MiB TotalAlloc = 228 MiB Sys = 138 MiB NumGC = 14 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 959.781458ms Each 12.56 Bytes Alloc = 121 MiB TotalAlloc = 274 MiB Sys = 158 MiB NumGC = 15 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 807.0679ms Each 12.55 Bytes Alloc = 142 MiB TotalAlloc = 319 MiB Sys = 178 MiB NumGC = 16 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 806.084817ms Each 12.54 Bytes Alloc = 162 MiB TotalAlloc = 365 MiB Sys = 198 MiB NumGC = 17 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 843.729833ms Each 12.54 Bytes Alloc = 182 MiB TotalAlloc = 410 MiB Sys = 219 MiB NumGC = 18 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 952.451746ms Each 12.53 Bytes Alloc = 202 MiB TotalAlloc = 456 MiB Sys = 239 MiB NumGC = 19 CompactMap = 1539351/1749659 elements on 2198 segments, 87.98% efficiency Taken = 933.36874ms --- PASS: TestMemoryUsage (10.49s) === RUN TestSegmentBsearchKey === RUN TestSegmentBsearchKey/empty_segment === RUN TestSegmentBsearchKey/new_key,_insert_at_beggining === RUN TestSegmentBsearchKey/new_key,_insert_at_end === RUN TestSegmentBsearchKey/new_key,_insert_second === RUN TestSegmentBsearchKey/new_key,_insert_in_middle === RUN TestSegmentBsearchKey/key_#1 === RUN TestSegmentBsearchKey/key_#2 === RUN TestSegmentBsearchKey/key_#3 === RUN TestSegmentBsearchKey/key_#4 === RUN TestSegmentBsearchKey/key_#5 --- PASS: TestSegmentBsearchKey (0.00s) --- PASS: TestSegmentBsearchKey/empty_segment (0.00s) --- PASS: TestSegmentBsearchKey/new_key,_insert_at_beggining (0.00s) --- PASS: TestSegmentBsearchKey/new_key,_insert_at_end (0.00s) --- PASS: TestSegmentBsearchKey/new_key,_insert_second (0.00s) --- PASS: TestSegmentBsearchKey/new_key,_insert_in_middle (0.00s) --- PASS: TestSegmentBsearchKey/key_#1 (0.00s) --- PASS: TestSegmentBsearchKey/key_#2 (0.00s) --- PASS: TestSegmentBsearchKey/key_#3 (0.00s) --- PASS: TestSegmentBsearchKey/key_#4 (0.00s) --- PASS: TestSegmentBsearchKey/key_#5 (0.00s) === RUN TestSegmentSet --- PASS: TestSegmentSet (0.00s) === RUN TestSegmentSetOrdering --- PASS: TestSegmentSetOrdering (0.06s) === RUN TestSegmentGet === RUN TestSegmentGet/invalid_key === RUN TestSegmentGet/key_#1 === RUN TestSegmentGet/key_#2 === RUN TestSegmentGet/key_#3 --- PASS: TestSegmentGet (0.00s) --- PASS: TestSegmentGet/invalid_key (0.00s) --- PASS: TestSegmentGet/key_#1 (0.00s) --- PASS: TestSegmentGet/key_#2 (0.00s) --- PASS: TestSegmentGet/key_#3 (0.00s) === RUN TestSegmentDelete --- PASS: TestSegmentDelete (0.00s) === RUN TestSegmentForKey === RUN TestSegmentForKey/first_segment === RUN TestSegmentForKey/second_segment,_gapless === RUN TestSegmentForKey/gapped_segment --- PASS: TestSegmentForKey (0.00s) --- PASS: TestSegmentForKey/first_segment (0.00s) --- PASS: TestSegmentForKey/second_segment,_gapless (0.00s) --- PASS: TestSegmentForKey/gapped_segment (0.00s) === RUN TestAscendingVisit --- PASS: TestAscendingVisit (0.00s) === RUN TestRandomInsert --- PASS: TestRandomInsert (0.56s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage/needle_map 11.110s === RUN TestMemoryUsage Each 15.11 Bytes Alloc = 24 MiB TotalAlloc = 114 MiB Sys = 45 MiB NumGC = 16 Taken = 1.276111497s Each 14.90 Bytes Alloc = 48 MiB TotalAlloc = 227 MiB Sys = 81 MiB NumGC = 20 Taken = 1.686608462s Each 14.82 Bytes Alloc = 71 MiB TotalAlloc = 341 MiB Sys = 125 MiB NumGC = 23 Taken = 1.163931801s Each 14.79 Bytes Alloc = 95 MiB TotalAlloc = 454 MiB Sys = 157 MiB NumGC = 25 Taken = 1.239403767s Each 14.76 Bytes Alloc = 119 MiB TotalAlloc = 568 MiB Sys = 197 MiB NumGC = 27 Taken = 893.586806ms Each 14.75 Bytes Alloc = 143 MiB TotalAlloc = 681 MiB Sys = 246 MiB NumGC = 28 Taken = 808.210401ms Each 14.74 Bytes Alloc = 166 MiB TotalAlloc = 795 MiB Sys = 270 MiB NumGC = 29 Taken = 783.896989ms Each 14.73 Bytes Alloc = 190 MiB TotalAlloc = 908 MiB Sys = 294 MiB NumGC = 30 Taken = 855.710013ms Each 14.73 Bytes Alloc = 214 MiB TotalAlloc = 1022 MiB Sys = 322 MiB NumGC = 31 Taken = 1.008604232s Each 14.72 Bytes Alloc = 238 MiB TotalAlloc = 1135 MiB Sys = 346 MiB NumGC = 32 Taken = 891.526934ms --- PASS: TestMemoryUsage (10.61s) === RUN TestSnowflakeSequencer I0701 09:56:14.431050 snowflake_sequencer.go:21 use snowflake seq id generator, nodeid:for_test hex_of_nodeid: 1 --- PASS: TestSnowflakeSequencer (0.05s) === RUN TestOverflow2 needle key: 150073 needle key: 150076 needle key: 150088 needle key: 150089 needle key: 150124 needle key: 150137 needle key: 150145 needle key: 150147 needle key: 150158 needle key: 150162 --- PASS: TestOverflow2 (0.00s) === RUN TestIssue52 key 10002 ok true 10002, 1250, 10002 key 10002 ok true 10002, 1250, 10002 --- PASS: TestIssue52 (0.00s) === RUN TestCompactMap --- PASS: TestCompactMap (0.05s) === RUN TestOverflow overflow[ 0 ]: 1 overflow[ 1 ]: 2 overflow[ 2 ]: 3 overflow[ 3 ]: 4 overflow[ 4 ]: 5 overflow[ 0 ]: 1 size -12 overflow[ 1 ]: 2 size 12 overflow[ 2 ]: 3 size 24 overflow[ 3 ]: 4 size -12 overflow[ 4 ]: 5 size 12 overflow[ 0 ]: 1 overflow[ 1 ]: 2 overflow[ 2 ]: 3 overflow[ 3 ]: 4 overflow[ 4 ]: 5 overflow[ 0 ]: 1 overflow[ 1 ]: 2 overflow[ 2 ]: 3 overflow[ 3 ]: 4 overflow[ 4 ]: 5 --- PASS: TestOverflow (0.00s) === RUN TestCompactSection_Get compact_map_test.go:204: 1574318345753513987 compact_map_test.go:215: 1574318350048481283 --- PASS: TestCompactSection_Get (0.78s) === RUN TestCompactSection_PutOutOfOrderItemBeyondLookBackWindow --- PASS: TestCompactSection_PutOutOfOrderItemBeyondLookBackWindow (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage/needle_map/old 11.490s === RUN TestReplicaPlacementSerialDeserial --- PASS: TestReplicaPlacementSerialDeserial (0.00s) === RUN TestReplicaPlacementHasReplication === RUN TestReplicaPlacementHasReplication/empty_replica_placement === RUN TestReplicaPlacementHasReplication/no_replication === RUN TestReplicaPlacementHasReplication/same_rack_replication === RUN TestReplicaPlacementHasReplication/diff_rack_replication === RUN TestReplicaPlacementHasReplication/DC_replication === RUN TestReplicaPlacementHasReplication/full_replication --- PASS: TestReplicaPlacementHasReplication (0.00s) --- PASS: TestReplicaPlacementHasReplication/empty_replica_placement (0.00s) --- PASS: TestReplicaPlacementHasReplication/no_replication (0.00s) --- PASS: TestReplicaPlacementHasReplication/same_rack_replication (0.00s) --- PASS: TestReplicaPlacementHasReplication/diff_rack_replication (0.00s) --- PASS: TestReplicaPlacementHasReplication/DC_replication (0.00s) --- PASS: TestReplicaPlacementHasReplication/full_replication (0.00s) === RUN TestSuperBlockReadWrite --- PASS: TestSuperBlockReadWrite (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/storage/super_block 0.006s ? github.com/seaweedfs/seaweedfs/weed/storage/types [no test files] ? github.com/seaweedfs/seaweedfs/weed/storage/volume_info [no test files] === RUN TestRemoveDataCenter data: map[dc1:map[rack1:map[server111:map[limit:3 volumes:[map[id:1 size:12312] map[id:2 size:12312] map[id:3 size:12312]]] server112:map[limit:10 volumes:[map[id:4 size:12312] map[id:5 size:12312] map[id:6 size:12312]]]] rack2:map[server121:map[limit:4 volumes:[map[id:4 size:12312] map[id:5 size:12312] map[id:6 size:12312]]] server122:map[limit:4 volumes:[]] server123:map[limit:5 volumes:[map[id:2 size:12312] map[id:3 size:12312] map[id:4 size:12312]]]]] dc2:map[] dc3:map[rack2:map[server321:map[limit:4 volumes:[map[id:1 size:12312] map[id:3 size:12312] map[id:5 size:12312]]]]]] I0701 09:56:04.064557 node.go:250 weedfs adds child dc1 I0701 09:56:04.064739 node.go:250 weedfs:dc1 adds child rack2 I0701 09:56:04.064743 node.go:250 weedfs:dc1:rack2 adds child server123 I0701 09:56:04.064748 node.go:250 weedfs:dc1:rack2:server123 adds child I0701 09:56:04.064755 node.go:250 weedfs:dc1:rack2 adds child server121 I0701 09:56:04.064758 node.go:250 weedfs:dc1:rack2:server121 adds child I0701 09:56:04.064761 node.go:250 weedfs:dc1:rack2 adds child server122 I0701 09:56:04.064763 node.go:250 weedfs:dc1:rack2:server122 adds child I0701 09:56:04.064766 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.064769 node.go:250 weedfs:dc1:rack1 adds child server111 I0701 09:56:04.064771 node.go:250 weedfs:dc1:rack1:server111 adds child I0701 09:56:04.064775 node.go:250 weedfs:dc1:rack1 adds child server112 I0701 09:56:04.064777 node.go:250 weedfs:dc1:rack1:server112 adds child I0701 09:56:04.064781 node.go:250 weedfs adds child dc2 I0701 09:56:04.064783 node.go:250 weedfs adds child dc3 I0701 09:56:04.064785 node.go:250 weedfs:dc3 adds child rack2 I0701 09:56:04.064787 node.go:250 weedfs:dc3:rack2 adds child server321 I0701 09:56:04.064790 node.go:250 weedfs:dc3:rack2:server321 adds child I0701 09:56:04.064795 node.go:264 weedfs removes dc2 I0701 09:56:04.064798 node.go:264 weedfs removes dc3 --- PASS: TestRemoveDataCenter (0.00s) === RUN TestHandlingVolumeServerHeartbeat I0701 09:56:04.064819 node.go:250 weedfs adds child dc1 I0701 09:56:04.064822 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.064826 node.go:250 weedfs:dc1:rack1 adds child 127.0.0.1:34534 I0701 09:56:04.064829 node.go:250 weedfs:dc1:rack1:127.0.0.1:34534 adds child ssd I0701 09:56:04.064833 node.go:250 weedfs:dc1:rack1:127.0.0.1:34534 adds child I0701 09:56:04.064862 volume_layout.go:417 Volume 1 becomes writable I0701 09:56:04.064867 volume_layout.go:417 Volume 2 becomes writable I0701 09:56:04.064870 volume_layout.go:417 Volume 3 becomes writable I0701 09:56:04.064873 volume_layout.go:417 Volume 4 becomes writable I0701 09:56:04.064880 volume_layout.go:417 Volume 5 becomes writable I0701 09:56:04.064882 volume_layout.go:417 Volume 6 becomes writable I0701 09:56:04.064885 volume_layout.go:417 Volume 7 becomes writable I0701 09:56:04.064888 volume_layout.go:417 Volume 8 becomes writable I0701 09:56:04.064890 volume_layout.go:417 Volume 9 becomes writable I0701 09:56:04.064892 volume_layout.go:417 Volume 10 becomes writable I0701 09:56:04.064895 volume_layout.go:417 Volume 11 becomes writable I0701 09:56:04.064897 volume_layout.go:417 Volume 12 becomes writable I0701 09:56:04.064900 volume_layout.go:417 Volume 13 becomes writable I0701 09:56:04.064902 volume_layout.go:417 Volume 14 becomes writable I0701 09:56:04.064912 data_node.go:81 Deleting volume id: 8 I0701 09:56:04.064915 data_node.go:81 Deleting volume id: 9 I0701 09:56:04.064917 data_node.go:81 Deleting volume id: 10 I0701 09:56:04.064920 data_node.go:81 Deleting volume id: 11 I0701 09:56:04.064923 data_node.go:81 Deleting volume id: 12 I0701 09:56:04.064927 data_node.go:81 Deleting volume id: 13 I0701 09:56:04.064929 data_node.go:81 Deleting volume id: 14 I0701 09:56:04.064931 data_node.go:81 Deleting volume id: 7 I0701 09:56:04.064936 topology.go:329 removing volume info: Id:8, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.064950 volume_layout.go:229 volume 8 does not have enough copies I0701 09:56:04.064953 volume_layout.go:237 volume 8 remove from writable I0701 09:56:04.064956 volume_layout.go:405 Volume 8 becomes unwritable I0701 09:56:04.064959 topology.go:329 removing volume info: Id:9, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.064965 volume_layout.go:229 volume 9 does not have enough copies I0701 09:56:04.064967 volume_layout.go:237 volume 9 remove from writable I0701 09:56:04.064969 volume_layout.go:405 Volume 9 becomes unwritable I0701 09:56:04.064971 topology.go:329 removing volume info: Id:10, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.064975 volume_layout.go:229 volume 10 does not have enough copies I0701 09:56:04.064976 volume_layout.go:237 volume 10 remove from writable I0701 09:56:04.064979 volume_layout.go:405 Volume 10 becomes unwritable I0701 09:56:04.064981 topology.go:329 removing volume info: Id:11, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.064985 volume_layout.go:229 volume 11 does not have enough copies I0701 09:56:04.064988 volume_layout.go:237 volume 11 remove from writable I0701 09:56:04.064992 volume_layout.go:405 Volume 11 becomes unwritable I0701 09:56:04.065002 topology.go:329 removing volume info: Id:12, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.065005 volume_layout.go:229 volume 12 does not have enough copies I0701 09:56:04.065007 volume_layout.go:237 volume 12 remove from writable I0701 09:56:04.065009 volume_layout.go:405 Volume 12 becomes unwritable I0701 09:56:04.065011 topology.go:329 removing volume info: Id:13, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.065015 volume_layout.go:229 volume 13 does not have enough copies I0701 09:56:04.065016 volume_layout.go:237 volume 13 remove from writable I0701 09:56:04.065018 volume_layout.go:405 Volume 13 becomes unwritable I0701 09:56:04.065020 topology.go:329 removing volume info: Id:14, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.065023 volume_layout.go:229 volume 14 does not have enough copies I0701 09:56:04.065025 volume_layout.go:237 volume 14 remove from writable I0701 09:56:04.065027 volume_layout.go:405 Volume 14 becomes unwritable I0701 09:56:04.065031 topology.go:329 removing volume info: Id:7, Size:25432, ReplicaPlacement:000, Collection:, Version:3, FileCount:2343, DeleteCount:345, DeletedByteCount:34524, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.065034 volume_layout.go:229 volume 7 does not have enough copies I0701 09:56:04.065036 volume_layout.go:237 volume 7 remove from writable I0701 09:56:04.065038 volume_layout.go:405 Volume 7 becomes unwritable I0701 09:56:04.065044 topology.go:329 removing volume info: Id:3, Size:0, ReplicaPlacement:000, Collection:, Version:3, FileCount:0, DeleteCount:0, DeletedByteCount:0, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.065047 volume_layout.go:229 volume 3 does not have enough copies I0701 09:56:04.065049 volume_layout.go:237 volume 3 remove from writable I0701 09:56:04.065051 volume_layout.go:405 Volume 3 becomes unwritable I0701 09:56:04.065055 volume_layout.go:417 Volume 3 becomes writable after add volume id 6 after add volume id 1 after add volume id 2 after add volume id 3 after add volume id 4 after add volume id 5 after add writable volume id 1 after add writable volume id 2 after add writable volume id 4 after add writable volume id 5 after add writable volume id 6 after add writable volume id 3 I0701 09:56:04.065095 topology_event_handling.go:86 Removing Volume 5 from the dead volume server 127.0.0.1:34534 I0701 09:56:04.065105 volume_layout.go:456 Volume 5 has 0 replica, less than required 1 I0701 09:56:04.065107 volume_layout.go:405 Volume 5 becomes unwritable I0701 09:56:04.065109 topology_event_handling.go:86 Removing Volume 6 from the dead volume server 127.0.0.1:34534 I0701 09:56:04.065112 volume_layout.go:456 Volume 6 has 0 replica, less than required 1 I0701 09:56:04.065114 volume_layout.go:405 Volume 6 becomes unwritable I0701 09:56:04.065116 topology_event_handling.go:86 Removing Volume 1 from the dead volume server 127.0.0.1:34534 I0701 09:56:04.065118 volume_layout.go:456 Volume 1 has 0 replica, less than required 1 I0701 09:56:04.065120 volume_layout.go:405 Volume 1 becomes unwritable I0701 09:56:04.065122 topology_event_handling.go:86 Removing Volume 2 from the dead volume server 127.0.0.1:34534 I0701 09:56:04.065124 volume_layout.go:456 Volume 2 has 0 replica, less than required 1 I0701 09:56:04.065127 volume_layout.go:405 Volume 2 becomes unwritable I0701 09:56:04.065130 topology_event_handling.go:86 Removing Volume 3 from the dead volume server 127.0.0.1:34534 I0701 09:56:04.065132 volume_layout.go:456 Volume 3 has 0 replica, less than required 1 I0701 09:56:04.065135 volume_layout.go:405 Volume 3 becomes unwritable I0701 09:56:04.065137 topology_event_handling.go:86 Removing Volume 4 from the dead volume server 127.0.0.1:34534 I0701 09:56:04.065140 volume_layout.go:456 Volume 4 has 0 replica, less than required 1 I0701 09:56:04.065142 volume_layout.go:405 Volume 4 becomes unwritable I0701 09:56:04.065149 node.go:264 weedfs:dc1:rack1 removes 127.0.0.1:34534 --- PASS: TestHandlingVolumeServerHeartbeat (0.00s) === RUN TestAddRemoveVolume I0701 09:56:04.065168 node.go:250 weedfs adds child dc1 I0701 09:56:04.065172 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.065174 node.go:250 weedfs:dc1:rack1 adds child 127.0.0.1:34534 I0701 09:56:04.065178 node.go:250 weedfs:dc1:rack1:127.0.0.1:34534 adds child I0701 09:56:04.065182 node.go:250 weedfs:dc1:rack1:127.0.0.1:34534 adds child ssd I0701 09:56:04.065193 volume_layout.go:417 Volume 1 becomes writable I0701 09:56:04.065197 topology.go:329 removing volume info: Id:1, Size:100, ReplicaPlacement:000, Collection:xcollection, Version:3, FileCount:123, DeleteCount:23, DeletedByteCount:45, ReadOnly:false from 127.0.0.1:34534 I0701 09:56:04.065202 volume_layout.go:229 volume 1 does not have enough copies I0701 09:56:04.065204 volume_layout.go:237 volume 1 remove from writable I0701 09:56:04.065206 volume_layout.go:405 Volume 1 becomes unwritable --- PASS: TestAddRemoveVolume (0.00s) === RUN TestListCollections I0701 09:56:04.065226 node.go:250 weedfs adds child dc1 I0701 09:56:04.065229 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.065233 node.go:250 weedfs:dc1:rack1 adds child 127.0.0.1:34534 I0701 09:56:04.065241 volume_layout.go:229 volume 1111 does not have enough copies I0701 09:56:04.065244 volume_layout.go:237 volume 1111 remove from writable I0701 09:56:04.065248 volume_layout.go:229 volume 2222 does not have enough copies I0701 09:56:04.065252 volume_layout.go:237 volume 2222 remove from writable I0701 09:56:04.065256 volume_layout.go:229 volume 3333 does not have enough copies I0701 09:56:04.065258 volume_layout.go:237 volume 3333 remove from writable === RUN TestListCollections/no_volume_types_selected === RUN TestListCollections/normal_volumes === RUN TestListCollections/EC_volumes === RUN TestListCollections/normal_+_EC_volumes --- PASS: TestListCollections (0.00s) --- PASS: TestListCollections/no_volume_types_selected (0.00s) --- PASS: TestListCollections/normal_volumes (0.00s) --- PASS: TestListCollections/EC_volumes (0.00s) --- PASS: TestListCollections/normal_+_EC_volumes (0.00s) === RUN TestFindEmptySlotsForOneVolume data: map[dc1:map[rack1:map[server111:map[limit:3 volumes:[map[id:1 size:12312] map[id:2 size:12312] map[id:3 size:12312]]] server112:map[limit:10 volumes:[map[id:4 size:12312] map[id:5 size:12312] map[id:6 size:12312]]]] rack2:map[server121:map[limit:4 volumes:[map[id:4 size:12312] map[id:5 size:12312] map[id:6 size:12312]]] server122:map[limit:4 volumes:[]] server123:map[limit:5 volumes:[map[id:2 size:12312] map[id:3 size:12312] map[id:4 size:12312]]]]] dc2:map[] dc3:map[rack2:map[server321:map[limit:4 volumes:[map[id:1 size:12312] map[id:3 size:12312] map[id:5 size:12312]]]]]] I0701 09:56:04.065399 node.go:250 weedfs adds child dc2 I0701 09:56:04.065402 node.go:250 weedfs adds child dc3 I0701 09:56:04.065404 node.go:250 weedfs:dc3 adds child rack2 I0701 09:56:04.065406 node.go:250 weedfs:dc3:rack2 adds child server321 I0701 09:56:04.065408 node.go:250 weedfs:dc3:rack2:server321 adds child I0701 09:56:04.065413 node.go:250 weedfs adds child dc1 I0701 09:56:04.065415 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.065417 node.go:250 weedfs:dc1:rack1 adds child server111 I0701 09:56:04.065419 node.go:250 weedfs:dc1:rack1:server111 adds child I0701 09:56:04.065423 node.go:250 weedfs:dc1:rack1 adds child server112 I0701 09:56:04.065425 node.go:250 weedfs:dc1:rack1:server112 adds child I0701 09:56:04.065429 node.go:250 weedfs:dc1 adds child rack2 I0701 09:56:04.065431 node.go:250 weedfs:dc1:rack2 adds child server121 I0701 09:56:04.065433 node.go:250 weedfs:dc1:rack2:server121 adds child I0701 09:56:04.065437 node.go:250 weedfs:dc1:rack2 adds child server122 I0701 09:56:04.065439 node.go:250 weedfs:dc1:rack2:server122 adds child I0701 09:56:04.065442 node.go:250 weedfs:dc1:rack2 adds child server123 I0701 09:56:04.065444 node.go:250 weedfs:dc1:rack2:server123 adds child assigned node : server122 assigned node : server123 assigned node : server121 --- PASS: TestFindEmptySlotsForOneVolume (0.00s) === RUN TestReplication011 data: map[dc1:map[rack1:map[server111:map[limit:300 volumes:[map[id:1 size:12312] map[id:2 size:12312] map[id:3 size:12312]]] server112:map[limit:300 volumes:[map[id:4 size:12312] map[id:5 size:12312] map[id:6 size:12312]]] server113:map[limit:300 volumes:[]] server114:map[limit:300 volumes:[]] server115:map[limit:300 volumes:[]] server116:map[limit:300 volumes:[]]] rack2:map[server121:map[limit:300 volumes:[map[id:4 size:12312] map[id:5 size:12312] map[id:6 size:12312]]] server122:map[limit:300 volumes:[]] server123:map[limit:300 volumes:[map[id:2 size:12312] map[id:3 size:12312] map[id:4 size:12312]]] server124:map[limit:300 volumes:[]] server125:map[limit:300 volumes:[]] server126:map[limit:300 volumes:[]]] rack3:map[server131:map[limit:300 volumes:[]] server132:map[limit:300 volumes:[]] server133:map[limit:300 volumes:[]] server134:map[limit:300 volumes:[]] server135:map[limit:300 volumes:[]] server136:map[limit:300 volumes:[]]]]] I0701 09:56:04.065542 node.go:250 weedfs adds child dc1 I0701 09:56:04.065545 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.065547 node.go:250 weedfs:dc1:rack1 adds child server111 I0701 09:56:04.065550 node.go:250 weedfs:dc1:rack1:server111 adds child I0701 09:56:04.065554 node.go:250 weedfs:dc1:rack1 adds child server112 I0701 09:56:04.065557 node.go:250 weedfs:dc1:rack1:server112 adds child I0701 09:56:04.065566 node.go:250 weedfs:dc1:rack1 adds child server113 I0701 09:56:04.065568 node.go:250 weedfs:dc1:rack1:server113 adds child I0701 09:56:04.065571 node.go:250 weedfs:dc1:rack1 adds child server114 I0701 09:56:04.065574 node.go:250 weedfs:dc1:rack1:server114 adds child I0701 09:56:04.065577 node.go:250 weedfs:dc1:rack1 adds child server115 I0701 09:56:04.065579 node.go:250 weedfs:dc1:rack1:server115 adds child I0701 09:56:04.065583 node.go:250 weedfs:dc1:rack1 adds child server116 I0701 09:56:04.065585 node.go:250 weedfs:dc1:rack1:server116 adds child I0701 09:56:04.065588 node.go:250 weedfs:dc1 adds child rack2 I0701 09:56:04.065590 node.go:250 weedfs:dc1:rack2 adds child server126 I0701 09:56:04.065592 node.go:250 weedfs:dc1:rack2:server126 adds child I0701 09:56:04.065595 node.go:250 weedfs:dc1:rack2 adds child server121 I0701 09:56:04.065598 node.go:250 weedfs:dc1:rack2:server121 adds child I0701 09:56:04.065602 node.go:250 weedfs:dc1:rack2 adds child server122 I0701 09:56:04.065607 node.go:250 weedfs:dc1:rack2:server122 adds child I0701 09:56:04.065609 node.go:250 weedfs:dc1:rack2 adds child server123 I0701 09:56:04.065612 node.go:250 weedfs:dc1:rack2:server123 adds child I0701 09:56:04.065616 node.go:250 weedfs:dc1:rack2 adds child server124 I0701 09:56:04.065618 node.go:250 weedfs:dc1:rack2:server124 adds child I0701 09:56:04.065621 node.go:250 weedfs:dc1:rack2 adds child server125 I0701 09:56:04.065623 node.go:250 weedfs:dc1:rack2:server125 adds child I0701 09:56:04.065626 node.go:250 weedfs:dc1 adds child rack3 I0701 09:56:04.065628 node.go:250 weedfs:dc1:rack3 adds child server135 I0701 09:56:04.065630 node.go:250 weedfs:dc1:rack3:server135 adds child I0701 09:56:04.065633 node.go:250 weedfs:dc1:rack3 adds child server136 I0701 09:56:04.065635 node.go:250 weedfs:dc1:rack3:server136 adds child I0701 09:56:04.065638 node.go:250 weedfs:dc1:rack3 adds child server131 I0701 09:56:04.065645 node.go:250 weedfs:dc1:rack3:server131 adds child I0701 09:56:04.065648 node.go:250 weedfs:dc1:rack3 adds child server132 I0701 09:56:04.065650 node.go:250 weedfs:dc1:rack3:server132 adds child I0701 09:56:04.065652 node.go:250 weedfs:dc1:rack3 adds child server133 I0701 09:56:04.065655 node.go:250 weedfs:dc1:rack3:server133 adds child I0701 09:56:04.065658 node.go:250 weedfs:dc1:rack3 adds child server134 I0701 09:56:04.065660 node.go:250 weedfs:dc1:rack3:server134 adds child assigned node : server116 assigned node : server114 assigned node : server123 --- PASS: TestReplication011 (0.00s) === RUN TestFindEmptySlotsForOneVolumeScheduleByWeight data: map[dc1:map[rack1:map[server111:map[limit:2000 volumes:[]]]] dc2:map[rack2:map[server222:map[limit:2000 volumes:[]]]] dc3:map[rack3:map[server333:map[limit:1000 volumes:[]]]] dc4:map[rack4:map[server444:map[limit:1000 volumes:[]]]] dc5:map[rack5:map[server555:map[limit:500 volumes:[]]]] dc6:map[rack6:map[server666:map[limit:500 volumes:[]]]]] I0701 09:56:04.065708 node.go:250 weedfs adds child dc1 I0701 09:56:04.065711 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.065713 node.go:250 weedfs:dc1:rack1 adds child server111 I0701 09:56:04.065715 node.go:250 weedfs:dc1:rack1:server111 adds child I0701 09:56:04.065722 node.go:250 weedfs adds child dc2 I0701 09:56:04.065725 node.go:250 weedfs:dc2 adds child rack2 I0701 09:56:04.065728 node.go:250 weedfs:dc2:rack2 adds child server222 I0701 09:56:04.065731 node.go:250 weedfs:dc2:rack2:server222 adds child I0701 09:56:04.065738 node.go:250 weedfs adds child dc3 I0701 09:56:04.065741 node.go:250 weedfs:dc3 adds child rack3 I0701 09:56:04.065744 node.go:250 weedfs:dc3:rack3 adds child server333 I0701 09:56:04.065746 node.go:250 weedfs:dc3:rack3:server333 adds child I0701 09:56:04.065750 node.go:250 weedfs adds child dc4 I0701 09:56:04.065752 node.go:250 weedfs:dc4 adds child rack4 I0701 09:56:04.065755 node.go:250 weedfs:dc4:rack4 adds child server444 I0701 09:56:04.065758 node.go:250 weedfs:dc4:rack4:server444 adds child I0701 09:56:04.065762 node.go:250 weedfs adds child dc5 I0701 09:56:04.065764 node.go:250 weedfs:dc5 adds child rack5 I0701 09:56:04.065768 node.go:250 weedfs:dc5:rack5 adds child server555 I0701 09:56:04.065771 node.go:250 weedfs:dc5:rack5:server555 adds child I0701 09:56:04.065777 node.go:250 weedfs adds child dc6 I0701 09:56:04.065779 node.go:250 weedfs:dc6 adds child rack6 I0701 09:56:04.065781 node.go:250 weedfs:dc6:rack6 adds child server666 I0701 09:56:04.065784 node.go:250 weedfs:dc6:rack6:server666 adds child server666 : 156 server222 : 538 server555 : 177 server111 : 540 server444 : 295 server333 : 294 --- PASS: TestFindEmptySlotsForOneVolumeScheduleByWeight (0.00s) === RUN TestPickForWrite data: map[dc1:map[rack1:map[serverdc111:map[ip:127.0.0.1 limit:100 volumes:[map[collection:test id:1 replication:001 size:12312] map[collection:test id:2 replication:100 size:12312] map[collection:test id:4 replication:100 size:12312] map[collection:test id:6 replication:010 size:12312]]]]] dc2:map[rack1:map[serverdc211:map[ip:127.0.0.2 limit:100 volumes:[map[collection:test id:2 replication:100 size:12312] map[collection:test id:3 replication:010 size:12312] map[collection:test id:5 replication:001 size:12312] map[collection:test id:6 replication:010 size:12312]]]]] dc3:map[rack1:map[serverdc311:map[ip:127.0.0.3 limit:100 volumes:[map[collection:test id:1 replication:001 size:12312] map[collection:test id:3 replication:010 size:12312] map[collection:test id:4 replication:100 size:12312] map[collection:test id:5 replication:001 size:12312]]]]]] I0701 09:56:04.067967 node.go:250 weedfs adds child dc1 I0701 09:56:04.067975 node.go:250 weedfs:dc1 adds child rack1 I0701 09:56:04.067977 node.go:250 weedfs:dc1:rack1 adds child serverdc111 I0701 09:56:04.067983 volume_layout.go:417 Volume 1 becomes writable I0701 09:56:04.067986 node.go:250 weedfs:dc1:rack1:serverdc111 adds child I0701 09:56:04.067992 volume_layout.go:417 Volume 2 becomes writable I0701 09:56:04.067996 volume_layout.go:417 Volume 4 becomes writable I0701 09:56:04.068000 volume_layout.go:417 Volume 6 becomes writable I0701 09:56:04.068003 node.go:250 weedfs adds child dc2 I0701 09:56:04.068005 node.go:250 weedfs:dc2 adds child rack1 I0701 09:56:04.068008 node.go:250 weedfs:dc2:rack1 adds child serverdc211 I0701 09:56:04.068011 volume_layout.go:405 Volume 2 becomes unwritable I0701 09:56:04.068013 volume_layout.go:417 Volume 2 becomes writable I0701 09:56:04.068021 node.go:250 weedfs:dc2:rack1:serverdc211 adds child I0701 09:56:04.068025 volume_layout.go:417 Volume 3 becomes writable I0701 09:56:04.068028 volume_layout.go:417 Volume 5 becomes writable I0701 09:56:04.068033 volume_layout.go:405 Volume 6 becomes unwritable I0701 09:56:04.068035 volume_layout.go:417 Volume 6 becomes writable I0701 09:56:04.068037 node.go:250 weedfs adds child dc3 I0701 09:56:04.068040 node.go:250 weedfs:dc3 adds child rack1 I0701 09:56:04.068042 node.go:250 weedfs:dc3:rack1 adds child serverdc311 I0701 09:56:04.068045 volume_layout.go:405 Volume 1 becomes unwritable I0701 09:56:04.068046 volume_layout.go:417 Volume 1 becomes writable I0701 09:56:04.068049 node.go:250 weedfs:dc3:rack1:serverdc311 adds child I0701 09:56:04.068052 volume_layout.go:405 Volume 3 becomes unwritable I0701 09:56:04.068054 volume_layout.go:417 Volume 3 becomes writable I0701 09:56:04.068057 volume_layout.go:405 Volume 4 becomes unwritable I0701 09:56:04.068058 volume_layout.go:417 Volume 4 becomes writable I0701 09:56:04.068061 volume_layout.go:405 Volume 5 becomes unwritable I0701 09:56:04.068063 volume_layout.go:417 Volume 5 becomes writable --- PASS: TestPickForWrite (0.00s) === RUN TestVolumesBinaryState === RUN TestVolumesBinaryState/mark_true_when_copies_exist === RUN TestVolumesBinaryState/mark_true_when_no_copies_exist --- PASS: TestVolumesBinaryState (0.00s) --- PASS: TestVolumesBinaryState/mark_true_when_copies_exist (0.00s) --- PASS: TestVolumesBinaryState/mark_true_when_no_copies_exist (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/topology 0.018s === RUN TestByteParsing --- PASS: TestByteParsing (0.00s) === RUN TestSameAsJavaImplementation Now we need to generate a 256-bit key for AES 256 GCM --- PASS: TestSameAsJavaImplementation (0.00s) === RUN TestToShortFileName --- PASS: TestToShortFileName (0.00s) === RUN TestHumanReadableIntsMax --- PASS: TestHumanReadableIntsMax (0.00s) === RUN TestHumanReadableInts --- PASS: TestHumanReadableInts (0.00s) === RUN TestAsyncPool -- Executing third function -- -- Executing second function -- -- Executing first function -- -- Third Function finished -- -- Executing fourth function -- -- Second Function finished -- -- Executing fifth function -- -- First Function finished -- 1 2 3 -- Fourth fifth finished -- -- Fourth Function finished -- 4 5 --- PASS: TestAsyncPool (0.12s) === RUN TestOrderedLock ActiveLock 1 acquired lock 1 ActiveLock 1 released lock 1 ActiveLock 2 acquired lock 0 ActiveLock 3 acquired lock 0 ActiveLock 4 acquired lock 0 ActiveLock 5 acquired lock 0 ActiveLock 2 released lock 0 ActiveLock 5 released lock 0 ActiveLock 3 released lock 0 ActiveLock 4 released lock 0 ActiveLock 6 acquired lock 1 ActiveLock 6 released lock 1 ActiveLock 7 acquired lock 0 ActiveLock 8 acquired lock 0 ActiveLock 9 acquired lock 0 ActiveLock 10 acquired lock 0 ActiveLock 9 released lock 0 ActiveLock 10 released lock 0 ActiveLock 8 released lock 0 ActiveLock 7 released lock 0 ActiveLock 11 acquired lock 1 ActiveLock 11 released lock 1 ActiveLock 12 acquired lock 0 ActiveLock 13 acquired lock 0 ActiveLock 14 acquired lock 0 ActiveLock 15 acquired lock 0 ActiveLock 13 released lock 0 ActiveLock 12 released lock 0 ActiveLock 15 released lock 0 ActiveLock 14 released lock 0 ActiveLock 16 acquired lock 1 ActiveLock 16 released lock 1 ActiveLock 17 acquired lock 0 ActiveLock 18 acquired lock 0 ActiveLock 19 acquired lock 0 ActiveLock 20 acquired lock 0 ActiveLock 20 released lock 0 ActiveLock 19 released lock 0 ActiveLock 17 released lock 0 ActiveLock 18 released lock 0 ActiveLock 21 acquired lock 1 ActiveLock 21 released lock 1 ActiveLock 22 acquired lock 0 ActiveLock 22 released lock 0 ActiveLock 23 acquired lock 0 ActiveLock 24 acquired lock 0 ActiveLock 25 acquired lock 0 ActiveLock 25 released lock 0 ActiveLock 23 released lock 0 ActiveLock 24 released lock 0 ActiveLock 26 acquired lock 1 ActiveLock 26 released lock 1 ActiveLock 27 acquired lock 0 ActiveLock 28 acquired lock 0 ActiveLock 29 acquired lock 0 ActiveLock 30 acquired lock 0 ActiveLock 30 released lock 0 ActiveLock 29 released lock 0 ActiveLock 28 released lock 0 ActiveLock 27 released lock 0 ActiveLock 31 acquired lock 1 ActiveLock 31 released lock 1 ActiveLock 32 acquired lock 0 ActiveLock 33 acquired lock 0 ActiveLock 34 acquired lock 0 ActiveLock 35 acquired lock 0 ActiveLock 34 released lock 0 ActiveLock 33 released lock 0 ActiveLock 35 released lock 0 ActiveLock 32 released lock 0 ActiveLock 36 acquired lock 1 ActiveLock 36 released lock 1 ActiveLock 37 acquired lock 0 ActiveLock 38 acquired lock 0 ActiveLock 38 released lock 0 ActiveLock 39 acquired lock 0 ActiveLock 40 acquired lock 0 ActiveLock 37 released lock 0 ActiveLock 40 released lock 0 ActiveLock 39 released lock 0 ActiveLock 41 acquired lock 1 ActiveLock 41 released lock 1 ActiveLock 42 acquired lock 0 ActiveLock 43 acquired lock 0 ActiveLock 44 acquired lock 0 ActiveLock 45 acquired lock 0 ActiveLock 45 released lock 0 ActiveLock 43 released lock 0 ActiveLock 44 released lock 0 ActiveLock 42 released lock 0 ActiveLock 46 acquired lock 1 ActiveLock 46 released lock 1 ActiveLock 47 acquired lock 0 ActiveLock 48 acquired lock 0 ActiveLock 49 acquired lock 0 ActiveLock 50 acquired lock 0 ActiveLock 47 released lock 0 ActiveLock 50 released lock 0 ActiveLock 48 released lock 0 ActiveLock 49 released lock 0 --- PASS: TestOrderedLock (1.20s) === RUN TestParseMinFreeSpace --- PASS: TestParseMinFreeSpace (0.00s) === RUN TestNewQueue --- PASS: TestNewQueue (0.00s) === RUN TestEnqueueAndConsume 1 2 3 ----------------------- 4 5 6 7 ----------------------- --- PASS: TestEnqueueAndConsume (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/util 1.336s ? github.com/seaweedfs/seaweedfs/weed/util/buffer_pool [no test files] === RUN TestJobQueue enqueued 5 items dequeue 1 dequeue 2 enqueue 6 enqueue 7 dequeue ... dequeued 3 dequeue ... dequeued 4 dequeue ... dequeued 5 dequeue ... dequeued 6 dequeue ... dequeued 7 enqueue 8 enqueue 9 enqueue 10 enqueue 11 enqueue 12 dequeued 8 dequeued 9 dequeued 10 dequeued 11 dequeued 12 --- PASS: TestJobQueue (0.00s) === RUN TestJobQueueClose dequeued 1 dequeued 2 dequeued 3 dequeued 4 dequeued 5 dequeued 6 dequeued 7 --- PASS: TestJobQueueClose (0.00s) PASS ok github.com/seaweedfs/seaweedfs/weed/util/buffered_queue 0.002s ? github.com/seaweedfs/seaweedfs/weed/util/buffered_writer [no test files] === RUN TestOnDisk I0701 09:56:04.065187 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c0_2_0.ldb, watermark 0, num of entries:0 I0701 09:56:04.065682 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c0_2_0.ldb... , watermark: 0 I0701 09:56:04.067326 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c0_2_1.ldb, watermark 0, num of entries:0 I0701 09:56:04.067827 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c0_2_1.ldb... , watermark: 0 I0701 09:56:04.069398 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c1_3_0.ldb, watermark 0, num of entries:0 I0701 09:56:04.069875 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c1_3_0.ldb... , watermark: 0 I0701 09:56:04.072724 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c1_3_1.ldb, watermark 0, num of entries:0 I0701 09:56:04.073189 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c1_3_1.ldb... , watermark: 0 I0701 09:56:04.073760 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c1_3_2.ldb, watermark 0, num of entries:0 I0701 09:56:04.075072 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c1_3_2.ldb... , watermark: 0 I0701 09:56:04.075560 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c2_2_0.ldb, watermark 0, num of entries:0 I0701 09:56:04.076286 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c2_2_0.ldb... , watermark: 0 I0701 09:56:04.077984 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c2_2_1.ldb, watermark 0, num of entries:0 I0701 09:56:04.078496 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c2_2_1.ldb... , watermark: 0 I0701 09:56:04.079488 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c0_2_0.ldb, watermark 0, num of entries:0 I0701 09:56:04.079886 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c0_2_0.ldb... , watermark: 0 I0701 09:56:04.081301 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c0_2_1.ldb, watermark 0, num of entries:0 I0701 09:56:04.081918 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c0_2_1.ldb... , watermark: 0 I0701 09:56:04.083353 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c0_2_0.ldb, watermark 0, num of entries:2 I0701 09:56:04.085018 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c0_2_0.ldb... , watermark: 0 I0701 09:56:04.086591 needle_map_leveldb.go:122 generateLevelDbFile /tmp/TestOnDisk2011646604/001/c0_2_1.ldb, watermark 0, num of entries:1 I0701 09:56:04.087287 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c0_2_1.ldb... , watermark: 0 I0701 09:56:04.087759 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c1_3_0.ldb... , watermark: 0 I0701 09:56:04.088400 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c1_3_1.ldb... , watermark: 0 I0701 09:56:04.089927 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c1_3_2.ldb... , watermark: 0 I0701 09:56:04.090423 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c2_2_0.ldb... , watermark: 0 I0701 09:56:04.090827 needle_map_leveldb.go:66 Loading /tmp/TestOnDisk2011646604/001/c2_2_1.ldb... , watermark: 0 chunk_cache_on_disk_test.go:98: failed to write to and read from cache: 2 --- FAIL: TestOnDisk (0.03s) FAIL FAIL github.com/seaweedfs/seaweedfs/weed/util/chunk_cache 0.038s ? github.com/seaweedfs/seaweedfs/weed/util/fla9 [no test files] ? github.com/seaweedfs/seaweedfs/weed/util/grace [no test files] ? github.com/seaweedfs/seaweedfs/weed/util/http [no test files] ? github.com/seaweedfs/seaweedfs/weed/util/http/client [no test files] ? github.com/seaweedfs/seaweedfs/weed/util/httpdown [no test files] === RUN TestNewLogBufferFirstBuffer processed all messages E0701 09:56:04.104784 log_read.go:115 LoopProcessLogData: test process log entry 1 ts_ns:1751363764104754549 partition_key_hash:-736260903 data:"\xbe\x80\xd8\x02\x93\x90\x08\xf0\xbdϕ\xfc\xf9\x8bG\xablk(#\xc0\xa0>\x9a\xe6\x18\x897\x9c!-P_\xac\x0e׼=v\x1e\x95\x1f\x8co\xb7Hk\nMa\xbc\x97:\xd3\xe1\xdeQ=\xecӓun/\xa2p\xe4Y\xb1*\x85\x17\x04.ǘ\xe0\xa0[\".\xfa\xfd\xf8\xccV=y\xf7\xa7킞P\xa5z\x12\x08\xbdm\xe1Fb\x11\x11\xef\xf3r\x9b\xd3\x0fH\xba\x95p\xf7\x1e\x84\xf4c\x91(\x82\xb0\xa6k\xecDn\xff\xf0\xde \xe4`\xa5\n\x10\x1e\xfc\xaf\xe2\xa4\xe5\xa2\xea\x12\xf4\x0f\xf7\xe84\x7f%\\?\xd8\xd1xS\xac)_6\xa6\xa6\xf8q2̇\xb1\xe1\x17J\x14\xd5E9):\xa3|\xa4\xf0^m\xf4o|;\xe1J:\x8cmַD\x94\xedT}\x97\xa7\xb3\xc2X\x1bB[\xb9Y])\x05\x17\x0e\x0e\xe0v\xa5b\xb1\xe1\xfe\xaeϷ#\xe0\x949!9ِ\xa7\x7f\xbb\x93y\xab\xba\xc0\x84\x86\x1d\xee\xe7\xd6!g݉\xaem\xae\x98S5=\xc2\x0c\x04QE\xf7|{\xa2ұ\xb7Y\xa7\x019\xce\xeb'\x9f\xe0\x10\x88Zo8_\x8b\xcf֣.\xda\x1d\xa3w\xa1\x0b\xa9\xc8/\xed\x15o\xa2Ĕ\"\x89\xf2!\x84\x97\x95s@\xde\x13\xb2\xaeE>\x01\xaeO\xa3h\xf8m3\x0f\xb6v\x81\x18b\x9eg\x9f\x9a>\x06\xe5\x7f\x94\xa4\xb1S\xb3\x8b\x12+x:dҁk}k,H\xf3eYU\xf1u\xb5\x9b=\x9b@\x8b_\xf4\x0e\xd5 [\xee\x07/\xc1k%jC5\xaeӸ\x81\xd6\xd5&\txd*\x9a\x82>f;\x02\xd1\x08\x96\xb4B\x86I%\x1c\x0e\xfd\x88~(5\xaac\x96\xf3yR/\x83W\x98\xb3\xb4\x03\xb8\x89\xd3I\x1e\xba\xadWX\xb9\xb4\xae\xa0\xce7\x7fo'\x9e\xf0J)\xa4\x88\xcf\x10\xef\x00社\x0b^\xb5\xa35V.\x18fB\x0e\xffo?@^\x8f\x8e?\n^v\xc6Y7\xc0\xebR\xf4\xa0&\x02\xffQ?\xfaǵ\x82ՠ\xe3\x9c˃\x8eK\xa1L\xcc\xd4G\xee$ǔ\xca\xccT\xd5\xd0\xe1ϤWf\x9e\xcfρ\xa0\xa5\x86\xfd\xd5\\\x14\x14\xe4\x89K\xff\x0c\xf3\xeb>\xfe\xff9\x03\xda?!S%\xc07\xdf!02r3R\xf4}v\x8d\\W\"\xad\xac\xfd\x0e\x15\xa3d\xa2p\xa0\xd0BObUp\x1e\xa4\xfb\x80\xbc\xbc\x85\xd1\xe3\xc9\x05\xe5\xb8C8\x0b\xcbd\xbdk:T\xdb\nC\xef\xd5>\x82\xcc\xd6G\x18J\x9e\xa1\xfd\x1a\xc3\x0f\xf6Q\xcc\x1b\x1b\xea\x9f\xd1\x00\xce\xe0|\xe9\xc9\xe9\xcf&\x9d\x08\xa2\xbcCA|&s\x1c\xd4~\xb5F\xd8Lk\xe2ͽ\x80i\x96\xb4\xc4Y\x0f\r\x86\xa8\x06\xed\x1bKu\x9d\xd6-\x15,\x05\xfbleG\xac\x05\xb9\x8b\xa3\x8b\xf6f'{\x06'0\xe9\xcdzI\xab*iy($\xbc5'\x91`\xae\x06;=<\x95\xe3 \x98Gp\xad\x04nE\xee\x14\xf6\xeax\xe9\x8e1\x89\x14*Tm\xd7T[tr\xca@el\xa5|\xafwBc\x0ei\xc3\xe2\xe0Ef\xae\xd7\x7fZ\x80\xdan\x9a0/\x8d\xfa\xdfN\xb9\x02)]\xc3\x0e\xb3\xd0J\xcdN7V\xb8\xec_E]g\xff\xd0N!\xf3.\x98\x9d\xc1\xa7C^\xfe̟\xe6\xb2C\x8f\x1a|K\x93\xb1f\xbb\xc6=U\xf1\xb2nS\xfaxsp:\xa5\xb69\x1e\x87ک9s\xb0\x81\xdf\n\x98\xe6M\xd8\xcd\x17_ԎR\xa8;ۅ\x14\xe9LL\xa1\xe7\x08x\xa9\x89b\xe6^\xbf\xd9(\xe1:+B\xcb7\xa5/\xeb\xd7W\x10\xaeэ\x1ax0\xe3\xc5S\xc4O\xae\xa4\xa5?\xabj\xa0M~\xff\r\x06+\xfd\x10\x00O\x9e\x1a\xae\x8e$\xf3=\xf6P\x85Q\x80\xb6\xf2M\xee\xd9\x13\xa7\x10\xec\x83\xf5!\x81\xef@\xe7G\x8bՌ\x01\xae\xb8ׂ%\x1b0\x01\x80\xef+7i0\x91\x85\xc2\xc0Z>S\x9f\xcc0\x0caۭ\xcc\xf2\x98~`\xbd7۳\xcf\\bH\xa1\xad\x9f\x81\x0c ERROR: A failure occurred in check(). Aborting... ==> ERROR: Build failed, check /home/alhp/workspace/chroot/build_cfdf6060-e862-4678-a9c1-13c90f055610/build